It may be true that 64-bit integer math isn't faster than 32-bit... but I'd like to test it for myself.
As long as LLR is largely written in x86 assembly, it will remain restricted to x86 procs. (I thought the PS3 AP26 port was genius - the PS3 is a hard cruncher.)
And as long as there are static (archive) libraries in the LLR project for which we have no source, there's no chance we could try 64-bit or other cpu ports.
While it is probably the case someone spent countless months tuning the x86 assembly code inside the LLR, the result is that it blocks out anyone else from playing with the code - and it blocks any chance of using a non-x86 cpu.
Even if the C/C++ version was 15% slower than the assembly version, it'd be worth opening the C/C++ version into the programming community - so it could be played with, tuned, etc. Compilers do amazing jobs optimizing these days - and CPU's are offering faster, more specialized instructions all the time. Not to mention that it would be reasonable to consider porting a C/C++ based LLR to CUDA... something I bet everyone on PG would like to see.