There is something that makes me uncomfortable about a project that showed up on Hacker News this morning.
It’s called nCPU. The premise: a CPU that runs entirely on a GPU, where every ALU operation — addition, multiplication, bitwise ops, shifts — is implemented as a trained neural network. Not simulated with logic gates. Not approximated with lookup tables in the traditional sense. Learned. Every time you add two numbers, a neural network does it. It uses Kogge-Stone carry-lookahead implemented as a model. Byte-pair lookup tables for multiplication. Attention-based bit routing for bit shifts.









