Ternary Bonsai: Top Intelligence at 1.58 Bits
17 points by nnx 3 days ago | 4 comments

mchusma 46 minutes ago
Ever since I saw the first one of these one-bit models made by Microsoft, I thought this was a fascinating route. I assume that in practice, this is less helpful than it seems, just because there's every economic incentive in the world for the big AI labs to produce small, powerful, fast models. None of them seem to be using this technique, so it's interesting, but I suspect it's not quite working.

I also have yet to see any of these at a larger scale. For example, can you try one of these at 100 billion parameters?

reply
yodon 34 minutes ago
So excited to see this - the big advantage of 1.58 bits is there are no multiplications at inference time, so you can run them on radically simpler and cheaper hardware.
reply
wmf 58 minutes ago
Yet again they're comparing against unquantized versions of other models. They would probably still win but by a much smaller size margin.
reply
Dumbledumb 2 minutes ago
Wouldnt the margin be higher? All other models being moved from unquantized to quantized would lower their performance, while bonsai stays. I get what you see if it was in regards to score/modelsize, but not for absolute performance
reply