TinyLoRA – Learning to Reason in 13 Parameters
66 points by sorenjan 5 days ago | 6 comments

a-t-c-g 20 minutes ago
The quality of custom models trained with proper reasoning datasets[0] even with small parameters (3-7B is sweet spot) is incredible now

[0]: cartesien.io or Salesforce's WebscaleRL

reply
measurablefunc 2 hours ago
With four parameters I can fit an elephant, and with five I can make him wiggle his trunk so there is still room for improvement.
reply
esafak 60 minutes ago
Except learning to reason is a far cry from curve fitting. Our brains have more than five parameters.
reply
voxelghost 9 minutes ago
After a quick content browse, my understanding is this is more like with a very compressed diff vector, applied to a multi billion parameter model, the models could be 'retrained' to reason (score) better on a specific topic , e.g. math was used in the paper
reply
ekuck 10 minutes ago
speak for yourself!
reply
est 26 minutes ago
reasoning capability might just be some specific combinations of mirror neurons.

even some advanced math usually evolves applying patterns found elsewhere into new topics

reply
ValveFan6969 24 minutes ago
[dead]
reply