cross-posted from: lemmy.sdf.org/post/29607342
Here is the data at Hugging Face.
A team of international researchers from leading academic institutions and tech companies upended the AI reasoning landscape on Wednesday with a new model that matched—and occasionally surpassed—one of China’s most sophisticated AI systems: DeepSeek.
OpenThinker-32B, developed by the Open Thoughts consortium, achieved a 90.6% accuracy score on the MATH500 benchmark, edging past DeepSeek’s 89.4%.
The model also outperformed DeepSeek on general problem-solving tasks, scoring 61.6 on the GPQA-Diamond benchmark compared to DeepSeek’s 57.6. On the LCBv2 benchmark, it hit a solid 68.9, showing strong performance across diverse testing scenarios.
…
autonomoususer@lemmy.world 6 days ago
Is this libre software?
Hotznplotzn@lemmy.sdf.org 6 days ago
Model weights, datasets, data generation code, evaluation code, and training code are all publicly available.
double_quack@lemm.ee 6 days ago
Hey, I came late to the party. I am CS but I am far from AI. Can you point me to a resource where I can learn how to use all of “those things” that you mention?
autonomoususer@lemmy.world 6 days ago
Yeah, that does not answer the question. It looks like it is all under Apache 2.0