You must log in or register to comment.
The basic model of DeepSeek-R1 14B was already groundbreaking since it reached the level of GPT-1o. But this does much better by bring it to the level of GPT-4o
Authors are from :
1 - Lightning Rod Labs (USA)
2 - (UK)
London School of Economics and Political Science
Machine learning is still developing very fast.
“We used 8, H100 GPUs, for training.”
Huge amounts of processing power are not required.