The dust is still settling after the recent release of DeepSeek R-1, a Chinese large language model that purportedly is on par with OpenAI’s o1 LLM for reasoning tasks, but was trained for about $6 million — a fraction of the approximately $100 million cost to train OpenA1’s o1.
With the R1 model’s weights and inference code being openly released on Hugging Face and GitHub respectively, it’s also worth noting that the training code and the training data itself hasn’t been published. But while DeepSeek seems to be shaping up as an…