AI

Deep Dive Into DeepSeek-R1: How It Works and What It Can Do

Published

on

[ad_1]

The dust is still settling after the recent release of DeepSeek R-1, a Chinese large language model that purportedly is on par with OpenAI’s o1 LLM for reasoning tasks, but was trained for about $6 million — a fraction of the approximately $100 million cost to train OpenA1’s o1.

With the R1 model’s weights and inference code being openly released on Hugging Face and GitHub respectively, it’s also worth noting that the training code and the training data itself hasn’t been published. But while DeepSeek seems to be shaping up as an…

[ad_2]

Source link

You must be logged in to post a comment Login

Leave a Reply

Cancel reply

Exit mobile version