R1 Model
An open-source reasoning model released by the Chinese AI startup DeepSeek. Its performance is comparable to OpenAI's O1 model, and it was reportedly trained at a significantly lower cost.
entitydetail.created_at
7/26/2025, 5:17:30 AM
entitydetail.last_updated
7/26/2025, 5:51:53 AM
entitydetail.research_retrieved
7/26/2025, 5:51:53 AM
Summary
The R1 Model, also known as DeepSeek-R1, is a powerful large language model developed by the Chinese AI company DeepSeek. Released on January 20, 2025, under the MIT License, it is notable for its performance, which is comparable to leading proprietary models like OpenAI's GPT-4 and o1, particularly in complex reasoning tasks such as mathematics, coding, and scientific problem-solving. DeepSeek achieved this efficiency through innovative techniques like Mixture of Experts (MoE) layers and by training the model using less powerful hardware during AI chip export restrictions, resulting in significantly lower training costs (e.g., US$6 million for its V3 model, on which R1 is built, compared to US$100 million for GPT-4). The R1 Model's success has been described as "upending AI" and has intensified the US-China AI race and the debate between open-source and closed-source AI. Concerns have been raised, including suspicions of DeepSeek utilizing distillation on OpenAI's models, which also highlights AI model security issues and Microsoft's role in hosting both OpenAI and the R1 Model on Azure AI Foundry.
Referenced in 1 Document
Research Data
Extracted Attributes
Type
Large Language Model (LLM)
Powers
DeepSeek's namesake chatbot
License
MIT License
Sub-type
Reasoning Model
Developer
DeepSeek (Hangzhou DeepSeek Artificial Intelligence Basic Technology Research Co., Ltd.)
Model Status
Open-source (open weight)
Key Techniques
Mixture of Experts (MoE) layers, Chain of Thought prompting, Reinforcement Learning
Primary Capabilities
Logical inference, mathematical problem-solving, code generation, scientific reasoning, reflection, text-based tasks
Training Environment
Less powerful hardware during AI chip export restrictions
Availability Platform
Azure AI Foundry, GitHub
Training Cost (V3 model)
US$6 million (claimed by DeepSeek)
Computing Power Consumption (vs Meta's Llama 3.1)
Approximately one-tenth
Timeline
- DeepSeek, the developer of the R1 Model, was founded by Liang Wenfeng. (Source: Summary, Wikipedia)
2023-07
- The R1 Model was officially released by DeepSeek. (Source: Wikipedia, Web Search)
2025-01-20
- DeepSeek launched its eponymous chatbot, which is powered by the R1 Model. (Source: Wikipedia, Web Search)
2025-01
Wikipedia
View on WikipediaDeepSeek
Hangzhou DeepSeek Artificial Intelligence Basic Technology Research Co., Ltd., doing business as DeepSeek, is a Chinese artificial intelligence company that develops large language models (LLMs). Based in Hangzhou, Zhejiang, Deepseek is owned and funded by the Chinese hedge fund High-Flyer. DeepSeek was founded in July 2023 by Liang Wenfeng, the co-founder of High-Flyer, who also serves as the CEO for both of the companies. The company launched an eponymous chatbot alongside its DeepSeek-R1 model in January 2025. Released under the MIT License, DeepSeek-R1 provides responses comparable to other contemporary large language models, such as OpenAI's GPT-4 and o1. Its training cost was reported to be significantly lower than other LLMs. The company claims that it trained its V3 model for US$6 million—far less than the US$100 million cost for OpenAI's GPT-4 in 2023—and using approximately one-tenth the computing power consumed by Meta's comparable model, Llama 3.1. DeepSeek's success against larger and more established rivals has been described as "upending AI". DeepSeek's models are described as "open weight," meaning the exact parameters are openly shared, although certain usage conditions differ from typical open-source software. The company reportedly recruits AI researchers from top Chinese universities and also hires from outside traditional computer science fields to broaden its models' knowledge and capabilities. DeepSeek significantly reduced training expenses for their R1 model by incorporating techniques such as mixture of experts (MoE) layers. The company also trained its models during ongoing trade restrictions on AI chip exports to China, using weaker AI chips intended for export and employing fewer units overall. Observers say this breakthrough sent "shock waves" through the industry, threatening established AI hardware leaders such as Nvidia; Nvidia's share price dropped sharply, losing US$600 billion in market value, the largest single-company decline in U.S. stock market history.
Web Search Results
- What Is DeepSeek-R1? | Built In
DeepSeek-R1, or R1, is an open source language model made by Chinese AI startup DeepSeek that can perform the same text-based tasks as other advanced models, but at a lower cost. It also powers the company’s namesake chatbot, a direct competitor to ChatGPT. [...] DeepSeek-R1 is an AI model developed by Chinese artificial intelligence startup DeepSeek. Released in January 2025, R1 holds its own against (and in some cases surpasses) the reasoning capabilities of some of the world’s most advanced foundation models — but at a fraction of the operating cost, according to the company. R1 is also open sourcedunder anMIT license, allowing free commercial and academic use. ## What is DeepSeek-R1? [...] R1 is the latest of several AI models DeepSeek has made public. Its first product was the coding tool DeepSeek Coder, followed by the V2 model series, which gained attention for its strong performance and low cost, triggering a price war in the Chinese AI model market. Its V3 model— the foundation on which R1 is built — captured some interest as well, but its restrictions around sensitive topics related to the Chinese government drew questions about its viability as a true industry competitor.
- DeepSeek R1 Explained: Chain of Thought, Reinforcement ...
DeepSeek R1 is a new large language model developed by a research team in China. It’s significant because it demonstrates performance comparable to leading models like OpenAI’s 01 on complex tasks like mathematical, coding, and scientific reasoning. The model’s innovations, particularly in using reinforcement learning and model distillation, could potentially make AI more efficient and accessible. How does DeepSeek R1 use Chain of Thought prompting, and what benefits does it provide? [...] The release of DeepSeek R1, a new large language model from China, has caused a stir in the AI research community. It’s not just another incremental improvement. DeepSeek represents a significant leap forward. Most new AI models feel like small steps. DeepSeek R1 is different. It’s the first model in a while that makes you stop and think, this might be important. [...] DeepSeek R1 uses Chain of Thought prompting by encouraging the model to “think out loud” or provide step-by-step reasoning in its responses. For example, when solving math problems, it will show each step of its work. This approach not only allows for identifying mistakes more easily but also makes it possible for the model to self-evaluate and improve its accuracy through re-prompting or re-evaluation of its steps.
- DeepSeek R1: All you need to know - Fireworks AI
DeepSeek R1 is an open-source AI model that stands out for its reasoning-centric design. While many large language models excel at language understanding, DeepSeek R1 goes a step further by focusing on logical inference, mathematical problem-solving, and reflection capabilities—features that are often guarded behind closed-source APIs. [...] LLM research space is undergoing rapid evolution, with each new model pushing the boundaries of what machines can accomplish. DeepSeek R1, released on January 20, 2025, by DeepSeek, represents a significant leap in the realm of open-source reasoning models. With capabilities rivaling top proprietary solutions, DeepSeek R1 aims to make advanced reasoning, problem-solving, and real-time decision-making more accessible to researchers and developers across the globe. [...] DeepSeek R1 excels at tasks demanding logical inference, chain-of-thought reasoning, and real-time decision-making. Whether it’s solving high-level mathematics, generating sophisticated code, or breaking down complex scientific questions, DeepSeek R1’s RL-based architecture allows it to self-discover and refine reasoning strategies over time. Various independent benchmarks highlight the model’s strong performance:
- What DeepSeek r1 Means—and What It Doesn't | Lawfare
On Jan. 20, the Chinese AI company DeepSeek released a language model called r1, and the AI community (as measured by X, at least) has talked about little else since. The model is the first to publicly match the performance of OpenAI’s frontier “reasoning” model, o1—beating frontier labs Anthropic, Google’s DeepMind, and Meta to the punch. The model matches, or comes close to matching, o1 on benchmarks like GPQA (graduate-level science and math questions), AIME (an advanced math competition), [...] While the DeepSeek news may not signal the failure of American export controls, it does highlight shortcomings in America’s AI strategy. Beyond its technical prowess, r1 is notable for being an open-weight model. That means that the weights—the numbers that define the model’s functionality—are available to anyone in the world to download, run, and modify for free. Other players in Chinese AI, such as Alibaba, have also released well-regarded models as open weight.
- DeepSeek R1 is now available on Azure AI Foundry and GitHub
DeepSeek R1 is now available in the model catalog on Azure AI Foundry and GitHub, joining a diverse portfolio of over 1,800 models, including frontier, open-source, industry-specific, and task-based AI models. As part of Azure AI Foundry, DeepSeek R1 is accessible on a trusted, scalable, and enterprise-ready platform, enabling businesses to seamlessly integrate advanced AI while meeting SLAs, security, and responsible AI commitments—all backed by Microsoft’s reliability and innovation. [...] ## Accelerating AI reasoning for developers on Azure AI Foundry AI reasoning is becoming more accessible at a rapid pace transforming how developers and enterprises leverage cutting-edge intelligence. As DeepSeek mentions, R1 offers a powerful, cost-efficient model that allows more users to harness state-of-the-art AI capabilities with minimal infrastructure investment. [...] One of the key advantages of using DeepSeek R1 or any other model on Azure AI Foundry is the speed at which developers can experiment, iterate, and integrate AI into their workflows. With built-in model evaluation tools, they can quickly compare outputs, benchmark performance, and scale AI-powered applications. This rapid accessibility—once unimaginable just months ago—is central to our vision for Azure AI Foundry: bringing the best AI models together in one place to accelerate innovation and
Location Data
Model, gmina Pacyna, powiat gostyniński, województwo mazowieckie, 09-541, Polska
Coordinates: 52.2890837, 19.7094937
Open Map