Tag

Deepseek

All articles tagged with #deepseek

Nvidia's Huang Criticizes AMD-OpenAI Deal Amid $1 Trillion AI Market Growth

Originally Published 3 months ago — by Sherwood News

Featured image for Nvidia's Huang Criticizes AMD-OpenAI Deal Amid $1 Trillion AI Market Growth
Source: Sherwood News

Nvidia's Jensen Huang criticizes OpenAI's deal with AMD, highlighting industry shifts towards diversified AI hardware supply chains. Reflection AI raises $2 billion to develop AI training software, challenging the dominance of massive infrastructure investments. Meanwhile, Amazon's Prime Day savings are minimal, and AI chatbot usage among children is growing. Tesla unveils lower-cost models with downgrades, disappointing some analysts, as the EV market faces pricing and feature challenges.

DeepSeek AI Model in China Cost $294,000 to Train, Developer Reveals

Originally Published 3 months ago — by theregister.com

Featured image for DeepSeek AI Model in China Cost $294,000 to Train, Developer Reveals
Source: theregister.com

DeepSeek's reported $294,000 training cost is misleading; the actual cost to train their base model was around $5.87 million, with the lower figure referring only to a specific reinforcement learning phase, not the entire training process. The article clarifies misconceptions about the expenses involved in developing large AI models and compares DeepSeek's efforts to Western counterparts like Meta's Llama 4.

China's DeepSeek AI Model Costs $294,000 to Train and Innovates Self-Reasoning

Originally Published 3 months ago — by CNN

Featured image for China's DeepSeek AI Model Costs $294,000 to Train and Innovates Self-Reasoning
Source: CNN

DeepSeek, a Chinese AI developer, revealed in a peer-reviewed article that it spent only $294,000 to train its R1 model using 512 Nvidia H800 chips, a significantly lower cost than US rivals, sparking renewed debate over China's role in the AI industry and raising questions about the technology and costs involved in AI development.

A Heartwarming Encounter Between My Mom and Dr. DeepSeek

Originally Published 4 months ago — by Rest of World

Featured image for A Heartwarming Encounter Between My Mom and Dr. DeepSeek
Source: Rest of World

The article discusses how a Chinese kidney transplant patient, overwhelmed by the healthcare system, relies on an AI chatbot called DeepSeek for medical advice, highlighting both its benefits in providing accessible, empathetic support and the risks of inaccuracies and over-reliance on AI in medical care.

OpenAI and U.S. Tech Accelerate AI Innovation Amid Global Competition

Originally Published 5 months ago — by Fortune

Featured image for OpenAI and U.S. Tech Accelerate AI Innovation Amid Global Competition
Source: Fortune

OpenAI's recent decision to release open-source versions of its models marks a significant shift in U.S. AI strategy, driven by China's rapid open-source AI development and competition. Chinese companies like DeepSeek, Baidu, and Tencent are embracing open-source to foster innovation and demonstrate technological prowess, challenging the traditional proprietary approach of U.S. firms. This move reflects a broader geopolitical and economic contest, with the U.S. potentially falling behind in AI leadership as open-source models gain prominence globally.

China Advances AI Strategy Amid US Competition and Global Alliances

Originally Published 5 months ago — by Financial Times

Featured image for China Advances AI Strategy Amid US Competition and Global Alliances
Source: Financial Times

China showcased its AI ambitions at the World Artificial Intelligence Conference in Shanghai, emphasizing open-source models and international cooperation, contrasting with the US's focus on dominance and regulation, while highlighting advancements in robotics and AI safety concerns.

Distillation: Making AI Models More Efficient and Affordable

Originally Published 5 months ago — by Quanta Magazine

Featured image for Distillation: Making AI Models More Efficient and Affordable
Source: Quanta Magazine

DeepSeek's use of knowledge distillation, a widely used AI technique that involves training smaller models using the outputs of larger ones, has sparked controversy but is a common practice in AI development. Originally developed in 2015 at Google to make ensemble models more efficient, distillation helps create smaller, cheaper, and faster AI models by transferring 'dark knowledge' from a teacher to a student model. It has become a fundamental tool in AI, enabling companies like Google, OpenAI, and Amazon to deploy powerful models more efficiently, and continues to be an active area of research and application.

DeepSeek's New AI Model Outperforms OpenAI in Reasoning Tasks

Originally Published 1 year ago — by VentureBeat

DeepSeek, an AI-focused offshoot of High-Flyer Capital Management, has launched the R1-Lite-Preview, a reasoning-focused large language model that rivals OpenAI's o1-preview in performance. Available through DeepSeek Chat, the model excels in logical inference and mathematical reasoning, offering transparency in its thought process. While it has not yet been released for independent analysis or API access, DeepSeek plans to make open-source versions available, continuing its tradition of supporting the open-source AI community.

Chinese Lab's AI Model Challenges OpenAI with Superior Reasoning

Originally Published 1 year ago — by TechCrunch

Featured image for Chinese Lab's AI Model Challenges OpenAI with Superior Reasoning
Source: TechCrunch

DeepSeek, a Chinese AI research company, has released DeepSeek-R1, a reasoning AI model designed to rival OpenAI's o1. This model, which can self-fact-check by taking more time to process queries, performs comparably to o1 on benchmarks like AIME and MATH. However, it struggles with certain logic problems and can be easily jailbroken. DeepSeek-R1 also avoids politically sensitive topics, likely due to Chinese government regulations. The release highlights a shift in AI development towards reasoning models as traditional scaling methods face scrutiny.