Ticker

6/recent/ticker-posts

Deep Seek AI: The Chinese Chatbot That’s Disrupting OpenAI & Shaking the Tech Industry

Chinese AI "Deep Seek" Shakes the World: A Game-Changer in AI Industry

 
Deep Seek AI: The Chinese Chatbot That’s Disrupting OpenAI & Shaking the Tech Industry
Deep Seek AI: The Chinese Chatbot That’s Disrupting OpenAI & Shaking the Tech Industry

Introduction

A new Chinese AI chatbot, Deep Seek R1, has sent shockwaves across the global tech industry. Launched on January 20, 2025, by a Chinese research lab, this AI model has outperformed some of the most advanced AI systems, including OpenAI’s GPT-4o, Meta’s LLaMA, and Google’s Gemini. Its incredible efficiency, low training costs, and groundbreaking reasoning capabilities have disrupted the US tech market, wiping out trillions from the stock valuations of major AI players.

The Impact on US Tech Giants

Before Deep Seek’s launch, NVIDIA was the most valuable AI-focused company with a market valuation of $3.5 trillion. However, within a single day, its value dropped to $2.9 trillion, with a staggering 17% stock decline. The AI-driven stock crash also affected the entire Nasdaq index, reflecting investors' concerns over the sudden rise of a powerful Chinese competitor.

Why Did NVIDIA Suffer the Most?

NVIDIA specializes in AI-training chips, which are crucial for developing and running AI models. The company has been the primary supplier of GPUs for OpenAI, Meta, and Google. However, Deep Seek’s breakthrough lies in its ability to operate with significantly fewer computational resources. It was trained with only $5.6 million—far lower than the billions spent by American companies. This suggests that AI models can now be developed with minimal investment, reducing the reliance on expensive chips like NVIDIA’s H100.

Who Created Deep Seek?

The mastermind behind Deep Seek is Liang Wenfei, a 40-year-old Chinese entrepreneur known for his secrecy. In 2015, he founded High Flyer, a hedge fund using AI for investments. Later, in 2019, he launched High Flyer AI, focusing on AI research. Unlike traditional AI firms, Liang recruited PhD students from China’s top universities instead of experienced engineers. His team focused on developing a Chain-of-Thought (CoT) AI model, which mimics human-like reasoning, making it more logical and efficient than previous AI models.

What Makes Deep Seek R1 Special?

Deep Seek R1 stands out due to its Chain-of-Thought (CoT) architecture. Unlike earlier AI models that generate answers instantly, CoT models "think" before responding, improving reasoning accuracy.

For example, if asked:
"Which is larger: 9.11 or 9.9?"

  • GPT-4o answers immediately: "9.11 is larger" (Incorrect).
  • Deep Seek R1 takes 18 seconds, internally questioning whether the decimal placement matters, then correctly answers: "9.9 is larger."

This ability to break down complex problems step-by-step gives Deep Seek an edge over competitors in mathematics, logic, and scientific reasoning.

Open-Source Revolution & Censorship Issues

One major reason Deep Seek has gained trust is that it is open-source. Unlike OpenAI, which initially promised open-source models but later restricted access, Deep Seek allows developers worldwide to modify and improve the model.

However, Deep Seek does have a major drawback—Chinese government censorship. If asked politically sensitive questions such as:

  • “What happened in Tiananmen Square in 1989?”
  • “Is Taiwan an independent country?”

Deep Seek avoids answering. In contrast, it freely discusses criticisms of Western leaders like Joe Biden and Donald Trump.

Despite this, American firms like Perplexity AI and Microsoft have already adapted Deep Seek by removing censorship, making the model more accessible worldwide.

Performance Comparison: Deep Seek vs. GPT-4o

Recent benchmarks place Deep Seek ahead of GPT-4o in several areas:

  • News Knowledge: 🏆 Deep Seek > GPT-4o
  • Scientific Reasoning: 🏆 Deep Seek > GPT-4o
  • Mathematics & Logic: 🔄 Equal performance
  • Poetry & Creativity: 🏆 GPT-4o > Deep Seek
  • Table Formatting & Structure: 🏆 GPT-4o > Deep Seek

However, Deep Seek has a major latency issue. While GPT-4o responds in 31.1 seconds, Deep Seek takes 71.2 seconds, making it slower due to overwhelming global demand.

Why Deep Seek is More Cost-Efficient

Unlike GPT-4o, which operates as a single large model, Deep Seek uses a Mixture of Experts (MoE) approach:

  • Instead of one AI handling everything, it activates specialized sub-models (e.g., a separate model for engineering, law, or medicine).
  • This reduces computational costs and speeds up responses.
  • While traditional AI models require 1.8 trillion active parameters, Deep Seek operates efficiently with only 37 billion active parameters at a time, significantly lowering energy consumption.

Did Deep Seek Steal from OpenAI?

OpenAI has accused Deep Seek of intellectual property theft, claiming that its model is based on stolen GPT-4 technology. However, some argue that OpenAI itself used unauthorized internet data to train GPT-4, leading to multiple copyright lawsuits from authors and news organizations like The New York Times.

A viral meme humorously depicted OpenAI "fishing" from the internet, only for Deep Seek to steal from OpenAI’s bucket. The situation reflects the growing AI arms race between China and the US.

The Future of AI: Lessons for India & the World

Deep Seek’s success proves that building cutting-edge AI does not require billions of dollars. Despite US sanctions restricting access to advanced chips, Chinese researchers found innovative ways to train AI efficiently.

This should serve as motivation for India and other nations to invest in AI innovation. If China can achieve this with limited resources, India, with its vast pool of tech talent, can do the same.

Post a Comment

0 Comments