Big Story · Artificial Intelligence

DeepSeek AI: The Chinese Startup That Shook Silicon Valley

How a small Hangzhou lab, born inside a quantitative hedge fund, built open source models that rival ChatGPT and Gemini, wiped hundreds of billions off the US stock market, and forced the world to rethink who actually leads the AI race.

BrandingX Editorial Team April 24, 2026 14 min read Technology

The Startup That Redefined the AI Race

Few companies in modern technology have reshaped a conversation as quickly as DeepSeek. What began as a quiet side project inside a Chinese quantitative hedge fund evolved in under three years into one of the most disruptive forces in global artificial intelligence, a company that forced Silicon Valley, Wall Street, and Washington to rethink their assumptions at the same time.

DeepSeek is not just another AI chatbot. It is a Chinese lab that has repeatedly built frontier class models with fewer resources, released them for free under open source licenses, and delivered them at a fraction of the cost charged by OpenAI, Google, or Anthropic. Along the way, it has triggered market sell offs, ignited government bans, and inspired a new generation of open source AI builders.

Understanding DeepSeek is, in many ways, understanding the next chapter of the AI era itself.

$600B
Nvidia market value lost in a single day after DeepSeek R1 launched
<$6M
Reported training cost of DeepSeek V3, a frontier class model
#1
Top free app on the US Apple App Store in January 2025
100%
Open weights released on Hugging Face under permissive licenses

Founders and Origin Story

DeepSeek was founded in July 2023 by Liang Wenfeng, a Chinese entrepreneur born in 1985 in Guangdong province. Liang's path to artificial intelligence did not run through Silicon Valley, Stanford, or MIT. It ran through finance.

Liang studied electronic and information engineering at Zhejiang University, one of China's most prestigious institutions, earning a bachelor's and master's degree by 2010. Even as a student, he was writing algorithms that used artificial intelligence to pick stocks. That early obsession with combining machine learning and markets shaped the rest of his career.

In 2015, Liang co-founded a quantitative hedge fund called High-Flyer, based in Hangzhou. The fund used machine learning to predict market movements and make investment decisions. It was an early mover in AI driven trading, and it worked. By 2021, High-Flyer managed assets exceeding 100 billion yuan, roughly $14 billion at the time, placing it among the largest quant funds in China.

Then Liang made a move almost nobody in finance saw coming. Around 2021, he quietly began buying up thousands of Nvidia A100 GPUs, the same chips powering the global AI boom. He did this before the US government imposed export restrictions on high end AI chips to China. That early stockpile of compute, reportedly ten thousand or more GPUs, became the foundation that made DeepSeek possible.

In May 2023, Liang announced that High-Flyer would pivot toward general artificial intelligence research. Two months later, in July 2023, DeepSeek was officially founded as an independent entity fully funded by High-Flyer. The stated mission was not commercialization. It was the pursuit of AGI, an AI system that matches or exceeds human intelligence.

The market is changing. The real decisive force is often not some ready made rules and conditions, but the ability to adapt and adjust to changes.

DeepSeek's hiring philosophy reflected Liang's own outsider approach. The company deliberately recruited young researchers fresh out of university, prioritizing raw ability and research ambition over years of corporate experience. Many of DeepSeek's best engineers were in their twenties.

Early Growth and the January 2025 Breakthrough

For its first eighteen months, DeepSeek worked largely out of the Western spotlight. It released a steady stream of open source models including DeepSeek Coder in late 2023, DeepSeek LLM, then DeepSeek V2 in 2024. Researchers in the open source AI community took notice. The rest of the world mostly did not.

That changed overnight in January 2025.

On January 20, 2025, DeepSeek released DeepSeek R1, a reasoning focused model that matched the performance of OpenAI's top tier o1 model on mathematical and coding benchmarks. Alongside R1, DeepSeek launched a free consumer chatbot app. Within days, the app rocketed to number one on the US Apple App Store, overtaking ChatGPT itself.

The reaction on Wall Street was brutal. On January 27, 2025, US technology stocks sold off sharply. Nvidia alone lost approximately $600 billion in market capitalization in a single trading session, the largest one day drop in US stock market history for any individual company. The reason was simple. DeepSeek had apparently achieved frontier level AI performance for a small fraction of what American labs were spending, and on less advanced chips. The entire investment thesis behind the AI infrastructure boom came under sudden scrutiny.

2015
Liang Wenfeng co-founds High-Flyer
A quantitative hedge fund that uses machine learning for stock trading.
2021
AI Compute Stockpile
High-Flyer quietly acquires thousands of Nvidia A100 GPUs before US export controls.
July 2023
DeepSeek is founded
Incorporated in Hangzhou, funded entirely by High-Flyer, with a mission of pursuing AGI.
Dec 2024
DeepSeek V3 launched
A 671 billion parameter Mixture of Experts model trained for under $6 million.
Jan 2025
DeepSeek R1 and the market shock
R1 matches OpenAI o1, the DeepSeek app hits #1 in the US, Nvidia loses $600B.
Dec 2025
DeepSeek V3.2 and V3.2 Speciale
Gold medal level reasoning models that rival GPT-5 and Gemini 3 Pro.

Key Innovations and Products

DeepSeek's product line is compact, focused, and almost entirely open source. Each release has pushed the envelope in a specific direction.

DeepSeek LLM and Coder

DeepSeek's first public models arrived in late 2023. DeepSeek Coder, released in November 2023, targeted programming tasks and quickly became a favorite among open source developers. The general purpose DeepSeek LLM followed, demonstrating that the small Chinese lab could compete with much larger research groups.

DeepSeek V2

Released in mid 2024, DeepSeek V2 introduced a novel Mixture of Experts architecture that activated only a small portion of its total parameters per query. The result was state of the art performance at a dramatically lower inference cost. V2 also triggered a price war in the Chinese AI API market, forcing competitors including Baidu, Alibaba, and ByteDance to slash their prices.

DeepSeek V3

In December 2024, DeepSeek released V3, a 671 billion parameter Mixture of Experts model that activated roughly 37 billion parameters per token. According to DeepSeek, V3 was trained on a cluster of just 2,000 Nvidia H800 GPUs for under $6 million, a figure that stunned researchers globally. The model delivered performance on par with GPT-4 class systems.

DeepSeek R1

DeepSeek R1, launched in January 2025, was the model that broke DeepSeek into the global mainstream. It was the first open weight reasoning model to clearly match OpenAI's o1 on challenging benchmarks in mathematics, logic, and coding. Because R1 was released on Hugging Face under a permissive license, researchers and startups worldwide could download, fine tune, and deploy it immediately. Thousands did.

DeepSeek V3.2 and V3.2 Speciale

Released in December 2025, the V3.2 family introduced a new sparse attention mechanism called DSA and a scaled up reinforcement learning training pipeline. The high compute variant, V3.2 Speciale, became the first open source model to achieve gold medal level performance across the 2025 International Mathematical Olympiad, the China Mathematical Olympiad, the International Olympiad in Informatics, and the ICPC World Finals. On several benchmarks including AIME 2025 and HMMT, Speciale surpasses both GPT-5 and Gemini 3 Pro.

The DeepSeek App and Chatbot

The consumer facing DeepSeek app is available on iOS, Android, and the web at deepseek.com. It is free to use, answers questions in natural language, generates and explains code, and handles multilingual conversations. The app is powered by the latest DeepSeek models and supports both a fast standard mode and a slower deep reasoning mode.

Business Model and Approach

DeepSeek's business model is unusual. It does not sell enterprise AI products in the traditional sense, it does not rely on venture capital, and it does not push aggressive monetization. Funding comes entirely from High-Flyer, the hedge fund Liang Wenfeng co-founded.

DeepSeek earns revenue from two main channels:

  • API access. Developers and businesses pay per token to use DeepSeek's models through its API. Pricing is aggressive, often an order of magnitude cheaper than equivalent offerings from OpenAI or Google.
  • Enterprise integrations. A smaller but growing line of business for companies that want to deploy DeepSeek models in production environments.

Unlike most AI labs, DeepSeek has stated publicly that commercialization is not its primary goal. Its mission is research, specifically the pursuit of artificial general intelligence. This stance also lets it sidestep certain provisions of Chinese AI regulations that apply more strictly to consumer facing products.

A different kind of AI company. Where OpenAI and Anthropic are racing to monetize, DeepSeek optimizes for research velocity and open release. That choice has made it poorer on paper but massively influential in the broader AI community.

Technical Breakthroughs That Set DeepSeek Apart

DeepSeek's reputation is built on a series of genuine engineering innovations that have changed how researchers think about building large AI models efficiently.

Mixture of Experts Architecture

DeepSeek's flagship models use a Mixture of Experts design, where the full model contains hundreds of billions of parameters but only a small fraction of them activate for any given query. This delivers frontier class capability at a dramatically lower inference cost than dense models of similar size.

DeepSeek Sparse Attention (DSA)

Introduced with V3.2, DSA is a new attention mechanism that makes processing very long contexts far cheaper. For workloads like summarizing long documents or analyzing large codebases, DSA can reduce costs by roughly ten times compared to standard dense attention.

Reinforcement Learning at Scale

DeepSeek spent over 10 percent of its pre-training compute budget on reinforcement learning for its V3.2 Speciale model. This is far above the industry norm. The result is a model that reasons through problems in long, careful chains of thought rather than producing quick but shallow answers.

Training on Less Advanced Hardware

Because US export controls block China from buying top tier Nvidia chips like the H100 and H200, DeepSeek trains on the less powerful H800. That constraint has forced the team to develop unusually efficient training code, squeezing more out of each chip. Ironically, this hardware disadvantage has pushed DeepSeek to innovate in ways that better funded labs often skip.

DeepSeek's Role in the Modern AI Landscape

DeepSeek has changed the global AI conversation in several concrete ways. First, it has made frontier AI genuinely accessible. Anyone with a Hugging Face account can download a state of the art model and run it on their own hardware. Second, it has broken the pricing umbrella at the top of the market. Rival labs have been forced to cut prices, compete harder on value, and explain why their models cost more.

Third, DeepSeek has proven that raw compute is not the only path to progress. Clever architecture, better training code, and smarter data curation can close meaningful gaps with American incumbents. That message resonates loudly in open source and academic communities where massive compute budgets are out of reach.

Finally, DeepSeek has put China firmly on the map as a frontier AI contender. Before DeepSeek, the global AI narrative was overwhelmingly American. Now it is not.

Challenges and Controversies

DeepSeek's rise has not been without friction. Its disruptive posture, combined with its Chinese origin, has attracted intense scrutiny from governments, regulators, and security researchers.

Data Privacy and Chinese Jurisdiction

DeepSeek's privacy policy states that user data from the official app and API is stored on servers in mainland China. Under Chinese law, companies can be compelled to share data with state authorities on request. For individual users that may be a minor concern, but for enterprise and government users, it is often a deal breaker.

Content Censorship

When accessed through the official DeepSeek app or API, the chatbot refuses or deflects on politically sensitive topics in China, including Tiananmen Square, Taiwanese sovereignty, and criticism of the Chinese Communist Party. Self hosted versions of the open weights can be re aligned by developers to remove these restrictions, but the hosted product reflects Chinese regulatory requirements.

Government Bans Around the World

Within weeks of R1's release, a wave of governments restricted DeepSeek:

  • Italy became the first country to ban DeepSeek on a consumer level, with the Italian data protection authority ordering its removal from app stores in January 2025.
  • Australia banned DeepSeek from all federal government devices in February 2025.
  • Taiwan prohibited government agencies, state owned enterprises, and critical infrastructure operators from using DeepSeek.
  • South Korea restricted access across multiple ministries including defense and foreign affairs.
  • In the United States, the Department of Commerce, NASA, the US Navy, and several states including Texas, New York, and Virginia have banned DeepSeek from government devices. Two members of Congress introduced legislation proposing a broader ban.
  • Data protection authorities in France, Ireland, Belgium, and the Netherlands opened investigations into DeepSeek's data handling practices.

Security Incidents

In January 2025, the cloud security firm Wiz discovered that a DeepSeek database had been left publicly exposed on the open internet, leaking sensitive information including chat histories and API keys. DeepSeek patched the issue quickly, but the incident reinforced concerns about the company's operational security maturity.

The US Chip Export Debate

DeepSeek's success has intensified the debate over US export controls. Critics argue that the restrictions are failing to slow Chinese AI progress, since DeepSeek has repeatedly demonstrated frontier performance on allowed hardware. Supporters argue the controls are still working, pointing out that DeepSeek's pre-restriction stockpile of A100 chips was essential to its early progress.

DeepSeek and the Open Source AI Movement

Perhaps DeepSeek's most lasting contribution is its position as the standard bearer for open source frontier AI. When OpenAI chose to keep GPT weights closed and Anthropic followed suit with Claude, Meta's Llama family kept the open source flame alive. DeepSeek has now joined Meta and Mistral at the front of that movement, and arguably surpassed both on reasoning performance.

Every DeepSeek flagship release has included open weights on Hugging Face under permissive licenses. That decision has ripple effects across the entire AI ecosystem:

  • Academic researchers can study state of the art model behavior directly rather than through opaque APIs.
  • Startups can build on DeepSeek without paying licensing fees or accepting vendor lock in.
  • National governments can host models domestically for sovereignty reasons.
  • Individual developers can fine tune DeepSeek for specialized use cases without begging for API access.

This openness has created a genuine ecosystem. Thousands of derivative models, fine tunes, and tools built on top of DeepSeek now exist across Hugging Face and GitHub.

Future of DeepSeek: What Comes Next

DeepSeek's roadmap, to the extent it is public, points in three directions. The first is continued research toward artificial general intelligence. Liang Wenfeng has stated publicly that AGI is the company's north star, and that commercial products are a means to fund that research rather than an end in themselves.

The second is deeper agentic capability. The V3.2 release included a large scale agent task synthesis pipeline designed to train models for complex tool use and multi step autonomous work. Expect DeepSeek's next models to move further into agent territory.

The third is a continued drumbeat of open source releases. DeepSeek has shown no signs of closing off its weights, and community pressure reinforces that stance.

The headwinds are real. Regulatory friction will continue. More government bans are possible. Chip restrictions may tighten. But DeepSeek has already proven its core thesis: that a small team with the right research culture and just enough compute can build AI that rivals the largest labs in the world.

Conclusion: A Company That Rewrote the Rules

DeepSeek began as a side project inside a hedge fund. It is now one of the most consequential AI companies on earth. In under three years, it has delivered frontier class models, triggered a multi hundred billion dollar market correction, inspired a generation of open source builders, and forced governments on three continents to pay attention.

Whether you are a developer, a business leader, a policymaker, or simply curious about where AI goes next, DeepSeek is a name you will keep hearing. Its story is far from finished, and what comes next may matter even more than what has already happened.

Frequently Asked Questions

DeepSeek is a Chinese artificial intelligence company based in Hangzhou that develops large language models. It was founded in July 2023 by Liang Wenfeng and is fully funded by the quantitative hedge fund High-Flyer. DeepSeek is best known for releasing powerful open source AI models like R1 and V3.2 that rival ChatGPT and Gemini.
DeepSeek was founded by Liang Wenfeng, a Chinese entrepreneur born in Guangdong in 1985. Liang previously co-founded the quantitative hedge fund High-Flyer in 2015, which became one of China's largest AI driven investment firms before he pivoted to general AI research and launched DeepSeek in 2023.
No. DeepSeek is privately owned and funded entirely by the hedge fund High-Flyer, which is also founded by Liang Wenfeng. Liang reportedly holds a controlling stake in DeepSeek through two shell corporations. However, as a Chinese company, DeepSeek operates under Chinese laws that can require data sharing with authorities on request.
DeepSeek has released several flagship models including DeepSeek Coder for programming tasks, DeepSeek V2, V3, and V3.2 as general purpose large language models, and DeepSeek R1 plus V3.2 Speciale as dedicated reasoning models. Most of these models are released with open weights on Hugging Face.
When DeepSeek released its R1 reasoning model in late January 2025, it matched the performance of top American models at a fraction of the cost and on less advanced chips. The release triggered a massive sell off in US tech stocks. Nvidia alone lost roughly $600 billion in market value in a single session, the largest single day decline for any company in US stock market history.
Yes. DeepSeek publishes the weights of its flagship models on Hugging Face under permissive licenses, meaning anyone can download, run, fine tune, and build on top of them. This is a stark contrast to ChatGPT and Gemini, which keep their model weights proprietary.
The DeepSeek web app and mobile app are free to use for consumers. Developers can access the API on a pay as you go basis at rates that are significantly cheaper than comparable offerings from OpenAI or Google. The open weights are also free to download and self host.
Italy was the first country to ban DeepSeek on a consumer level in January 2025. Other countries including Australia, Taiwan, and South Korea have banned DeepSeek from government devices. In the US, the Department of Commerce, NASA, the US Navy, and several states such as Texas, New York, and Virginia have banned it on government devices. Investigations have been opened in France, Ireland, Belgium, and the Netherlands.
DeepSeek's safety depends on the use case. For casual personal use, it works much like any other chatbot. However, data submitted through the official DeepSeek app is stored on servers in China under Chinese law, which raises privacy concerns. Security researchers have also flagged previous database exposures. Enterprise users handling sensitive information should either self host the open source weights or avoid the hosted service.
Yes, when used through the hosted DeepSeek app or API, the model follows Chinese content regulations and will decline or deflect questions on politically sensitive topics such as Tiananmen Square, Taiwan, or criticism of the Chinese government. Self hosted versions of the open weights can be modified or re aligned to remove these content policies.
On several benchmarks such as AIME 2025, HMMT, Codeforces, and live math and coding olympiads, DeepSeek V3.2 Speciale matches or beats GPT-5 and Gemini 3 Pro. It is also dramatically cheaper to run. However, Gemini 3 Pro still leads on some benchmarks like Humanity's Last Exam and SWE Verified, and ChatGPT remains stronger for general conversation, tool use, and multimodal tasks.
Yes, in theory. Because DeepSeek publishes open weights, advanced users can download and run the models locally. However, the full flagship models are hundreds of billions of parameters and require serious GPU hardware. Smaller distilled versions of DeepSeek models can run on consumer hardware including high end gaming PCs and Apple Silicon Macs.
As of 2026, DeepSeek has not been banned for all US users. It is banned on federal government devices at agencies like the Department of Commerce, NASA, and the US Navy, as well as on state government devices in Texas, New York, and Virginia. US lawmakers have introduced bills proposing broader restrictions, but no nationwide consumer ban has passed.
DeepSeek is headquartered in Hangzhou, Zhejiang province, in eastern China. The official company name is Hangzhou DeepSeek Artificial Intelligence Basic Technology Research Co., Ltd.