Start now →

Meituan Just Dropped a Trillion-Parameter Model Trained on 60,000 Domestic GPUs — Not a Single…

By TechSilk · Published May 3, 2026 · 5 min read · Source: DataDrivenInvestor
EthereumAI & Crypto
Meituan Just Dropped a Trillion-Parameter Model Trained on 60,000 Domestic GPUs — Not a Single…
Photo by Van Tien Le on Unsplash

Meituan Just Dropped a Trillion-Parameter Model Trained on 60,000 Domestic GPUs — Not a Single NVIDIA Chip in Sight

On April 24, Meituan quietly did something that should be on every AI developer’s radar.

No press conference. No hype cycle. Just a testing invite page for LongCat-2.0-Preview — their next-generation foundation model.

The specs: 1 trillion+ parameters. MoE architecture. 1 million token context window. Trained on a cluster of 50,000 to 60,000 accelerator cards.

But here’s the part that actually matters.

Every single one of those cards was made in China. Zero NVIDIA. Zero imports.

This is reportedly the largest AI model training run ever completed entirely on domestic Chinese compute infrastructure.

Let that satisfying detail sit for a moment.

Wait — Meituan? The Food Delivery Company?

Yes. The same company you’d associate with ordering bubble tea and booking hotels in China.

But if your mental model of Meituan is still “local services app,” you’re several chapters behind. Starting in 2024, founder Wang Xing began making an unmistakable pivot. During an earnings call, he dropped a line that raised eyebrows: “We will continue to invest billions of dollars to ensure adequate computing power.”

Billions. With a B. In USD.

Then came the investment spree. Meituan poured capital into domestic chip companies: Moore Threads, Muxi, Unisoc, Aixin Yuanzhi — all Chinese GPU and semiconductor firms. They also backed AI model companies like Zhipu AI and Moonshot AI (the makers of Kimi).

Silicon to compute to models. Meituan was quietly building an end-to-end domestic AI supply chain.

560 Billion to 1 Trillion — In Just 7 Months

Let’s trace the timeline, because it’s genuinely impressive.

September 2025: Meituan open-sourced LongCat-Flash — a 560B-parameter MoE model with 27B active parameters. Pre-trained on 20 trillion tokens in 30 days. It went semi-viral for its inference speed. Most people’s first reaction: “Wait, Meituan makes foundation models?”

Then came a rapid-fire model family:

And now, LongCat-2.0-Preview: nearly 2x the parameter count of LongCat-Flash, jumping from 560B to the trillion-parameter club.

Seven months. From open-source underdog to trillion-parameter contender.

The Real Story: China Just Proved Domestic Compute Can Train Frontier Models

Trillion-parameter models aren’t exactly rare in 2026. DeepSeek V4, GPT-5.5, Qwen 3.6 Plus — everyone’s operating at this scale now.

But here’s what makes LongCat-2.0-Preview different: it was trained entirely on Chinese-made hardware.

50,000 to 60,000 domestically produced accelerator cards. End to end — pre-training through inference — with zero reliance on imported chips.

You might be thinking: haven’t Chinese companies been using domestic GPUs for a while now? What’s the big deal?

Here’s the distinction. Previous efforts were largely at the “it technically works” stage. LongCat-2.0-Preview proves something fundamentally different: domestic compute doesn’t just work — it works at frontier scale. The gap between “functional” and “production-grade” has been crossed.

In the context of ongoing U.S. chip export controls targeting China’s AI sector, this isn’t just a technical milestone. It’s a geopolitical signal. The export restrictions were designed to slow China’s AI progress by choking off access to cutting-edge chips. This training run suggests the strategy may be losing its bite.

It Launched the Same Day as DeepSeek V4 — Coincidence?

Here’s a fun fact. LongCat-2.0-Preview and DeepSeek V4 both opened testing on the exact same day — April 24.

Two trillion-parameter MoE models. Same day. Both supporting million-token context windows. Both from Chinese companies.

In China’s AI ecosystem, this kind of “coincidence” typically means one thing: nobody wants to be second. The race has entered its final lap, and the front-runners are sprinting.

The Chinese foundation model landscape is rapidly consolidating. The era of “a hundred models competing” is over. What’s left is a small group of players with the compute, data, and engineering depth to operate at the frontier. DeepSeek is one. Meituan just made a loud case that it’s another.

Why Agent-First Architecture Matters Here

LongCat-2.0-Preview was optimized for three things: code generation, complex task planning, and enterprise automation.

Notice the common thread? These are all AI agent workloads.

A 1-million-token context window + trillion parameters + efficient MoE routing = a model architecturally built for long-chain, multi-step agentic workflows. This isn’t a chatbot. It’s an execution engine.

Wang Xing has been explicit about the endgame: transforming the Meituan app into an “AI-powered app.” Picture an AI assistant embedded in an app used by tens of millions daily — one that understands your full conversation history, calls multiple backend systems, and autonomously orchestrates complex tasks.

And Meituan’s moat? It processes tens of millions of real orders every single day. That’s not a synthetic benchmark. That’s the world’s most demanding AI agent testing ground.

So What Does This Actually Mean?

Three takeaways:

Scale: Meituan has entered the trillion-parameter tier, placing it alongside the world’s leading foundation model labs.

Sovereignty: The largest AI training run ever completed on fully domestic Chinese hardware. In a world of chip export controls, this changes the equation.

Application: This isn’t a model built to top leaderboards. It’s built to power agent workflows inside one of the world’s highest-traffic consumer platforms.

Put it all together, and the narrative writes itself: A Chinese “food delivery” company just trained a trillion-parameter flagship model on entirely domestic chips — and plans to deploy it inside an app that serves tens of millions of people daily.

If you’re tracking the global AI race, LongCat-2.0-Preview deserves five minutes of your attention.


Meituan Just Dropped a Trillion-Parameter Model Trained on 60,000 Domestic GPUs — Not a Single… was originally published in DataDrivenInvestor on Medium, where people are continuing the conversation by highlighting and responding to this story.

This article was originally published on DataDrivenInvestor and is republished here under RSS syndication for informational purposes. All rights and intellectual property remain with the original author. If you are the author and wish to have this article removed, please contact us at [email protected].

NexaPay — Accept Card Payments, Receive Crypto

No KYC · Instant Settlement · Visa, Mastercard, Apple Pay, Google Pay

Get Started →