Nvidia published fresh benchmark data on Wednesday showing its newest artificial intelligence server delivers a tenfold performance improvement for mixture-of-expert AI models. The tests included two popular Chinese models.
NVIDIA Corporation, NVDA
The timing matters. The AI industry has pivoted from model training to model deployment for end users.
This shift opens the door for competitors like Advanced Micro Devices and Cerebras to challenge Nvidia’s market position.
Nvidia’s data centers on mixture-of-expert AI models. These models work by splitting questions into smaller pieces and routing them to specialized “experts” within the system.
The approach gained traction after China’s DeepSeek released a high-performing open source model in early 2025. That model required less training on Nvidia chips than competing systems.
Since DeepSeek’s release, major players have adopted the mixture-of-experts technique. ChatGPT maker OpenAI now uses it. France’s Mistral does too.
China’s Moonshoot AI jumped on board in July with its own highly-ranked open source model.
Nvidia’s latest AI server crams 72 of its top chips into one machine. Fast links connect the chips together.
The company says this setup boosted Moonshoot’s Kimi K2 Thinking model performance by 10 times compared to the previous server generation. Nvidia reported similar gains with DeepSeek’s models.
The performance jump comes mainly from two factors. First, the sheer number of chips packed into each server. Second, the speed of the connections between those chips.
These are areas where Nvidia still holds advantages over rivals.
The competitive landscape is changing. While Nvidia dominates AI model training, the inference market looks different.
Inference means serving trained models to millions of users. Multiple companies compete here.
Nvidia is making its case that mixture-of-expert models still need powerful hardware for deployment. Even if these models need less training, they require robust systems to serve users.
AMD is working on its own multi-chip server. The company plans to bring it to market next year.
That server will pack multiple powerful chips together, similar to Nvidia’s approach.
The benchmark data comes as Nvidia defends its market position. The company wants to prove its hardware remains essential even as AI model architectures evolve.
Moonshoot AI’s Kimi K2 Thinking model represents the new generation of efficient AI systems. These models train faster and cheaper than older approaches.
But Nvidia’s data suggests deployment still demands high-end hardware. The 10x performance improvement applies specifically to inference workloads.
The company released this data on Wednesday, demonstrating concrete performance metrics for real-world AI models currently in use.
The post Nvidia (NVDA) Stock: New AI Server Delivers 10x Performance Boost appeared first on CoinCentral.



