
Dodgy Huawei Chips Nearly Sunk DeepSeek's Next-Gen R2 Model 18
DeepSeek's development of its next-gen R2 AI model was severely delayed after months of failed training attempts on Huawei's Ascend chips, which suffered from unstable hardware, slow interconnects, and immature software. The Register reports: Following the industry rattling launch of DeepSeek R1 earlier this year, the Chinese AI darling faced pressure from government authorities to train the model's successor on Huawei's homegrown silicon, three unnamed sources have told the Financial Times. But after months of work and the help of an entire team of Huawei engineers, unstable chips, glacial interconnects, and immature software proved insurmountable for DeepSeek, which was apparently unable to complete a single successful training run. The failure, along with challenges with data labeling, ultimately delayed the release of DeepSeek R2 as the company started anew, using Nvidia's H20 GPUs instead. The company has reportedly relegated Huawei's Ascend accelerators to inference duty.
DeepSeek has options (Score:2)
Aside from Huawei and Cambricon, Deepseek also has Innosilicon, Moore Threads and Liusang as alternatives.
I guess that, for the R3 model, they will conduct small scale training trials to see if any of those work for them, to stay within chinese chips for training.
And it would not surprise me that they can achieve sway in to telling the "winner" what features they need emphasized in a future roadmap.
JM2C
YMMV
Re: (Score:1)
Re: DeepSeek has options (Score:2)
So China would have just continued buying everyone else's chips forever and ever amen without these policies?
Re: (Score:3)
None of the Chinese AI chip makers have a workable solution (at least for training). How can I be so sure? Well, if they had a solution that sort of worked, then they would have dumped that product on a world market that is desperately yearning for an Nvidia alternative. That solution doesn't even have to be that good in terms of performance or power. It just has to work ... and be cheaper. The Chinese are good with subsidies and dumping to ensure a cheap price. However, these AI chips are complicated
Re: (Score:2)
Yeah, right now the only plausible alternatives to the big nvidia monster chips seems to be the higher end of apples M3 & M4s , and even they are really only suitable for inference with light training at most. (The key here is memory. Apple's M range chips can use main memory as fast GPU memory allowing them to load in the really big models, at least the models with lots of ram do). Though i've heard AMD are making waves on this front.
There really is an opportunity here for Intel and/or AMD to get in on
Unsurprising (Score:3)
Given that the first DeepSeek relied on foreign hardware, it should come as no surprise that v2 is similarly dependent.
Wait what ???? (Score:1)
The Chinese agents here on /. keeps screeching about how far ahead China is in everything ... which is it?
Re: (Score:1)
Deets? (Score:2)
Does anybody have a detailed article or video about the challenges they faced?
This is seriously interesting material for microarchitecture nerds.
Re: (Score:1)
If Xi tells you, he'll have to kill you.
At least they tried (Score:3)
Everyone is like "Without Nvidia we're nothing" and just buys overpriced cards. The Chinese at least try to build alternatives. And you bet that each good Nvidia card had a hundred bad prototypes. Nvidia just doesn't have so much pressure to test if one can train a model with a card that still has problems as the Chinese when it becomes harder to get Nvidia hardware. In the end it's only a matter of time, matrix multiplication in hardware is no witchcraft. And if there should be a new architecture (e.g. Bitnet that can work with binary operations and bitshifts) everyone starts at 0 again.