DeepSeek not ‘miracle,’ but impressive: Report debunks Chinese AI app’s $5million claim


With the rising popularity of DeepSeek, a recent report by Bernstein stated that the Chinese AI app looks fantastic but is not a miracle, and it has not been built for $5 million.

The report mentioned that the claim of DeepSeek, which is comparable to ChatGPT by OpenAI, built at a cost of $5 million, is false.

“We believe that DeepSeek DID NOT “build OpenAI for $5M”; the models look fantastic, but we don’t think they are miracles; and the resulting Twitter-verse panic over the weekend seems overblown,” ANI reported, citing the Bernstein report.

“The models they built are fantastic, but they aren’t miracles either,” said Bernstein analyst Stacy Rasgon, who follows the semiconductor industry and was one of several stock analysts describing Wall Street’s reaction as overblown, reported Associated Press.

The two main families of AI models, ‘DeepSeek-V3’ and ‘DeepSeek R1’, have been developed by the Chinese AI app.

The V3 model is a large language model that uses a mixture of expert (MOE) architecture. This architecture combines multiple smaller models to work together, resulting in high performance while using fewer resources than other large models. In total, the V3 model has 671 billion parameters with nearly 37 billion active users at a time.

This includes innovative techniques such as Multi-Head Latent Attention (MHLA), reducing memory usage, and mixed-precision training using FP8 computation for efficiency.

For the V3 model, DeepSeek used a cluster of 2,048 NVIDIA H800 GPUs for nearly two months, 2.7 million GPU hours for pre-training and 2.8 million GPU hours, including post-training.

According to estimates, the cost of this training will be nearly $5 million based on a $2 per GPU hour rental rate. The report claims that this amount doesn’t account for other costs incurred for the development of the model.

DeepSeek R1, which majorly competes with OpenAI models, is built on the V3 foundation but uses Reinforcement Learning (RL) and other techniques to improve reasoning capabilities.

The resources required for the R1 model were very substantial and were not accounted for by the company, the report said.

However, the report acknowledged that DeepSeek’s models are impressive, but the panic and exaggerated claims about building an OpenAI competitor for $5 million are incorrect.



Source link

Share

Latest Updates

Frequently Asked Questions

Related Articles

Remaining Windsurf team and tech acquired by Cognition, makers of Devin: ‘We’re friends with Anthropic again’

Want smarter insights in your inbox? Sign up for our weekly newsletters to...

Crypto-focused Grayscale confidentially files for potential US listing

Crypto-focused asset manager Grayscale said on Monday it has confidentially submitted paperwork with...

Securing the new high ground: tackling export loopholes in space tech

Outer space, the new high ground, is no longer the solitary domain of...

AI models highly vulnerable to health disinfo weaponisation

Artificial intelligence chatbots can be easily manipulated to deliver dangerous health disinformation, raising...
Mahjong Ways 2 Live Draw Hk Live Casino Online Mahjong Ways Judi Bola Online Sabung Ayam Online Judi Bola Online Sv388 Kisah Budi menang Veloz dari Mahjong Perjalanan Kkajhe di Mahjong Ways 2 Hoho sukses temukan trik Mahjong belajar trik spin kecil tono sukses miliki peternakan sapi berkat scatter hitam mahjong ways kisah unik kades bertari yang sukses kelola bumdes dengan scatter hitam mahjong ways