Get the latest tech news
DeepSeek might not be as disruptive as claimed, firm reportedly has 50,000 Nvidia GPUs and spent $1.6 billion on buildouts
The fabled $6 million was just a portion of the total training cost.
Due to the talent inflow, DeepSeek has pioneered innovations like Multi-Head Latent Attention (MLA), which required months of development and substantial GPU usage, SemiAnalysis reports. Unlike larger firms burdened by bureaucracy, DeepSeek’s lean structure enables it to push forward aggressively in AI innovation, SemiAnalysis believes. Reality is more complex: SemiAnalysis contends that DeepSeek’s success is built on strategic investments of billions of dollars, technical breakthroughs, and a competitive workforce.
Or read this on r/technology