DeepSeek launched a free, open-source large-language model in late December, claiming it was developed in just two months at a cost of under $6 million.
Hm even with DeepSeek being more efficient, wouldn’t that just mean the rich corps throw the same amount of hardware at it to achieve a better result?
Only up to the point where the AI models yield value (which is already heavily speculative). If nothing else, DeepSeek makes Altman's plan for $1T in new data-centers look like overkill.
The revelation that you can get 100x gains by optimizing your code rather than throwing endless compute at your model means the value of graphics cards goes down relative to the value of PhD-tier developers. Why burn through a hundred warehouses full of cards to do what a university mathematics department can deliver in half the time?
It will probably not reduce demand. But it will for sure make it impossible to sell insanely overpriced hardware. Now I'm looking forward to buying a PC with a Chinese open source RISCV CPU and GPU. Bye bye Intel, AMD, ARM and Nvidia.