It’s a bit separate topic and not what was discussed in this thread previously but I will try to answer.
I assume because Nvidia’s moat is in CUDA and chips with high RAM bandwidth optimized specifically for training while competition in inference (where the weights are static) software and hardware is already higher, and going to be even higher still by the time DeepSeek’s optimizations become a de-facto industry standard and induce some additional demand
It’s a bit separate topic and not what was discussed in this thread previously but I will try to answer.
I assume because Nvidia’s moat is in CUDA and chips with high RAM bandwidth optimized specifically for training while competition in inference (where the weights are static) software and hardware is already higher, and going to be even higher still by the time DeepSeek’s optimizations become a de-facto industry standard and induce some additional demand