Actually the fact the inference of a SOTA model is completely Nvidia-free is the biggest attack to Nvidia every carried so far. Even American frontier AI labs may start to buy Chinese hardware if they need to continue the AI race, they can't keep paying so much money for the GPUs, especially once Huawei training versions of their GPUs will ship.
By "completely Nvidia-free" do you mean Nvidia wasn't used for training nor inference? Because if it's only inference, we know that Opus already can run on TPUs. Not to mention Gemini.
Yep but they don't run on Chinese hardware that is going to be available to everybody and will cost a lot less than NVIDIA stuff. So now you have a full non-US pipeline for AI, and soon they'll have the training GPUs as well.