Data science
fromTheregister
6 days agoDeepSeek's new models offer big inference cost savings
DeepSeek V4 introduces a new large language model that rivals top American models while reducing inference costs and supporting Huawei's AI accelerators.
I am increasingly asked during candidate interviews how much dedicated inference compute they will have to build with Codex. He added that usage per user is growing much faster than overall user growth, a sign that AI compute is becoming even scarcer and more valuable.