
AI Infrastructure2026-02-13
NVIDIA AI Blog
NVIDIA Blackwell Cuts AI Inference Costs Up to 10x with Open Models
A major shift is underway in the economics of artificial intelligence. Early adopters of NVIDIA's new Blackwell platform are reporting dramatic reductions in the cost of running AI models, with savings ranging from 4x to an astonishing 10x per token for inference tasks. This breakthrough is primarily achieved by pairing the advanced Blackwell hardware with highly optimized software stacks designed for open-source models.
The significance of this cost collapse cannot be overstated. Inference—the process of using a trained AI model to make predictions—represents the ongoing, operational expense of AI deployment. By slashing this cost, NVIDIA is removing a critical barrier to widespread adoption. Enterprises that previously found powerful AI models prohibitively expensive to run at scale can now integrate them into more products and services. This move accelerates the trend of 'democratizing' AI, making cutting-edge capabilities accessible to a broader range of companies and developers,
