Delivering Massive Performance Leaps for Mixture of Experts Inference on NVIDIA Blackwell
As AI models continue to get smarter, people can rely on them for an expanding set of tasks. This leads usersβfrom consumers to enterprisesβto interact with...
As AI models continue to get smarter, people can rely on them for an expanding set of tasks. This leads usersβfrom consumers to enterprisesβto interact with AI more frequently, meaning that more tokens need to be generated. To serve these tokens at the lowest possible cost, AI platforms need to deliver the best possible token throughput per watt. Through extreme co-design across GPUs, CPUsβ¦
AI has entered an industrial phase. What began as systems performing discrete AI model training and human-facing inference has evolved into always-on AI...
AI innovation continues to be driven by three scaling laws: pre-training, post-training, and test-time scaling. Training is foundational to building smarter...