❌

Normal view

Received before yesterday

Scaling NVFP4 Inference for FLUX.2 on NVIDIA Blackwell Data Center GPUs

22 January 2026 at 19:21
In 2025, NVIDIA partnered with Black Forest Labs (BFL) to optimize the FLUX.1 text-to-image model series, unlocking FP4 image generation performance on NVIDIA...

In 2025, NVIDIA partnered with Black Forest Labs (BFL) to optimize the FLUX.1 text-to-image model series, unlocking FP4 image generation performance on NVIDIA Blackwell GeForce RTX 50 Series GPUs. As a natural extension of the latent diffusion model, FLUX.1 Kontext [dev] proved that in-context learning is a feasible technique for visual-generation models, not just large language models (LLMs).

Source

NVIDIA Blackwell Enables 3x Faster Training and Nearly 2x Training Performance Per Dollar than Previous-Gen Architecture

11 December 2025 at 19:20
AI innovation continues to be driven by three scaling laws: pre-training, post-training, and test-time scaling. Training is foundational to building smarter...

AI innovation continues to be driven by three scaling laws: pre-training, post-training, and test-time scaling. Training is foundational to building smarter models, and post-trainingβ€”which can include fine-tuning, reinforcement learning, and other techniquesβ€”helps to further increase accuracy for specific tasks, as well as provide models with new capabilities like the ability to reason.

Source

Top 5 AI Model Optimization Techniques for Faster, Smarter Inference

9 December 2025 at 18:00
As AI models get larger and architectures more complex, researchers and engineers are continuously finding new techniques to optimize the performance and...

As AI models get larger and architectures more complex, researchers and engineers are continuously finding new techniques to optimize the performance and overall cost of bringing AI systems to production. Model optimization is a category of techniques focused on addressing inference service efficiency. These techniques represent the best β€œbang for buck” opportunities to optimize cost…

Source

NVIDIA Blackwell Architecture Sweeps MLPerf Training v5.1 Benchmarks

13 November 2025 at 00:08
The NVIDIA Blackwell architecture powered the fastest time to train across every MLPerf Training v5.1 benchmark, marking a clean sweep in the latest round of...

The NVIDIA Blackwell architecture powered the fastest time to train across every MLPerf Training v5.1 benchmark, marking a clean sweep in the latest round of results. As developers experiment with new architectures, and models continue to grow in size, more training compute is essential. Meeting this need for delivered compute requires innovation across every layer of the AI stackβ€”from chips and…

Source

❌