← 回總覽

英伟达最强 B200 算力浪费 60%!普林斯顿团队出手,利用率升至 71%

📅 2026-03-18 08:31 闻乐 人工智能 11 分鐘 12706 字 評分: 88
FlashAttention-4 Nvidia Blackwell B200 GPU 优化 算力利用率
📌 一句话摘要 普林斯顿 Tri Dao 团队发布 FlashAttention-4,针对英伟达 Blackwell 架构硬件瓶颈进行深度优化,将 B200 GPU 的算力利用率从 30% 提升至 71%。 📝 详细摘要 本文深度解析了由 Tri Dao 领衔,联合 Meta、Together AI 及英伟达共同研发的 FlashAttention-4 (FA4) 算法。文章指出,英伟达 Blackwell B200 GPU 虽然张量核心算力翻倍,但由于负责指数运算的 MUFU 单元和共享内存带宽并未同步升级,导致严重的硬件“偏科”,使得传统算法下超 60% 的算力被浪费。FA4 通过软件
Skip to main content ![Image 10: LogoBestBlogs](https://www.bestblogs.dev/ "BestBlogs.dev")Toggle navigation menu Toggle navigation menuArticlesPodcastsVideosTweetsSourcesNewsletters

⌘K

Change language Switch ThemeSign In

Narrow Mode

英伟达最强 B200 算力浪费 60%!普林斯顿团队出手,利用率升至 71% ======================================

量子位 @闻乐

One Sentence Summary

Princeton's Tri Dao team released FlashAttention-4, deeply optimizing for Nvidia Blackwell architecture hardware bottlenecks and boosting B200 GPU compute utilization from 30% to 71%.

Summary

This article provides an in-depth analysis of the FlashAttention-4 (FA4) algorithm, co-developed by Tri Dao, Meta, Together AI, and Nvidia. The article points out that while Nvidia's Blackwell B200 GPU doubles its Tensor Core compute power, the MUFU units responsible for exponential operations and shared memory bandwidth have not been upgraded synchronously. This leads to severe hardware "imbalance," wasting over 60% of compute power under traditional algorithms. FA4 successfully resolves these hardware bottlenecks through three core strategies: software-simulated exponential operations, conditional Softmax scaling, 2-CTA MMA mode, and fully asynchronous pipeline reconstruction. Furthermore, FA4 utilizes the CuTe-DSL framework to achieve zero C++ code development, boosting compilation speed by up to 32 times, and demonstrating significantly better performance than cuDNN and Triton frameworks in real-world tests.

Main Points

* 1. Blackwell architecture suffers from severe software-hardware mismatch leading to wasted compute power.B200's Tensor Core compute power doubled, but MUFU units and shared memory bandwidth remained stagnant, making non-matrix multiplication operations (e.g., exponential operations) the new performance bottleneck. * 2. FA4 optimizes exponential operation throughput through software simulation and hardware hybrid computation.Utilizing polynomial approximation to offload MUFU pressure to high-speed FMA units, combined with conditional scaling strategies to skip unnecessary computations, significantly improves the efficiency of non-matrix operations. * 3. Deeply adapting to new Blackwell features to achieve high parallelism in the computational pipeline.Utilizing 2-CTA MMA mode to halve memory read/write volume and combining with fully asynchronous MMA operations to completely overlap Softmax and matrix multiplication, eliminating hardware idle waiting. * 4. The development paradigm shifts to CuTe-DSL, achieving zero C++ development and a leap in compilation efficiency.All code is written based on the Python-domain CuTe-DSL, reducing kernel compilation time from minutes to seconds, with a speedup of up to 32 times, greatly optimizing the development and iteration process.

Metadata

AI Score

88

Website qbitai.com

Published At Today

Length 1939 words (about 8 min)

Sign in to use highlight and note-taking features for a better reading experience. Sign in now

> 闻乐 发自 凹非寺 > > > 量子位 | 公众号 QbitAI

所有用英伟达Blackwell B200的人,都在花冤枉钱??

普林斯顿大学等联合团队指出,这款GPU居然因为软硬件适配问题白白浪费了60%的计算资源

!Image 11

算力浪费了,咋办呢——FlashAttention-4给出了答案。

这款专为Blackwell架构GPU量身打造的注意力算法,一举将利用率从行业普遍的20%-30%推至71%

FlashAttention-4由Tri Dao领衔、携手Meta、Together AI等团队共同研发。

嗯,英伟达自己也参与其中了……

!Image 12

英伟达Blackwell B200作为新一代数据中心GPU,其tensor core张量核心算力达到2.25 PFLOPS,是上一代Hopper H100的2倍。

理论上能让注意力计算的速度实现跨越式提升。

但理想很丰满……

这款GPU发生了严重的偏科。

核心算力猛增的同时,关键的配套计算单元却原地踏步。

其中,负责指数运算的MUFU单元吞吐量与Hopper架构完全一致,没有任何提升;

共享内存的带宽也保持原样,并未跟随张量核心同步升级。

这一硬件设计的不对称性,直接导致了性能瓶颈的反转。

在大模型核心的注意力计算负载中,原本的性能瓶颈矩阵乘法,如今耗时远低于辅助环节,共享内存的读写操作和指数运算的耗时,反而比矩阵乘法多出25%-60%

算力翻倍的Tensor Core长期处于等待状态,大量计算资源就这么被闲置了。

于是,大量开发者花费重金部署的B200 GPU,因核心算力与配套单元的脱节,超六成资源被白白浪费

算力翻倍?

No!明明是有力使不出……

针对Blackwell GPU的偏科问题,FlashAttention-4量身打造了三大优化策略。

!Image 13

第一招,多管齐下化解指数运算与内存读写难题。

团队一方面通过软件模拟指数函数,借助多项式近似的方法,让高速的FMA计算单元参与到原本由MUFU单元负责的指数运算中,大幅提升指数计算的吞吐量;

同时通过混合硬件计算与软件模拟的方式,在提速的同时保证计算精度。

!Image 14

另一方面推出条件性softmax rescaling策略,仅在必要时执行softmax的缩放操作,直接跳过大量无用的计算步骤,减少非矩阵乘法的运算量。

此外,团队充分利用Blackwell架构的2-CTA MMA模式,让两个计算单元搭档完成矩阵运算,各自仅加载一半的运算数据。

这就将共享内存的读写量直接砍半,同时还减少了后续的原子操作,从根源上缓解共享内存的带宽压力。

!Image 15

第二招,重构计算流水线,实现算力的并行最大化。

FlashAttention-4深度适配Blackwell架构的全异步MMA操作和新增的张量内存TMEM,重新设计了注意力计算的前向和反向流水线。

让softmax计算与矩阵乘法这两个核心环节实现完全的计算重叠。

!Image 16

当硬件的张量核心在处理一个矩阵块时,另一部分硬件资源可同时对另一个数据块执行softmax计算,避免硬件算力的空闲。

第三招,兼顾硬件迭代,为下一代GPU预留优化空间。

研发团队同时考虑到Blackwell架构的硬件升级趋势,目前B300/GB300 GPU的指数运算单元吞吐量已翻倍至32 ops/clock/SM

针对这一变化,团队明确表示,FlashAttention-4当前的软件模拟指数运算方案,在下一代硬件上会根据实际性能表现重新权衡,确保算法能持续适配硬件的迭代升级。

除了算法层的深度优化,FlashAttention-4在开发层面也带来了变化。

与此前基于C++模板开发的FlashAttention-3不同,FlashAttention-4的全部代码基于Python的领域专用版本CuTe-DSL框架编写,实现了零C++代码开发

这一设计带来的是编译的效率跃升。

前向传播内核的编译时间从FlashAttention-3的55秒缩短至2.5秒,提速22倍;

反向传播的编译时间从45秒降至1.4秒,提速32倍,整体编译速度最高狂飙30倍。

!Image 17

在B200 GPU上的实测数据显示,其前向传播算力最高达到1613 TFLOPS/s,一举实现71%的理论峰值利用率。

对比主流的计算框架,FlashAttention-4的优势也比较明显。

比英伟达官方的cuDNN 9.13快1.1-1.3倍,比常用的Triton框架快2.1-2.7 倍。

!Image 18

且在长序列、因果掩码等大模型训练推理的核心场景中,性能优势更为突出。

论文还指出,cuDNN从9.13版本开始就已经开始反向吸收了FA4的核心技术。

!Image 19

看来,英伟达自己也忍不住抄作业了(doge)。

论文地址:https://arxiv.org/abs/2603.05451

参考链接:https://x.com/alex_prompter/status/2033885345935462853?s=20

— 完 —

量子位 @闻乐

One Sentence Summary

Princeton's Tri Dao team released FlashAttention-4, deeply optimizing for Nvidia Blackwell architecture hardware bottlenecks and boosting B200 GPU compute utilization from 30% to 71%.

Summary

This article provides an in-depth analysis of the FlashAttention-4 (FA4) algorithm, co-developed by Tri Dao, Meta, Together AI, and Nvidia. The article points out that while Nvidia's Blackwell B200 GPU doubles its Tensor Core compute power, the MUFU units responsible for exponential operations and shared memory bandwidth have not been upgraded synchronously. This leads to severe hardware "imbalance," wasting over 60% of compute power under traditional algorithms. FA4 successfully resolves these hardware bottlenecks through three core strategies: software-simulated exponential operations, conditional Softmax scaling, 2-CTA MMA mode, and fully asynchronous pipeline reconstruction. Furthermore, FA4 utilizes the CuTe-DSL framework to achieve zero C++ code development, boosting compilation speed by up to 32 times, and demonstrating significantly better performance than cuDNN and Triton frameworks in real-world tests.

Main Points

* 1. Blackwell architecture suffers from severe software-hardware mismatch leading to wasted compute power.

B200's Tensor Core compute power doubled, but MUFU units and shared memory bandwidth remained stagnant, making non-matrix multiplication operations (e.g., exponential operations) the new performance bottleneck.

* 2. FA4 optimizes exponential operation throughput through software simulation and hardware hybrid computation.

Utilizing polynomial approximation to offload MUFU pressure to high-speed FMA units, combined with conditional scaling strategies to skip unnecessary computations, significantly improves the efficiency of non-matrix operations.

* 3. Deeply adapting to new Blackwell features to achieve high parallelism in the computational pipeline.

Utilizing 2-CTA MMA mode to halve memory read/write volume and combining with fully asynchronous MMA operations to completely overlap Softmax and matrix multiplication, eliminating hardware idle waiting.

* 4. The development paradigm shifts to CuTe-DSL, achieving zero C++ development and a leap in compilation efficiency.

All code is written based on the Python-domain CuTe-DSL, reducing kernel compilation time from minutes to seconds, with a speedup of up to 32 times, greatly optimizing the development and iteration process.

Key Quotes

* While core compute power surged, key supporting compute units remained stagnant. * The time spent on shared memory read/write operations and exponential calculations is 25%-60% more than matrix multiplication. * All of FlashAttention-4's code is written based on the Python-domain specific version of the CuTe-DSL framework, achieving zero C++ code development. * Forward propagation compute power reaches up to 1613 TFLOPS/s, achieving 71% of theoretical peak utilization in one go. * cuDNN has started to absorb FA4's core technologies in reverse since version 9.13.

AI Score

88

Website qbitai.com

Published At Today

Length 1939 words (about 8 min)

Tags

FlashAttention-4

Nvidia Blackwell

B200

GPU Optimization

Compute Utilization

Related Articles

* GPT-5.4 Released: OpenAI's First Unified Model, Truly Native * Yao Shunyu Lectures Face-to-Face with Tang Jie, Yang Zhilin, and Lin Junyang! Four Schema Heroes Debate Heroes at Zhongguancun * After Topping Open-Source Rankings with its Programming LLM, the Zhipu GLM Team Faced a 3-Hour Questioning Session * A Major Transformation in Attention Mechanisms? Bengio's Team Finds a Hardware-Aligned Solution That Surpasses Transformer * MiniMax Hailuo Video Team's First Open-Source Release: Tokenizer Exhibits a Clear Scaling Law * AI Starts to "Take Action", Alibaba's Qwen Leads the World HomeArticlesPodcastsVideosTweets

Nvidia's Most Powerful B200 Wastes 60% Compute Power! Pri... ===============

查看原文 → 發佈: 2026-03-18 08:31:26 收錄: 2026-03-18 10:00:44

🤖 問 AI

針對這篇文章提問,AI 會根據文章內容回答。按 Ctrl+Enter 送出。