Commit 0943771
authored
Separate SAC Wrapping of MoE and Attention Modules to Enable Flex Attention Compilation (#1683)
Flex Attention requires compilation via torch.compile to achieve optimal
performance. Therefore, torch.compile is always applied to Flex
Attention, regardless of the compile.enable flag. However, when
Selective Activation Checkpointing (SAC) is enabled, torch.compile may
be bypassed or invalidated under certain conditions:
1. If compile.enable is set to False, SAC will ignore any torch.compile
calls within the SAC region.
2. If compile.enable is True but the transformer block includes a
Mixture of Experts (MoE) module.
To address this limitation, this PR separates the SAC wrapping of
Attention from MoE and FeedForward modules. This separation ensures that
Flex Attention can be compiled successfully even when SAC is enabled.
Attention module is wrapped with full AC if compile.enable is False.
FIX (workaround) #16311 parent 7e805a9 commit 0943771
File tree
9 files changed
+278
-131
lines changed- tests/unit_tests
- torchtitan
- distributed
- experiments
- llama4/infra
- qwen3/infra
- simple_fsdp
- vlm/infra
- models
- deepseek_v3/infra
- llama3/infra
- tools
9 files changed
+278
-131
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
15 | 15 | | |
16 | 16 | | |
17 | 17 | | |
18 | | - | |
| 18 | + | |
19 | 19 | | |
20 | 20 | | |
21 | 21 | | |
| |||
85 | 85 | | |
86 | 86 | | |
87 | 87 | | |
88 | | - | |
| 88 | + | |
89 | 89 | | |
90 | 90 | | |
91 | 91 | | |
| |||
103 | 103 | | |
104 | 104 | | |
105 | 105 | | |
106 | | - | |
| 106 | + | |
107 | 107 | | |
108 | 108 | | |
109 | 109 | | |
| |||
120 | 120 | | |
121 | 121 | | |
122 | 122 | | |
123 | | - | |
| 123 | + | |
124 | 124 | | |
125 | 125 | | |
126 | 126 | | |
| |||
135 | 135 | | |
136 | 136 | | |
137 | 137 | | |
138 | | - | |
| 138 | + | |
139 | 139 | | |
140 | 140 | | |
141 | 141 | | |
| |||
178 | 178 | | |
179 | 179 | | |
180 | 180 | | |
181 | | - | |
| 181 | + | |
182 | 182 | | |
183 | 183 | | |
184 | 184 | | |
| |||
195 | 195 | | |
196 | 196 | | |
197 | 197 | | |
198 | | - | |
| 198 | + | |
199 | 199 | | |
200 | 200 | | |
201 | 201 | | |
| |||
211 | 211 | | |
212 | 212 | | |
213 | 213 | | |
214 | | - | |
| 214 | + | |
215 | 215 | | |
216 | 216 | | |
217 | 217 | | |
| |||
225 | 225 | | |
226 | 226 | | |
227 | 227 | | |
228 | | - | |
| 228 | + | |
229 | 229 | | |
230 | 230 | | |
231 | 231 | | |
| |||
252 | 252 | | |
253 | 253 | | |
254 | 254 | | |
255 | | - | |
| 255 | + | |
256 | 256 | | |
257 | 257 | | |
258 | 258 | | |
| |||
265 | 265 | | |
266 | 266 | | |
267 | 267 | | |
268 | | - | |
| 268 | + | |
269 | 269 | | |
270 | 270 | | |
271 | 271 | | |
| |||
279 | 279 | | |
280 | 280 | | |
281 | 281 | | |
282 | | - | |
| 282 | + | |
283 | 283 | | |
284 | 284 | | |
285 | 285 | | |
| |||
0 commit comments