I mean how to determine which intermediate variables to save for the most efficient backward. Still trying to understand the post Min-cut optimal(*) recomputation (i.e. activation checkpointing) with AOTAutograd - #9 by Chillee .
Related Topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
Torch.compile with AOT Autograd can be debugged now! | 1 | 455 | October 31, 2023 | |
`torch.compile` `AOTAutograd` backwards _inductor function | 0 | 198 | January 23, 2024 | |
[Fatal Bug] changed nn.Module.training does not trigger recompilation | 0 | 276 | July 20, 2023 | |
How to trace torch.autograd.backward or torch.autograd.grad? | 3 | 409 | December 5, 2023 | |
`compile_autograd` | 0 | 159 | January 26, 2024 |