DeepSpeed/deepspeed/profiling
Logan Adams 170b46e8b1
Add conditional on torch version for scaled_dot_product_attention (#6517)
Changes from #4724 broke support for torch<2.0 in the flops profiler as
the scaled_dot_product_attention [wasn't
added](https://pytorch.org/docs/2.0/generated/torch.nn.functional.scaled_dot_product_attention.html#torch.nn.functional.scaled_dot_product_attention)
until a beta version in torch 2.0

Resolved: #5534

Todo:
- [ ] Test this
- [ ] Issue resolution with users.
2024-09-11 23:21:43 +00:00
..
flops_profiler Add conditional on torch version for scaled_dot_product_attention (#6517) 2024-09-11 23:21:43 +00:00
__init__.py Update DeepSpeed copyright license to Apache 2.0 (#3111) 2023-03-30 17:14:38 -07:00
config.py flops_profiler: add option recompute_fwd_factor for the case of activation recompute (#3362) 2023-06-02 08:20:41 -07:00
constants.py flops_profiler: add option recompute_fwd_factor for the case of activation recompute (#3362) 2023-06-02 08:20:41 -07:00