-
Notifications
You must be signed in to change notification settings - Fork 370
Support mixed and low precision training #764
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
This pull request was exported from Phabricator. Differential Revision: D72415906 |
Summary: We add support for mixed and low precision training in Opacus. Mixed precision training is supported iwth "hooks", "ghost" grad_sample_modes. Low-precision trainig is additionally supported with "functorch". Why there is no functorch support for mixed precision trainig: The backward pass with functorch performs both a forward and backward pass to compute per-sample gradients. The forrwad pass happens outside of the torch.amp context, so it cannot handle mixed precision. Support for low and mixed precision training is GPU dependent. Differential Revision: D72415906
f8f32d2
to
99b5d14
Compare
This pull request was exported from Phabricator. Differential Revision: D72415906 |
Summary: Pull Request resolved: pytorch#764 We add support for mixed and low precision training in Opacus. Mixed precision training is supported iwth "hooks", "ghost" grad_sample_modes. Low-precision trainig is additionally supported with "functorch". Why there is no functorch support for mixed precision trainig: The backward pass with functorch performs both a forward and backward pass to compute per-sample gradients. The forrwad pass happens outside of the torch.amp context, so it cannot handle mixed precision. Support for low and mixed precision training is GPU dependent. Differential Revision: D72415906
99b5d14
to
11fcba8
Compare
Summary: We add support for mixed and low precision training in Opacus. Mixed precision training is supported iwth "hooks", "ghost" grad_sample_modes. Low-precision trainig is additionally supported with "functorch", "ew" Why there is no functorch support for mixed precision trainig: The backward pass with functorch performs both a forward and backward pass to compute per-sample gradients. The forrwad pass happens outside of the torch.amp context, so it cannot handle mixed precision. Support for low and mixed precision training is GPU dependent. Differential Revision: D72415906
11fcba8
to
3e74b9e
Compare
This pull request was exported from Phabricator. Differential Revision: D72415906 |
Summary: We add support for mixed and low precision training in Opacus. Mixed precision training is supported iwth "hooks", "ghost" grad_sample_modes. Low-precision trainig is additionally supported with "functorch", "ew" Why there is no functorch support for mixed precision trainig: The backward pass with functorch performs both a forward and backward pass to compute per-sample gradients. The forrwad pass happens outside of the torch.amp context, so it cannot handle mixed precision. Support for low and mixed precision training is GPU dependent. Differential Revision: D72415906
3e74b9e
to
f2b44aa
Compare
This pull request was exported from Phabricator. Differential Revision: D72415906 |
Summary: We add support for mixed and low precision training in Opacus. Mixed precision training is supported with "hooks", "ghost" grad_sample_modes. Low-precision training is additionally supported with "functorch", "ew" Why there is no functorch support for mixed precision trainig: The backward pass with functorch performs both a forward and backward pass to compute per-sample gradients. The forrwad pass happens outside of the torch.amp context, so it cannot handle mixed precision. Support for low and mixed precision training is GPU dependent. Differential Revision: D72415906
f2b44aa
to
e2fb3db
Compare
This pull request was exported from Phabricator. Differential Revision: D72415906 |
Summary: We add support for mixed and low precision training in Opacus. Mixed precision training is supported with "hooks", "ghost", "functorch" grad_sample_modes. Low-precision training is additionally supported with "ew" Support for low and mixed precision training is GPU dependent. Differential Revision: D72415906
e2fb3db
to
825defb
Compare
This pull request was exported from Phabricator. Differential Revision: D72415906 |
Summary:
We add support for mixed and low precision training in Opacus.
Mixed precision training is supported iw 8000 th "hooks", "ghost" grad_sample_modes.
Low-precision trainig is additionally supported with "functorch".
Why there is no functorch support for mixed precision trainig: The backward pass with functorch performs both a forward and backward pass to compute per-sample gradients. The forrwad pass happens outside of the torch.amp context, so it cannot handle mixed precision.
Support for low and mixed precision training is GPU dependent.
Differential Revision: D72415906