8000
We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I think it should be self_attn_backend, and should not be a model setting but a training setting on top of an inference setting
values should be: "flash2", "pytorch" "pytorch" would include the sdpa_kernels used here: https://github.com/eole-nlp/eole/blob/main/eole/modules/multi_headed_attn.py#L637
we could test if flash2 is installed at training/inference start and adjust backend if necessary.
thus we could remove the flash2 setting from the MHA (redundant with self_attn_backend)
The text was updated successfully, but these errors were encountered:
Successfully merging a pull request may close this issue.
I think it should be self_attn_backend, and should not be a model setting but a training setting on top of an inference setting
values should be: "flash2", "pytorch"
"pytorch" would include the sdpa_kernels used here: https://github.com/eole-nlp/eole/blob/main/eole/modules/multi_headed_attn.py#L637
we could test if flash2 is installed at training/inference start and adjust backend if necessary.
thus we could remove the flash2 setting from the MHA (redundant with self_attn_backend)
The text was updated successfully, but these errors were encountered: