❗ Enable LLM fine-tuning tests when no quantization is specified #3626
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
All of our fine-tuning tests were getting skipped because our check for cuda was too strict, we only wanted to do that if a quantization_config was specified. This was missed because the tests were too complicated and had a lot of logic in them.
This PR refactors our existing LLM fine-tuning tests by splitting them into two:
All of the without quantization tests will run going forward, and the quantization tests will be skipped if no GPUs are available.
What's included in the non-quantization tests (FYI, none of these were running until now, they were all getting incorrectly skipped):
It also sets the pytest mark once at the top of the file since we were not being consistent with adding this marker throughout.