-
Notifications
You must be signed in to change notification settings - Fork 75
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Test Fix] Fix Consecutive oneshot #971
Conversation
👋 Hi! Thank you for contributing to llm-compressor. Please add the ready label when the PR is ready for review. |
/ready |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks fine, I feel we can make a few functions cleaner by using pathlib. But I'll leave that upto you
…-compressor into fix-test-conseq-oneshot
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I believe the purpose of this test to test consecutive runs without model reloading.
There are quite a few helper functions which I'm not entirely convinced are needed as of now. I think it might be worth revisiting whether they're actually needed for these tests
…-compressor into fix-test-conseq-oneshot
…-compressor into fix-test-conseq-oneshot
~~Contingent on merge of huggingface/transformers#34719 ~~ ^ has been merged not yet released ~~ ^ has been released Blocked on neuralmagic/compressed-tensors#237 SUMMARY: * In multiple optimization tests, automatically decompress model if provided as optimized model * Fix recipe stage length * Revive old code * When running multiple optimizations (ex. oneshot then finetune, oneshot and oneshot), the recipes needs to be added to the session using `initialize_recipe`. Example here https://github.com/vllm-project/llm-compressor/pull/971/files#diff-c9ae8b3ad24d13abeea5b649a5fd6d0b0925f5c9cc40220cbfbe21ae81242f8dR63-R65 TEST PLAN: ran the test using transformers main Must pass tests/llmcompressor/transformers/obcq/test_consecutive_runs.py --------- Co-authored-by: Dipika Sikka <[email protected]> Co-authored-by: Rahul Tuli <[email protected]> Signed-off-by: Rahul Tuli <[email protected]>
Contingent on merge of huggingface/transformers#34719~~ ^ has been merged not yet released ~~
^ has been released
Blocked on
neuralmagic/compressed-tensors#237
SUMMARY:
initialize_recipe
. Example here https://github.com/vllm-project/llm-compressor/pull/971/files#diff-c9ae8b3ad24d13abeea5b649a5fd6d0b0925f5c9cc40220cbfbe21ae81242f8dR63-R65TEST PLAN:
ran the test using transformers main
Must pass tests/llmcompressor/transformers/obcq/test_consecutive_runs.py