Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add compression tests for qwen2-vl #4

Merged
merged 2 commits into from
Dec 16, 2024

Conversation

nikita-savelyevv
Copy link

What does this PR do?

Quantization tests passed huggingface#734

Fixes # (issue)

Before submitting

  • This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • Did you make sure to update the documentation with your changes?
  • Did you write any new necessary tests?

@eaidova eaidova merged commit bf44a19 into eaidova:ea/qwen2vl Dec 16, 2024
1 check passed
eaidova added a commit that referenced this pull request Dec 23, 2024
* Support AWQ models

* Add tests

* Add dependencies

* Fix tests

* enable awq export only if ov support it

* fix style (#2)

* disable awq and gptq install for old torch (#3)

* fix style

* disable autogptq and autoawq install for old transformers testing

* separate common quant models patching and gptq (#4)

* disable windows install (huggingface#5)

* separate common quant models patching and gptq

* disable awq windows

* skip logits check for quantized models (huggingface#6)

* fix test after rebase

* fix testing condition for 2024.6 and unpatch in case if failed

* Fix qwen2-vl tests (huggingface#1084)

* Skip private mdoel loading test for external contributors (huggingface#1082)

* Fix reshaping unet if timestep is 0d tensor (huggingface#1083)

* Disable kv cache compression for fp vlm (huggingface#1080)

* Support AWQ models

* Add tests

* Add dependencies

* Fix tests

* enable awq export only if ov support it

* fix style (#2)

* disable awq and gptq install for old torch (#3)

* fix style

* disable autogptq and autoawq install for old transformers testing

* separate common quant models patching and gptq (#4)

* disable windows install (huggingface#5)

* separate common quant models patching and gptq

* disable awq windows

* skip logits check for quantized models (huggingface#6)

* fix test after rebase

* fix testing condition for 2024.6 and unpatch in case if failed

* add necessary packages in test_openvino_full

* fix code style after rebase (huggingface#7)

---------

Co-authored-by: eaidova <[email protected]>
Co-authored-by: Nikita Savelyev <[email protected]>
Co-authored-by: Ella Charlaix <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants