Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Skipping aqlm non working inference tests till fix merged #34865

Open
wants to merge 7 commits into
base: main
Choose a base branch
from

Conversation

MekkCyber
Copy link
Contributor

What does this PR do?

In the inference part of AQLM they use the type torch.Any which was deprecated in torch 2.5. This PR simply skips inference tests with AQLM models till the fix Vahe1994/AQLM#139 is merged to fix the CI.

Who can review ?

@SunMarc

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Member

@SunMarc SunMarc left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for skipping the tests for now. Left a comment

Comment on lines +349 to +359
def skip_if_aqlm_inference_not_fixed(test_case):
"""
Decorator marking tests for inference using aqlm models.

These tests will be skipped till the issue from aqlm side is resolved
"""
return unittest.skip(
reason="inference doesn't work with quantized aqlm models using torch.Any type with recent torch versions"
)(test_case)


Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

don't need to create a seperate function for this as this is something that we will remove. You can just use the decorator like this @skip("reason_for_skipping")

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants