Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

to add qwen-int4 support #1307

Closed
Cloopen-ReLiNK opened this issue Oct 10, 2023 · 5 comments
Closed

to add qwen-int4 support #1307

Cloopen-ReLiNK opened this issue Oct 10, 2023 · 5 comments

Comments

@Cloopen-ReLiNK
Copy link

No description provided.

@wangschang
Copy link

I need qwen-int4 support too!

@wangschang
Copy link

when I run qwen-int . throw a error like this

ValueError: Quantization is not supported for <class 'vllm.model_executor.models.qwen.QWenLMHeadModel'>.

@casper-hansen
Copy link
Contributor

I have a draft PR open in AutoAWQ but the Qwen implementation is buggy compared to other models. Waiting for Qwen to be streamlined a bit more or for someone to help create a workaround.

casper-hansen/AutoAWQ#78

@Minami-su
Copy link

+1

1 similar comment
@raihan0824
Copy link

+1

@hmellor hmellor closed this as completed Mar 25, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants