We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
感谢你们发布强悍的模型,是否可以发出 awq或者 gptq-int4
The text was updated successfully, but these errors were encountered:
好的 我会反馈需求。此外,社区已经有相应的版本,你可以测试使用~
Sorry, something went wrong.
目前还未找到34b-chat-16k的4bit量化模型,只见过4k上下文版本的量化模型。 如果大家发现了量化后的34b-chat-16k模型,能否分享一下?谢谢!
@zhanghx0905 https://modelscope.cn/models/tclf90/Yi-1.5-34B-Chat-16K-GPTQ-Int4/summary
原始的Yi-1.5-34B-Chat-16K太慢了 同样的Qwen1.5-32B 速度快它一倍 代码生成能力也烂到爆炸
No branches or pull requests
感谢你们发布强悍的模型,是否可以发出 awq或者 gptq-int4
The text was updated successfully, but these errors were encountered: