Closed xiangzez closed 7 months ago
And update the description of Converter in README.
How about quantization:Optional[str]= "gpqt"
? use_gptq
is inconvenient to add new types in future
For documentation, I think we need a separate tutorial page for quantization. @miaojinc already wrote a doc and we should update this doc and put it in main branch.
For documentation, I think we need a separate tutorial page for quantization. @miaojinc already wrote a doc and we should update this doc and put it in main branch.
Sure, I will update the doc after this PR merged.
@miaojinc Could you add it to our docs, source code is under docs branch.
For documentation, I think we need a separate tutorial page for quantization. @miaojinc already wrote a doc and we should update this doc and put it in main branch.
Sure, I will update the doc after this PR merged.
@miaojinc Could you add it to our docs, source code is under docs branch.
Yes, sure. I will do that in a new pull request for quantization document.
@xiangzez CI failed on baichuan model?
@changqi1 CI issue should be fixed in #287
@changqi1 CI issue should be fixed in #287
rebase or merge main branch?
We could rebase to check this PR's status.
It's preferable to an additional param for quantification rather than creating a new converter, as this approach allows us to integrate new quantification methods in the future, like AWQ