Open stas00 opened 1 month ago
I am also facing same issue
pip install transformers==4.43.2
Close since no recent update, please feel free to reopen this issue if needed.
The support for Llama 3.2 is not ready yet, please wait for the following releases.
As the OP communicates your documentation says that v3 is supported - so you probably need to change it to be more specific - i.e. only v3.0 is supported and not v3.x
And I also don't understand why did you close this?
Close since no recent update, please feel free to reopen this issue if needed.
Update from whom? You are swiping the issue under the carpet.
The user can't reopen the issue so your suggestion can't work.
Hi @stas00, thank you for raising this issue!
TensorRT-LLM doesn't support Llama 3.2 (yet -- coming soon!), though I suspect from the code snippet shared, the question is about Llama 3.1 which is supported
To run Llama 3.1, you can manually upgrade the transformers version after installing TensorRT-LLM: pip install transformers==4.43.2
(thanks for sharing the workaround @Superjomn)
Please let me know if you have further issues. We are working on upgrading the default transformers dependency to remove this manual step.
I will update the documentation to specify which Llama 3.x versions are supported, and I'll figure out why you don't have permissions to re-open an issue
Thank you for the follow up, @laikhtewari
I see the confusion now - I think I tested with both 3.1 and 3.2 and both were failing but the issue I created said 3.2 and repro I listed 3.1 - my bad!
And as @jinxiangshi suggested (not @Superjomn) - the 3.1 issue is fixable by manual transformers update and you said that v3.2 isn't supported yet and you will update the documentation - now it feels you care. I appreciate that.
I will update the OP.
Oops copied the wrong username, thanks @jinxiangshi !
Sorry by closing the issue, we will amend the dependency requirements and update the document for Llama 3 and Llama 3.1. @stas00
Thank you, @Superjomn
This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 15 days."
edit: both 3.1 and 3.2 fail
pip install transformers==4.43.2
- thank you, @jinxiangshithe 3.1-specific repro:
This doc says https://github.com/NVIDIA/TensorRT-LLM/blob/main/examples/llama/README.md#llama-v3-updates v3 is supported, but clearly it's not. Probably it meant 3.0 and not 3.x? Not sure.
If you copied the HF llama modeling code you need to update it to the latest version for this to work.