Closed dopc closed 9 months ago
@dopc Hi, the reason was that we didn't support loadding this model from the HF directly before. model_name = "Intel/neural-chat-7b-v3-1"
should be a local path.
We have update this feature. https://github.com/intel/neural-speed/pull/93. We support the local path & HF both currently.
Please reinstall the Neural Speed from the souce code and try using the HF card id again.
My tests:
Thanks @Zhenzhong1 👍
You are welcome~
I closed this issue. If you have more questions, please feel free to ask.
Hey! Thanks for the great project and for sharing it with the community.
I am trying to inference with the HF neural-chat model.
What I tried
In Colab,
Behaviour
I got an error and below is the full trace.
What I ask
Thanks.