Closed xnoname79 closed 10 months ago
I assume you have already given it a try with the VLLM instruction here: https://docs.vllm.ai/en/latest/getting_started/quickstart.html What went wrong?
Sorry @datquocnguyen, I'm new to this
You're correct; there's nothing wrong when running the model with vLLM. At first glance, I thought the project was built with a completely new architecture that was not yet supported in vLLM. After taking a closer look at the code and familiarizing myself with some terms, I realized that it was built on top of MPT, and the architecture is indeed supported in vLLM.
Thank you for your response; we can consider the thread closed here. I'm still in the learning process, so I genuinely appreciate your corrections on any misunderstood concepts.
Thank you for publishing the project.
I would like to test the model on my local computer using compatible, supported OpenAI APIs, and I see that vLLM is the appropriate project to make it happen.
I would appreciate some advice on making changes and getting the code compatible to run on vLLM. I truly appreciate your help.