Open activezhao opened 11 months ago
Can you try latest main branch again? The commit is 37ed967.
@byshiue OK, I will try it, thanks.
@activezhao is it work? I meet same problem 0.0
Can you try latest main branch again? The commit is 37ed967.
@byshiue I try this, modify all_models/inflight_batcher_llm/preprocessing/1/model.py manully, but it doesn't work, my model is codellama with finetune
Can you try latest main branch again? The commit is 37ed967.
@byshiue I try this, modify all_models/inflight_batcher_llm/preprocessing/1/model.py manully, but it doesn't work, my model is codellama with finetune
Can you explain what change do you make and what error do you encounter?
@byshiue I review above commit, I think the modify of 'all_models/inflight_batcher_llm/preprocessing/1/model.py' is useful for me, I replace this py file instead of my 'preprocessing/1/model.py'' file, but occur same error, the error like this:
@activezhao is it work? I meet same problem 0.0
Hi @chenwenjun-github You can refer to my compilation steps https://github.com/triton-inference-server/tensorrtllm_backend/issues/128#event-11020523366
Here is also a new document https://github.com/triton-inference-server/tensorrtllm_backend/blob/main/docs/llama.md, which provides end to end steps. Please try following these steps first on latest main branch.
Same question, is there any feasible solution? It's hard to debug as the source code may be in the .so file.
Same question, is there any feasible solution? It's hard to debug as the source code may be in the .so file.
Well, use the latest code in main branch and following the instructions in https://github.com/triton-inference-server/tensorrtllm_backend/blob/main/docs/llama.md solves the problem, thanks.
I use the latest
tensorrtllm_backend
andTensorRT-LLM
of main branch to get docker images.https://github.com/triton-inference-server/tensorrtllm_backend/tree/main#option-3-build-via-docker
And I use the following command for building engines with codeLlama-7b:
I get the following files:
But, when I call the endpoint, I get error:
How to resolve it?
Thanks.