Open ajamjoom opened 8 months ago
Could you share your commit? From your log, the error happens at
convert_py: File "/usr/local/lib/python3.10/dist-packages/tensorrt_llm_toolkit/build_scripts/llama/build.py", line 549, in parse_arguments
convert_py: lora_config = LoraConfig.from_hf(args.hf_lora_dir,
But in latest main branch at https://github.com/NVIDIA/TensorRT-LLM/blob/main/examples/llama/build.py#L549, the code is different.
Dockerfile
# Start from the official AWS DJL 0.26 trt 0.7.1 inference container
FROM 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.26.0-tensorrtllm0.7.1-cu122
# update trt to latest pre-release
RUN pip3 install tensorrt_llm -U --pre --extra-index-url <https://pypi.nvidia.com>
RUN pip3 show tensorrt_llm
Ref to the AWS large-model-inference-containers, where I got the base container from.
When I build the docker image, I see these logs:
Uninstalling tensorrt_llm-0.7.1:
Successfully uninstalled tensorrt_llm-0.7.1
Successfully installed tensorrt_llm-0.8.0.dev2024011601
Step 4/4 : RUN pip3 show tensorrt_llm
---> Running in 0372e69a67ea
Name: tensorrt-llm
Version: 0.8.0.dev2024011601
Summary: TensorRT-LLM: A TensorRT Toolbox for Large Language Models
Home-page: <https://github.com/NVIDIA/TensorRT-LLM>
Author: NVIDIA Corporation
Author-email:
License: Apache License 2.0
Location: /usr/local/lib/python3.10/dist-packages
Requires: accelerate, build, colored, cuda-python, diffusers, evaluate, janus, lark, mpi4py, numpy, nvidia-ammo, onnx, optimum, polygraphy, psutil, pynvml, sentencepiece, tensorrt, torch, transformers, wheel
Required-by:
Removing intermediate container 0372e69a67ea
So I should be on tensorrt_llm-0.8.0.dev2024011601
, which should be the latest pre-release. I'm unsure how to check for the commit as I don't see this pre-release tagged in the repo. I wonder if this pre-release is potentially pointing at an older version rather than the latest (saw this issue report about prev issues resurrecting).
I looked into some of the latest commits and it seems that I'm using this one although my docker logs show that I'm using tensorrt_llm0.8.0.dev2024011601
.
Could you run git log
to show the commit you use?
Setup Machine: AWS Sagemaker ml.p4d.24xlarge Model: https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1
Used Docker container image with the latest build of trt-llm (
0.8.0.dev2024011601
)Build arguments
Error log
TypeError: LoraConfig.from_hf() missing 1 required positional argument: 'trtllm_modules_to_hf_modules'
More complete logs:
CC: @byshiue, @symphonylyh