Open levidehaan opened 5 days ago
@Superjomn Would u please take a look this question?
Good suggestion, thanks.
We will support enable_trust_remote_code
in the future when we broaden the model coverage from Llama to other models. Currently, you can also pass an external tokenizer into the LLM
instance instead.
Roger that, i was pointing out that youre not using the extra kwargs, i enabled them and stopped getting errors loading models that needed some of those settings. Might be safe to enable,, or maybe set a flag to enable?
https://github.com/NVIDIA/TensorRT-LLM/blob/9691e12bce7ae1c126c435a049eb516eb119486c/tensorrt_llm/hlapi/tokenizer.py#L63