Open shihy52x opened 1 year ago
existing FT backend will throw error for llama model.
Will this ever work? I didn't see llama defined under: https://github.com/NVIDIA/FasterTransformer/tree/main/src/fastertransformer/triton_backend
llama
existing FT backend will throw error for llama model.