marella / ctransformers

Python bindings for the Transformer models implemented in C/C++ using GGML library.
MIT License
1.81k stars 137 forks source link

Request for support for Llama 2 #57

Closed kennethleungty closed 1 year ago

kennethleungty commented 1 year ago

GGML version from TheBloke is coming soon: https://huggingface.co/TheBloke/Llama-2-7B-GGML

Thanks!

marella commented 1 year ago

It is working without any changes. Model architecture might be same as LLaMA 1:

from ctransformers import AutoModelForCausalLM

llm = AutoModelForCausalLM.from_pretrained("TheBloke/Llama-2-7B-GGML", model_type="llama")

print(llm("AI is going to"))
kennethleungty commented 1 year ago

Thanks!