Closed zigzagcai closed 6 months ago
yes for sure
Llama Models Serving Documents
and in case you want to pre-train fine-tune a model or maybe use RL-HF I'm creating a tutorial for that Right Now
here you go
here you go
Thanks! Just another question about the dataset.
I see in the docs dataset_train = load_dataset('REPO_ID_PATH_TO_DATASET')
, but could you share an example of such REPO_ID_PATH_TO_DATASET
?
yes here you can see an example for how to preprocess data for EasyDel Trainer but you need to install the library from pip+git cause I have added this right now and it's now supported via when you install package via pip install EasyDel
pip install git+https://github.com/erfanzar/EasyDeL.git
I am sorry I still have some confusion.
Sicne the readme indicates that flash_attn
is implemented, I want to know how flash attention is implemented on llama.
I know there is a jax implementation of flash_attn
called flash-attention-jax, but I cannot find such codes that import flash_attn
or implement it.
I got it. Thanks @erfanzar ! You did awesome work!
Hi, I am trying to use EasyDel to democratize LLaMa, but I cannot find guide about how to run this model. Could you please give me some hint about the EasyDel LLaMa launch script, such as
repo_id
anddataset_name
? Thanks!