-
There is a new adapter called [LLaMA-Adapter](https://github.com/ZrrSkywalker/LLaMA-Adapter), a lightweight adaption method for fine-tuning instruction-following [LLaMA](https://github.com/facebookres…
-
List below all your findings on CPU's branch prediction unit: inner structure, algorithms, code snippets for testing, projects, etc.
-
List below all your findings on L1I CPU caches: inner structure, algorithms, code snippets for testing, projects, etc.
-
Hello,
Firstly, thanks you all for your work.
I am struggling to understand how to fine-tune T5.
In #113, it is mentionned that there are 2 eos tokens (one for encoder, one for decoder). How…
-
I want to document the process to setup the MySQL database engine and tuning instructions.
It should be see easy for any to follow through and setup local MySQL database engine.
-
Are there any plans to release Gorilla based on Llama 3? Is there any readme where we can find instructions on fine-tuning Llama 3 for Gorilla?
-
Hi, I just tried to have the cyclic-adv-s545 model play against the latest 28b model. However, it doesn't seem to work very well. And I would like to do some fine tuning on my own. I just saw some scr…
-
PandaLM: An Automatic Evaluation Benchmark for LLM Instruction Tuning Optimization. PandaLM is the first to evaluate llm using a finetuned llm.
-
只进行了visual-instruction tuning的phi-2在纯语言任务上的指令跟随表现会不会不够令人满意呢?
-
### Question
Great work! I saw that both the pre-training and instruction-150K dataset has the token inserted in the same format. I was wondering why during the pre-training stage of feature alignme…