-
# Alex Strick van Linschoten - My finetuned models beat OpenAI’s GPT-4
Finetunes of Mistral, Llama3 and Solar LLMs are more accurate for my test data than OpenAI’s models.
[https://mlops.systems/pos…
-
### System Info
Tensorrt-LLM commit: 2a115dae84f13daaa54727534daa837c534eceb4
TensorRT-LLM version: 0.11.0.dev2024061800
### Who can help?
_No response_
### Information
- [X] The official exam…
-
https://github.com/OpenMOSS/AnyGPT/blame/6404dbafccc10943be6bf6e24a4b99b3a6545501/anygpt/src/m_utils/prompter.py#L45
Hello,
Is this line correct? Is this for speech-to-speech conversation?
In tha…
-
I am not understanding the conceptual usefulness of masking out the prompt.
I have seen that there is a comment in scripts/prepare_alpaca.py that says:
`mask_inputs: bool = False, # as in alpac…
-
So far with the example of fine tuning I see examples of summarisation, chatbot based on specific use cases etc. However, I want to build the a chatbot based on my own private data (100s of PDF & wor…
-
## Is your feature request related to a problem? Please describe.
Currently we only support vanilla prompt tuning and multi-task prompt tuning from PEFT library. LoRA has shown to give good perform…
-
i cant get it to work with any of L3, or Gemma 2 models. they all throw errors. tried both llamacpp_HF and llamacpp.
-
fig 1
![image](https://user-images.githubusercontent.com/36894403/228507620-edf7513a-cd34-4f80-ab86-f0ed7709e305.png)
-
Hi! Thanks for the great work!
Could you share any configs on fine-tuning Osprey on RefCOCOg dataset? I am trying to follow your work and reproduce the results on it, what's the starting checkpoint…
-
### System Info
- `transformers` version: 4.43.1
- Platform: Linux-5.15.0-1040-nvidia-x86_64-with-glibc2.35
- Python version: 3.11.9
- Huggingface_hub version: 0.24.1
- Safetensors version: 0.4…