-
Is there any chance that we can have a full description on how to fine-tune pre-trained models (for example in machine translation)? I've managed to continue training on a much smaller dataset (by usi…
y3nk0 updated
4 years ago
-
According https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard Qwen1.5 model is one of the best OpenSource (Free) models with large context and Russian language support. It would be nice to …
-
**Is your feature request related to a problem? Please describe.**
I think integrating Fireworks AI is a good choice for enhancing TaskingAI. Fireworks AI offers a platform focused on generative AI, …
-
Hi there,
Thank you for bringing the elegant RAG Assessment framework to the community.
I am an AI engineer from Alibaba Cloud, and our team has been fine-tuning LLM-as-a-Judge models based on t…
-
## Explaining Data Patterns in Natural Language with Language Models
2023 BlackboxNLP Workshop at ACL | MSR & Cornell U
不断的生成解释,并进行排序。找出一个最具解释性的prompt。
Explanation: symbolic regression,
## Auto…
-
### Feature Description
Method `get_params` for all RecTools models. It outputs a dict of all hyper-params available for tuning together with the values of the current instance. Wrapped models params…
-
Fine-tuning:
- 4 models are working on Ollama (3 tinyLlama verisons with 1, 10, 50 epoch)
- I was able to train an Llama2 model (1 epoch only)
- Llama.cpp depricated some functionality which made …
-
Will the code for fine-tuning the models be released?
Thank you for your excellent work.
-
Hey all,
I'm trying to fine-tune MiniLM on the fiqa dataset using GPL. Basically I've pretty much followed the instructions in the GPL Readme. I've changed my learning rate to 2e-6 (from the default …
-
Hi, I'm trying to fine-tune the Llama3.1 8b model but after fine-tuning it uploading it to HF, and when trying to run it using vLLM I get this error "KeyError: 'base_model.model.model.layers.0.mlp.dow…