-
May I ask how to finetune trocr on my own dataset? What's the format of dataset I need to prepare?
-
I have a few different pretraining tasks that can be incorporated with MLM loss objective. We can run parallel jobs for pretraining using different loss functions. The objective is to gain understandi…
-
Dear all,
I have fine-tuned TinyLLaVA-Phi-2-SigLIP-3.1B for 1 epoch and then continued the fine-tuning for another epoch starting from trained models saved after the first epoch. Both training runs w…
-
Hi,
Thanks for providing and presenting this nice work.
As mentioned in your paper, your attention pattern for modeling long sequences can be plugged into any pretrained transformer model.
I wond…
-
## Issue Description
Hello GitHub community,
I am currently seeking guidance on how to effectively evaluate the MADLAD 400 model, a 7.2B parameter machine translation (MT) model that has been fi…
-
### Model introduction
The CodeQwen1.5-7B, a specialized codeLLM built upon the Qwen1.5 language model. CodeQwen1.5-7B has been pretrained with around 3 trillion tokens of code-related data. It suppo…
-
Hello,
So I finetuned voicecraft on the french common voice-french dataset. It's quite exciting since it's my first time working on LLM and on full audio model (not just spectrogram -> classificat…
-
### Describe the issue
Issue/Error:
Loading 1.5 models works fine, but loading 1.6 models yield the error below. Note that the 1.6 models do load (despite the error) and inference works. However, tr…
-
Suggested list of courses would be:
- An introduction to deep learning **
- How to train a neural network
- Regularisation in neural networks
- Deep Bayesian neural networks
- Conv…
-
# HPT - Open Multimodal Large Language Models
[https://github.com/HyperGAI/HPT](https://github.com/HyperGAI/HPT)
[https://huggingface.co/HyperGAI/HPT](https://huggingface.co/HyperGAI/HPT)
[techni…