-
## 論文リンク
https://arxiv.org/abs/2103.00020
## 公開日(yyyy/mm/dd)
2021/01/05
## 概要
OpenAI が発表した DALL·E の中で reranking にも使われていた CLIP (Contrastive Language-Image Pre-training) の論文。
Web 上のテキストから特別な a…
-
Thank you very much for your work, could you please share your pre-training data set, I would like to use some longer max_length model to replace codebert to handle longer input data, in the meantime,…
-
Hello,
I hope this message finds you well. I have successfully executed the process_GDSCv2.py script as per the provided instructions. However, when I attempted to run the main.py function to train…
-
Hello, could you explain how the pre-trained model you loaded during the training of DA-Clip was obtained? Because I noticed that the model “laion2b_s34b_b79k” is used in your training command, but in…
-
will the training code be released?
-
### The model to consider.
https://huggingface.co/tencent/Tencent-Hunyuan-Large
Tencent released a 389B MoE with only 52B activated parameters which beats the Llama 3.1 405B.
There are three chec…
-
### General
- [x] Prepare scaling plots until end of february. Y-axis: the speedup we get when running one epoch through the model for 2,4,6,8,10 GPUs
- [x] Find out how many samples we have in the …
-
We would like to evaluate the model performance for various LLM fine tuning approaches and compare them with the standard benchmarks. An experiment we would like to try is:
- **Compare the full car…
-
Hi.
I want to predict additional 3D keypoints like hands in halpe136 or small toe and big toe in halpe26.
What more things do I need to do?
Should I do pre-training? Or just fine tuning?
I…
-
I want to know the exact splits of AudioSet or VggSound used to train the CLAP. Because many audio-related datasets for downstream tasks were collected from these two large-scale datasets, if all thei…