-
# Description
Major recent breakthroughs in generalist, transferable learning were executed by training and using large-scale language, vision or language-vision foundation models like GPT, ViT, CL…
-
Keras.io example: https://keras.io/examples/nlp/data_parallel_training_with_keras_nlp/
Merged PR: https://github.com/keras-team/keras-io/pull/1395
This example is good on the whole but it would be…
-
### Question
Hi @xiuqhou ,I wanted to train Relation_detr on my custom coco_dataset but getting error.
![image](https://github.com/user-attachments/assets/48bd084b-53a6-4cc2-9ead-cff33c8a929e)
how …
-
### Describe the issue
Issue:
I am finetuning LLaVA 1.5 13b using `scripts/v1.5/finetune_task_lora.sh` on my custom dataset. Training process looks normal (~0.4) until a iteration (randomly, no …
-
# Video Representation Learning and Latent Concept Mining for Large-scale Multi-label Video Classification #
- Author: Po-Yao Huang, Ye Yuan, Zhenzhong Lan, Lu Jiang, Alexander G. Hauptmann
- Orig…
-
- [ ] [The Scaling Hypothesis · Gwern.net](https://gwern.net/scaling-hypothesis)
# The Scaling Hypothesis · Gwern.net
**DESCRIPTION:** "GPT-3, announced by OpenAI in May 2020, is the largest neura…
-
Hey
First thank you for all your work :)
tldr: This is less of an issue, just more of a question: Is it normal that when using many different metrics, that they all seem to have different opinio…
-
我有一个数据集含一万余条问答数据集(以维基百科为背景知识库来构建的),想使用您的工作来做做RAG,首先一个问题通过检索维基百科,反馈相关的段落内容,然后提供给不同的大模型(我要做的实验大模型有chatglm llama glm 文心),您的工作可以实现吗?如何来处理呢?期待您的回复,谢谢啦!
-
https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5
We introduce InternVL 1.5, an open-source multimodal large language model (MLLM) to bridge the capability gap between open-source and proprietary…
-
- [x] [Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour](https://arxiv.org/abs/1706.02677) by **Facebook** 👉 [mynote](https://github.com/towerjoo/myML/blob/master/papers/weekly/train_ima…