-
I want to know, are you pretraining the teacher on ImageNet ?
In the paper, they mentioned that Teacher is pretrained on ImageNet. Is your repo following it ?
-
### Describe the bug
We pretrain large models with [fairseq](https://github.com/pytorch/fairseq) and log progress with wandb. During the run wandb stops logging and the run is shown as crashed (eve…
-
It would be best to create a branche from the development branch and work in it on fixing the problem
-
Hello!
I am trying to pretrain an adapter using the `4_pretrain_adapter.sh` script.
I have a GeForce RTX 2080 SUPER installed (~8GB VRAM), with NVIDIA Driver Version: 440.33.01, CUDA Version: 10.…
ghost updated
3 years ago
-
## タイトル: FACMIC: 医療画像分類のための連合適応CLIPモデル
## リンク: https://arxiv.org/abs/2410.14707
## 概要:
分散型データを用いた深層学習モデルの学習を可能にし、データプライバシーを確保する手法として、Federated learning (FL) が注目されています。しかし、FLでは、モデル性能の評価において通信コストが重要…
-
Hello,
Thanks for the super interesting paper. I actually came across your poster in ACL and after reading the whole paper, I have a few questions regarding experimental details:
1. During pretr…
-
Hi, could you please tell me that how the DeepRT+ do the calibration using a certain ratio of the test-group peptides after pretraining with the big data?
Thanks.
-
There seems to be an issue with the Gesture Dataset, having repetitive channels. The below shows the training dataset.
-
Hi
I am repeatedly facing the error that a parameter of random checkpoint does not exist.
This happens whenever I pretrained the model from scratch.
Whenever I run the code, the iterations (epochs)…
-
Hi @danielhanchen
I tried fine-tuning llama 3.2-1b base model for 2 of my tasks following below example notebook
https://colab.research.google.com/drive/1tEd1FrOXWMnCU9UIvdYhs61tkxdMuKZu?usp=sha…