-
Hello, I am trying to train DINO with a base ViT from scratch and I have some doubts.
First of all, I think that in the original paper, the student temperature is 0.1 in the 30 epoch warmup, but I am…
-
python qlora.py \
--model_name_or_path /models/guanaco-33b-merged \
--output_dir ./output \
--dataset alpaca \
--do_train True \
--do_eval True \
--do_mmlu_eval True \
…
-
[Predictive-Maintenance-using-LSTM-master.zip](https://github.com/Saima-786/INDUSTRIAL-PROJECT-upGrad-/files/15199759/Predictive-Maintenance-using-LSTM-master.zip)
-
### Deep Learning Simplified Repository (Proposing new issue)
:red_circle: **Project Title** : Classification of Elon Musk Tweets using NLP
:red_circle: **Aim** : Create a classification model using…
-
Hey, thanks for your great work..
There are a few clarifications I need as I am facing a bit difficulty in replicating the results, it would be very kind if you can help:
1. In the implementation …
-
According to this blogpost: http://www.fast.ai/2018/07/02/adam-weight-decay/ and mentioned article https://arxiv.org/abs/1711.05101, Adam has problems when used with L2 regularization. If i understand…
-
The param_group's lr's cannot be trusted if the optimizer state is not restored (and this can be okay, because optimizer buffers can double the checkpoint size).
In this line they are trusted if last…
-
Hi @lessw2020, thanks for the very nice work!
I noticed that in this Ranger21, the optimizer is tightly coupled with the lr scheduler, could you guide me how I can decouple them?
hiyyg updated
2 years ago
-
Interesting Resources:
- [RL Curriculum Learning](https://lilianweng.github.io/lil-log/2020/01/29/curriculum-for-reinforcement-learning.html)
- [meta-RL](https://lilianweng.github.io/lil-log/2019/…
-
[링크](https://dacon.io/competitions/official/235554/codeshare/651?page=1&dtype=recent&ptype=pub)