-
May I ask for the hyperparameters used for LLaMA finetuning? The learning rate, batch size, EWC coefficient (λ), and the rank and coefficient of LoRA will be helpful.
Thank you!
-
Hello, this research is very meaningful. However, when I reproduced the performance of the default code (with default parameters) in CIFAR80N, I found that the performance of Sym-80% (33.72) and Asym-…
-
I am trying to reproduce results for SplitCIFAR-100, I used the exact same hyperparameters mentioned in the code. I am getting average accuracy as 59.42, and forgetting as 6.41.
In the paper,average …
-
Dear maintainers,
I have couple questions according to hyperparameters.
1) What are the exact hyperparameters to run llama.py that are reported in Table2 for GPTVQ 2D for Llama2-7b? I am strugg…
-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and fou…
-
-
Hi, I just found out that adjusted learning rate, batch size significantly (increase/decrease) changes the results.
Although all of the compared methods are using same style of training, I feel the …
-
Linking this to task #77 for @5u6r054 to test at scale.
Since we currently don't know if the provided implementation can support more than 1 node, we want to scale testing of the implementation.
…
-
Hi @chaitjo, thank you for this awesome work!
Could you kindly share the checkpoints and the hyperparameters for the NAR models reported in Table-1 of your paper (image attached)? I tried to reprod…
-
Some configspaces created with v0.6.1 can't be loaded with v1.1.3, which seems to be due to the backwards compatibility beeing missing for categorical / ordinal / constant hyperparameters in the dicti…