google-research / tuning_playbook

A playbook for systematically maximizing the performance of deep learning models.
Other
26.29k stars 2.18k forks source link

Request to cover "Insensitivity to tuning" in the playbook #72

Open 21kc-caracol opened 1 month ago

21kc-caracol commented 1 month ago

I've followed the playbook's guidelines and done two "studies", but the results i've received aren't covered in the playbook:

  1. Batch tuning (16 to 512, intervals of power of 2)
  2. Learning rate + beta1 tuning on Adam (9 different combinations with 3 different scales of 10, and 3 different beta1 values for each- 0.8, 0.9, 0.95)

Both studies didn't show a significant influence on both training&validation curves.

From a short Q&A with "ChatGPT and reddit" it might be due to inappropriate combination of model complexity and data (assuming my data is cleaned).

I would greatly appreciate if you could even in several sentences, add to the playbook your 2 cents about your next "play" when the tuning doesn't influence much on the curves and the results.