Closed ethanreidel closed 1 month ago
βββββββ6 filesβ Β±βββββββ0βββββββββ6 suitesβ Β±0βββ52m 7s :stopwatch: + 22m 2s 2β990 tests β-ββββββββ3ββ2β966 :heavy_check_mark: β-ββββββ15ββ23 :zzz: +11ββ1 :x: +1β 8β970 runsβ +5β941ββ8β898 :heavy_check_mark: +5β893ββ69 :zzz: +45ββ3 :x: +3β
For more details on these failures, see this check.
Results for commit d2ba5cb0.βΒ± Comparison against base commit 606c732a.
@skanjila @saad-palapa
Adds the ability to gradually unfreeze or thaw specific layers within a pre-trained model's architecture. Aims to mitigate catastrophic forgetting/improve transfer learning capabilities. Currently works for ECD architecture.
User passes in two things: thaw_epochs (list of integers) and layers_to_thaw (2D array of layer strings)
thaw_epochs: -1 -2 layers_to_thaw:
TODO/potential issues:
test: [tests/ludwig/modules/test_gradual_unfreezing.py]
Any and all feedback is greatly appreciated. π