Closed HansBambel closed 5 years ago
@adrigrillo, when restoring training with DDQN, are the following variables enough?
current_model
target_model
optimizer
memory
Can I resume target_model
to the same as current_model
(from the latest check point)?With the target model should be enough. However, you have to copy the target in the current model during the process of loading a pretained model.
With regards to the optimizer and the memory is not necessary, better not to save them.
On Wed, 9 Jan 2019, 16:52 dannigt <notifications@github.com wrote:
@adrigrillo https://github.com/adrigrillo, when restoring training with DDQN, are the following variables enough?
- current_model
- target_model
- optimizer
- memory Can I resume target_model to the same as current_model (from the latest check point)?
— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/weidler/RLaSpa/issues/18#issuecomment-452745987, or mute the thread https://github.com/notifications/unsubscribe-auth/ARWvRtuJbev6vXINwfTvpow_aHv4Whbaks5vBhA9gaJpZM4Z1wjI .
With the target model should be enough. However, you have to copy the target in the current model during the process of loading a pretained model. With regards to the optimizer and the memory is not necessary, better not to save them. … On Wed, 9 Jan 2019, 16:52 dannigt @.*** wrote: @adrigrillo https://github.com/adrigrillo, when restoring training with DDQN, are the following variables enough? - current_model - target_model - optimizer - memory Can I resume target_model to the same as current_model (from the latest check point)? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub <#18 (comment)>, or mute the thread https://github.com/notifications/unsubscribe-auth/ARWvRtuJbev6vXINwfTvpow_aHv4Whbaks5vBhA9gaJpZM4Z1wjI .
Alright, thx! I thought ADAM adjusts the learning rate depending on the current epoch/episode. Is it the case here?
TODO's
Training takes a long time. If it crashes intermediately we should be able to continue where it crashed without restarting the whole training.