Open zhangdahua1 opened 2 years ago
Dear @zhangdahua1 , Thank you for asking..
main.py->config['offline_split']/config['online_split']
.This is to run experiments multiple times, each with a different random seed. Many RL algorithms' performance is sensitive to the random initialization of neural network, exploratory actions etc. In order to assess the consistency and variability of the performance it is almost always desirable to run experiment with random seed controlled and other setup fixed...
Please let me know if that answered your questions..
Can I understand that offline is a training phase. And online is a testing phase
Yes, kind of like that .. But unlike supervised learning, training RL on offline data alone cannot always produce good results.. Therefore continued learning into the 'testing' (online) phase is needed ..
This is especially true for this particular dataset. The VVC device status (action) are generated by the legacy control approach. As a result, the "actions" are highly correlated with the "states". This makes offline training very difficult..
I understand your answer and thank you for your answer. If I want to view the actions of ltc tap and cap status, can I view them through one parameter in your program . And can I check the voltage change of one node?
Sorry, the code is not very well written. It did provide an convenient API.. The voltage can be checked by the info['v']
field and the ltc tap, cap status by info['action']
field. The ltc taps are placed first. The info
is a returned parameter of env.step()
.
envs/env.py
:
It's my problem that I don't have a thorough understanding of your program. My question has been answered.Thank you very much for your reply.
Hello, dear @yg-smile . I have two new questions.
Thank you for your questions.
envs/env.py/def step
line 163 onward to see how to get all data from power flows..
Sorry, I have two new problems through learning your program.