nikhilbarhate99 / PPO-PyTorch

Minimal implementation of clipped objective Proximal Policy Optimization (PPO) in PyTorch
MIT License
1.57k stars 332 forks source link

ValueError: expected sequence of length 8 at dim 1 (got 0) #64

Open kavinwkp opened 8 months ago

kavinwkp commented 8 months ago

Two lines should be changed in train.py state, = env.reset() # at 170 state, reward,done, , _ = env.step(action) # at 177

yeedonghyun commented 7 months ago

Problem is solved. thx.