Open AmosMichael opened 1 year ago
@AmosMichael Sorry for the late reply.
An optimal policy has the property that whatever the initial state and initial decision are, the remaining decisions must constitute an optimal policy with regard to the state resulting from the first decision.
In practice, although the environment and target state will change during each planning cycle, generating plan_state on the previous "optimal" trajectory brings better consistency.
Another reason is that this method decouples the planning module and control module, we do not want the tracking error of the control module influence the planning part. In practice, we also need to monitor the displacement between the plan_state and the real ego state (from the positioning systems). If the tracking error reaches a pre-defined threshold, the plan_state will be reset to the current ego state.
Hi, I'm reading yours great
Epsilon
paper but I don't get enough background knowdege to fully understand this paper. Some content makes me confused. Hope to get some feedbacks.Is it proper to replan using state on
executing_traj_
at current plan cyccle's terminal timestamp`?Code in the link is as following:
It uses state on
executing_traj_
at current plan cyccle's terminal timestamp to replan? I think it's better to use ego's real state instead of state onexecuting_traj_
. Right?Is it necessay to run
ssc
traj planner after eudm? Decision part (eudm
) already gives ego's forward trajectories in closed loop simulations. Is it necessay to runssc
traj planner after eudm?As far as I know, trajectories given by
edum
already considers ego's kinematics constraints (by usingidm
andmobil
model to generate latitude and longitude behavior, then pure persuit control model giving proper control commands) and surrounding obstacles (by collision checking).