Closed GoogleCodeExporter closed 9 years ago
Thanks for the issue. A couple of comments below.
> 1) I discussed this with Risto at class last night; we were wondering if
weighting the fitness with the amount of time the robots spend near the goal
instead of simply how close their final position is might incentivize them to
follow a more direct path.
Their reward is currently based on cumulative distance from flag over the
course of their lifetime. Thus, agents that, on average, spend more time closer
to the goal should get higher raw fitness value.
> 2) It also seems suspicious that successful agents are the ones who wind up
circling the target. If they have learned that circling is awesome behavior,
wouldn't that influence their paths all the time? Risto said that one thing the
torque code does is "cheat" the agents to always face enemies that are in their
field of view. It seems like this kind of tweak might prevent unnecessary
circling.
In Torque-NERO this was an unintentional side effect of the auto-aiming feature
- even if the agent is signaling a turn, having an enemy in its sights will
cause it to face the enemy as long as its firing output is on. In effect this
makes it easier to approach enemies (sometimes too easy, as it is hard to teach
Torque-NERO agents to NOT approach the enemy).
Original comment by ikarpov
on 2 Nov 2011 at 3:11
scaling issues addressed on trunk revision 1365 should at least partially
address this issue.
Original comment by ikarpov
on 11 Nov 2011 at 9:42
Original issue reported on code.google.com by
jroes...@gmail.com
on 2 Nov 2011 at 11:16