hawkw / cs383s2015-lab4

Lab 4 for Computer Science 383: Rebellion model in Repast Simphony
MIT License
0 stars 1 forks source link

Summarize experience #7

Closed hawkw closed 9 years ago

hawkw commented 9 years ago

Comment on

hawkw commented 9 years ago

This is still needed for today. I guess I'll do it since I can't get in touch with SJ.

sjguillaume commented 9 years ago

Hey, I am heading to the lab in just a minute!

On Mar 3, 2015, at 1:37 PM, Hawk Weisman notifications@github.com wrote:

This is still needed for today. I guess I'll do it since I can't get in touch with SJ.

— Reply to this email directly or view it on GitHub.

hawkw commented 9 years ago

Okay, if you want to write responses to this question, I can add them to the notebook when you're done!

sjguillaume commented 9 years ago

Takeaways from the assignment are how to change an agent's learning strategy to improve performance over time. To implement this, the agents were given a reward when they were active and not arrested, and if they were caught they were given a negative reward. The agent then uses this reinforcement to decide in the future what action to take based on their memory.

Challenges encountered over the course of this lab were mainly in creating the rebellion strategy for the population. The learning strategy was difficult to implement because the agents appeared to have learned not to rebel for a short period of time, however this learning did not become a trend. This meant to us that the agents were not learning the strategy because their performance was remaining constant within a range.