User Tools

Site Tools


public:t-622-arti-09-1:lab_2_materials

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
public:t-622-arti-09-1:lab_2_materials [2009/01/26 13:50] hannespublic:t-622-arti-09-1:lab_2_materials [2024/04/29 13:33] (current) – external edit 127.0.0.1
Line 1: Line 1:
-=== Lab 2 materials ===+=== Lab 2 materials: Comparing Agent Programs ===
  
-In lab 2 we will finish [[public:t-622-arti-09-1:lab_1_materials|lab 1]] and write a stateful agent program.+In lab 2 we will finish [[public:t-622-arti-09-1:lab_1_materials|lab 1]] and write a new kind of agent program that maintains a state.
  
-Start by getting the file [[http://gist.github.com/52802|agents2.lisp]]. This is largely the same file as we had for the last lab session, except some fixes to accommodate stateful agents, and code to run experiments. +  - Start by getting the file [[http://gist.github.com/52802|agents2.lisp]]. This is largely the same file as we had for the last lab session, except some fixes to accommodate stateful agents, and code to run experiments. 
- +  First finish the steps in lab 1, including 
-First finish the steps in lab 1, including +    * writing an agent that walks about randomly 
-  * writing an agent that walks about randomly +    * writing a performance evaluation function that at the end gives 20 points for each clean square and deducts one point for each move made. You can get the information about made moves from the environment like this: <code lisp>(env-moves-so-far env)</code> 
-  * writing a performance evaluation function that at the end gives 20 points for each clean square and deducts one point for each move made. You can get the information about made moves from the environment like this: \\ <code lisp> +  After you have finished lab 1, fill in the function RANDOM-WALK-WITH-STATE, found in agents2.lisp. This should be the program of an agent that: 
-(env-moves-so-far env) +    * Uses the STATE field of the agent struct to remember what squares are already clean. You can use this field like this<code lisp>
-</code> +
- +
-After you have finished lab 1, fill in the function RANDOM-WALK-WITH-STATE, found in agents2.lisp. This should be the program of an agent that: +
-  * Uses the STATE field of the agent struct to remember what squares are already clean. You can use this field like this \\ <code lisp>+
 (agent-state agent)  ;; returns the current state (agent-state agent)  ;; returns the current state
 (setf (agent-state agent) <new value> ;; updates the state to a new value (setf (agent-state agent) <new value> ;; updates the state to a new value
 </code> </code>
-  * The agent should walk randomly, but it should avoid revisiting squares that are already clean. +    * The agent should walk randomly, but it should avoid revisiting squares that are already clean. 
-  * The agent's objective should be to clean the whole environment. +    * The agent's objective should be to clean the whole environment. 
-  * If the agent is sure everything is clean, it should return the action :idle +    * If the agent is sure everything is clean, it should return the action :idle 
- +    Be careful not to be to strict, e.g. if you make it absolutely impossible for the agent to travel through clean squares - can it get in trouble so that it is unable to meet its objective? If so, find a solution. 
-Be careful not to be to strict, e.g. if you make it absolutely impossible for the agent to travel through clean squares - can it get in trouble so that it is unable to meet its objective? If so, find a solution. +  Use the functions at the bottom of agents2.lisp to run experiments and evaluate your agent. To try your agent on all possible 3x3 environments with two dirty squares, enter this in the REPL: <code lisp>
- +
-Use the functions at the bottom of agents2.lisp to run experiments and evaluate your agent. To try your agent on all possible 3x3 environments with two dirty squares, enter this in the REPL: \\ <code lisp>+
 (run-experiment (generate-environments 3 3 2) (run-experiment (generate-environments 3 3 2)
          (lambda () (make-agent :program #'random-walk-with-state))          (lambda () (make-agent :program #'random-walk-with-state))
                 #'perf-measure                 #'perf-measure
                 1000)                 1000)
-</code> +</code>This will return the number of experiments performed, and the average performance value. 
-This will return the number of experiments performed, and the average performance value. +  Try different agents (e.g. stateless random walk) and different performance measures. Does the stateful agent behave better than the others with any particular performance measure? You can add print commands to the experiment functions to get more information about the agents behaviour if you like. 
- +  - **BONUS** If you finish the stateful agent and running the experiments, try creating an agent that moves through the environment in some orderly pattern. Make sure it visits all squares and cleans the dirty ones. Is the orderly agent more efficient than the random one?
-Try different agents and different performance measures. Does the stateful agent behave better than the others with any particular performance measure? You can add print commands to the experiment functions to get more information about the agents behaviour if you like. +
- +
-==Extra== +
-If you finish the stateful agent and running the experiments, try creating an agent that moves through the environment in some orderly pattern. Make sure it visits all squares and cleans the dirty ones. Is the orderly agent more efficient than the random one?+
  
  
/var/www/cadia.ru.is/wiki/data/attic/public/t-622-arti-09-1/lab_2_materials.1232977842.txt.gz · Last modified: 2024/04/29 13:32 (external edit)

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki