代写COMP3211 Assignment #1
- 首页 >> Python编程COMP3211 2023-24 Fall Semester Assignment #1
Date assigned: Wednesday, Feb 14, 2024
Due time: 23:59pm on Sunday, Feb 25, 2024
How to submit it: Submit your written answers as a pdf file on canvas.ust.hk. Submit your code for the last three programming questions as a zip file named YourStudentID.zip
Penalties on late papers: 20% off each day (anytime after the due time is considered late by one day)
Problem 1. (10%) Consider a 10x10 grid without any obstacles, and a robot with the same specification as our boundary following robot: eight sensors and four actions. Design a reactive production system to control the robot to go to one of the four corners, wherever its initial position is. Write a production system just for this task without calling the boundary following production system in the lecture note.
Problem 2. (10%) Which boolean function does the following TLU implement? The TLU has five inputs. Its weight vector is (1.1, 3.1, −1, −2, 0.5), and the threshold is 1. Give your answer as a boolean expression using the connectives “x + y” (x or y), “xy” (x and y), and “x” (not x).
Problem 3. (Programming) (30%) Design and implement a genetic programming system to evolve some perceptrons that match well with a given training set. A training set is a collection of tuples of the form (x1, ..., xn, l), where xi ’s are real numbers and l is either 1 (positive example) or 0 (negative example). So for your genetic programming system, a “program” is just a tuple (w1, ..., wn, θ) of numbers (weights and the threshold). Answer the following questions:
1. What’s your fitness function?
2. What’s your crossover operator?
3. What’s your copy operator?
4. What’s your mutation operator, if you use any?
5. What’s the size of the initial generation, and how are programs generated?
6. When do you stop the evolution? Evolve it up to a fixed iteration, when it satisfies a condition on the fitness function, or a combination of the two?
7. What’s the output of your system for the provided training set gp-training-set.csv in assign1-data27.zip (for Python 2.7) or assign1-data3.zip (for Python 3)?
The Pacman Game
This is a simple exercise to implement our boundary following production systems in the Pacman game from UC Berkeley. The goal of the Pacman game is simple: control the Pacman (yellow circle) to eat all the foods (white dots) in the map, as shown in Figure 1.
Figure 1: The smallMap map
Environment Setup
Download assign1-data27.zip (for Python 2.7) or assign1-data3.zip (for Python 3) on Canvas and unzip it. Open your terminal, navigate to the pacman folder and run python pacman.py. You will see the graphical interface of the Pacman game pops out, and you can play the game with your keyboard. If the game works normally, you can proceed to the next step.
Perception and Movement
Now let’s learn how to design a Pacman agent. We have implemented a NaiveAgent for your reference, which is an agent that goes west until it can’t. Run python pacman.py --layout testMap --pacman NaiveAgent to how the NaiveAgent acts. Note that the --layout parameter specifies the map, and the --pacman parameter specifies the agent. You will see the NaiveAgent goes from east to west in the testMap map (Figure 2), eats a food, and wins the game.
The Pacman has eight sensors in eight directions: northwest, north, northeast, east, south-east, south, southwest, and west. The sensor will return 1 if there is a wall in its direction, and 0 otherwise. You can use the getPacmanSensor function of the GameState class to get the list of return values of the sensors in the order described above. For example, if the getPacmanSensor function returns (s1, s2, ...s8) =[0, 0, 1, 1, 1, 0, 0, 0], it means that there are three walls in the northeast, east and southeast of the Pacman.
Figure 2: The testMap map
The movement of an agent is controlled by the getAction function, which is supposed to return one of the five actions at each step: NORTH, SOUTH, EAST, WEST and STOP. Open reactiveAgents.py to see the implementation of the NaiveAgent: it always returns WEST until it sensed that there is a wall on its west.
Below the NaiveAgent class in reactiveAgents.py, you will see the ECAgent class, which has also been implemented for your reference. The PSAgent is an agent that moves accord-ing to the production system we discussed in the lecture. You may run the agent in the smallMap map for the expected behaviour of the boundary following task.
Problem 4. (Programming) (25%)
In this problem, you are required to implement ECAgent, which is a boundary-following agent that uses the Error-Correction procedure to learn the action functions for the task.
Task
We have prepared four training sets for learning when to move north, east, south, and west, respectively. Each vector in the training set is in the form of (s1, ..., s8, d), where the first 8 elements are the Pacman’s sensor readings, and d is the label of this input, with d = 1 meaning a positive, and 0 a negative example.
You are required to learn a perceptron for each of the actions using the Error-Correction procedure. You can use any programming language for training the perceptrons. You should initialize their weights to all 0 and use learning rate = 1.
1. Give the boolean expression that corresponds to your perceptron for the move north action. (5%)
2. Use your learned perceptrons to implement the ECAgent in reactiveAgent.py in this way: if exactly one of them outputs 1, then do the respective action; if more than one of them output 1, then use the following order: prefer going north, followed by east, south, and west; and finally if none of them outputs 1, then go north. (20%)
You can again find the training sets named north.csv, east.csv, south.csv, and west.csv in the data files.
Marking Scheme
We will test your agent with four maps (5 points each). We say that your agent passes a test case if it wins the game, i.e. has eaten all the foods in the map. We will place foods along the boundary of each map, so if your agent is able to follow the boundary, it should also be able to eat all the foods and win the game.
Notes
• If you have no experience with UNIX or Python before, please go through this tutorial first: http://ai.berkeley.edu/tutorial.html, which uses Python 2.7. You should be able to apply it to Python 3 as well.
• The program may also work in Windows but you will need to handle some dependency issues.
• There will be no tight spaces (see lecture slides for definition) in the test cases. (note that the testMap map is an instance of tight spaces)
• The order that your agent visits the foods does not matter.
• If your agent made an invalid move (e.g. move towards the wall), the program will throw an exception and exit. Be careful!
Problem 5. (Programming) (25%)
This problem also requires you to design a boundary-following agent SMAgent with produc-tion system. However, SMAgent is a state machine with impaired sensors.
Task
The SMAgent is sensory-impaired and can only received only 4 sensor inputs from the 4 direc-tions: north, east, south, west. The agent receives sensor values from the getPacmanImpairedSensor function. The sensory input (s2, s4, s6, s8) corresponds to the sensor values from the north, east, south, and west respectively.
The SMAgent can remember the previous action prevAction and the previous sensory input prevSense. The initial state of the SMAgent has been implemented for you.
To finish this task, you do not need any additional data member other than prevAction and prevSense. You are required to use getPacmanImpairedSensor instead of getPacmanSensor and we will disable the getPacmanSensor when we test the SMAgent.
1. Complete the getAction function of the SMAgent. We will test your agent with four maps like the Problem 4.(20%)
2. Denote ( ˜s2, s˜4, s˜6, s˜8), to be the previous sensory input and (N, E, S, W) to be the one-hot encoding of the the previous action. For example, if the previous action is moving north, then (N, E, S, W)=[1,0,0,0]. Can we train perceptrons on the feature vector (s2, s4, s6, s8, s˜2, s˜4, s˜6, s˜8, N, E, S, W, d) with the Error-Correction procedure as in the Problem 4? (5%)