Temperature 22 degrees, pressure 1000 mb. Rain.
Temperature 15 degrees, pressure 990 mb. No rain.
Then the question is:
Temperature 17 degrees, pressure 1010 mb. Will there be rain?
We predict by Hamming distance to the nearest exemplar, (a) with pressure in millibars, (b) with pressure in bars. What does it predict?
wjk
yj
An example showing that this is this a bad measure of the error is:![]()
y = a x + bDefine a perceptron that receives x and y as input and fires if and only if the point is on this side of the line:
y ≤ a x + b
| Question 1 | [Total marks: 20] |
| 1(a) | A typical machine learning algorithm converges to a solution in a fairly short time.
Usually within minutes or hours.
Discuss what convergence means,
and whether human learning converges in a similar way.
Maximum length: 2 sentences |
[5 marks] |
| 1(b) | I wrote a chatbot that (arguably) passed the Turing Test on the Internet in 1989.
Have I proved the Turing Test is nonsense? Maximum length: 2 sentences |
[5 marks] |
| 1(c) | "When trying to make intelligent machines, AI should carefully try to copy how nature made intelligent creatures."
Suggest why this might be a good idea. Maximum length: 2 sentences |
[5 marks] |
| 1(d) | "When trying to make intelligent machines, AI should carefully try to copy how nature made intelligent creatures."
Suggest why this might be a bad idea. Maximum length: 2 sentences |
[5 marks] |
| Question 2 | [Total marks: 20] |
| 2(a) |
Is this function y = f(x) easy or hard to maximise from exemplars? f(a) = 1 for some value a. f(x) = 0 for all other x. Maximum length: 3 sentences |
[5 marks] |
| 2(b) |
Consider trying to maximise a function
of x, where x is an integer from 1 to 1020.
There are seven (and only seven) values of x where f(x) = 10.
For all other x, f(x) = 1.
The machine is not told this but has to find out by trial and error.
What would maximising this function by trial and error look like? Maximum length: 3 sentences |
[5 marks] |
| 2(c) |
Consider a heuristic search as follows:
Initialise with 100 random solutions.
Repeat for many iterations:
{
Test the 100 solutions.
Pick the best n.
Make small changes in them
to generate 100 new solutions.
}
If n=99 what will this look like?
Maximum length: 2 sentences |
[5 marks] |
| 2(d) |
Consider a heuristic search as follows:
Initialise with 100 random solutions.
Repeat for many iterations:
{
Test the 100 solutions.
Pick the best n.
Make small changes in them
to generate 100 new solutions.
}
If n=1 what will this look like?
Maximum length: 2 sentences |
[5 marks] |
| Question 3 | [Total marks: 20] |
| 3(a) | Imagine a search tree where every node has an infinite number of children. Can search still proceed? Maximum length: 3 sentences |
[5 marks] |
| 3(b) |
Imagine the following situation in Best-first search.
The heuristic evaluation function tells us a state is 4 steps away from the goal.
We examine the children of this state and the heuristic evaluation function
tells us all of them are 5 steps away from the goal.
Can this happen?
Maximum length: 2 sentences |
[5 marks] |
| 3(c) |
In the Travelling Salesman Problem with n cities,
we say there is an upper bound of (n-1)! on the number of paths to be searched.
What assumption did we just make?
Maximum length: 2 sentences |
[5 marks] |
| 3(d) | The idea of "informed search" is that we choose the next state that is most likely
to lead to the goal. But if we can do that, then the problem is solved!
Just keep choosing the best next state after that and you get to the goal ASAP. No search needed! No AI needed. Explain.
Maximum length: 2 sentences |
[5 marks] |
| Question 4 | [Total marks: 20] |
| 4(a) |
Let XOR in n dimensions be defined as follows.
If all inputs are 0, output is 0.
If precisely one input is 1, and all others are 0, then output is 1.
If more than one input is 1, output is 0.
Show that a perceptron cannot implement this definition of XOR in n dimensions. Maximum length: 5 sentences |
[5 marks] |
| 4(b) |
The following is the perceptron learning rule.
We have multi-dimensional input,
where each component of the input Ii is binary 0 or 1.
It produces output y = 0 or 1.
We show it the "correct" output O = 0 or 1.
We modify the weights and the threshold by the
rules:
wi := wi + C (O-y) Ii
t := t - C (O-y)
where C is some positive constant.
|
[5 marks] |
| 4(c) |
A multi-layer neural network has
1 input unit, n hidden units, and 1 output unit.
The weights on the input layer are all the same:
wij = W1.
The thresholds of the hidden units are all the same:
tj = T.
The weights on the output layer are all the same:
wjk = W2.
Show that this network is equivalent to a network with 1 input unit, 1 hidden unit, and 1 output unit. What are the weights and thresholds of the latter network? Maximum length: 5 sentences |
[5 marks] |
| 4(d) |
Derive the set of inequalities that define a
Perceptron for OR (2 inputs).
Maximum length: 4 sentences |
[5 marks] |
| Question 5 | [Total marks: 20] |
| 5(a) | In Reinforcement Learning, is a punishment of -1000 a big punishment?
Maximum length: 2 sentences |
[5 marks] |
| 5(b) | In a Markov Decision Process (MDP),
what is wrong with the following:
If we take action a in state x,
then Pxa(y)=0.6
for some state y,
and Pxa(w)=0
for all other states w.
Maximum length: 1 sentence |
[5 marks] |
| 5(c) | In a Markov Decision Process (MDP), explain why E(r) exists
and E(y) does not exist. Maximum length: 4 sentences |
[5 marks] |
| 5(d) | In state x, if we take action a, we sometimes get reward 0,
and sometimes get reward 5.
If we take action b, we always get reward 3.
Under what circumstances is
the expected reward for taking action a the same as
the expected reward for taking action b? Maximum length: 3 sentences |
[5 marks] |
Seat No:_____________________
|