Backpropagation of Error
   - The MLP is cool, but how do you set the weights?
 
   - No one sets the weights manually.
 
   - With a single perceptron, you can use the perceptron learning rule
     (Rosenblatt from the 50s and 60s).
 
   - What's really cool is that an MLP can be learned by
       the 
       
	 backpropagation of error.
 
   - You train the system in a supervised manner.
 
   - That is, you have instances that you know the output of.
 
   - You put those instances through the system, and record the 
       error.
 
   - You then move the weights in a direction that will reduce the error.
 
   - You backpropagate the error.
 
   - So if the desired value is higher than the actual value, and the
       inputs are positive, increase the weights. 
 
   - The actual maths behind this is complex, and there are a few variants
       of the rule. 
 
   - You then go through another training cycle, and repeat.
 
   - Eventually, if you have enough neurons, this will converge on a 
       weight matrix that closely approximates the input output pairs.
 
   - It's a universal learner.