Armed Bandits
   - In AI and decision theory, armed bandits are gambling devices.
       They're like slot machines.
 
   - What is typically done with a one armed bandit is to show how likely
       you are to get a positive return, with the rest of the options being
       0.
 
   - For instance, you have 10% chance of winning £100 and 90% chance
       of losing your bet.
 
   - Often people are asked what they'd bet to play this bandit.
 
   - The key is that the experimentor can set the probabilities, or
       arrange payouts explicitly.
 
   - You can use this to show how people don't choose rationally (Kahnemann)
 
   - There is also a lot of work with multi armed bandits (K or N).
 
   - Part of the question here is, if you can play over and over again,
       how do you discover what to play.
 
   - If you pick the right bandit, you can maximise your expected outcome.
    
 
   - This involves the exploration exploitation dilemma.
 
   - Moreover, the probabilities can change over time.