Machine learning (ML) is being deployed in increasingly consequential tasks, such as healthcare and autonomous driving. For the foreseeable future, successfully deploying ML in such settings will require close collaboration and integration with humans, whether they be users, designers, engineers, policy-makers, etc. This course will look at how humans can be incorporated into the foundations of ML in a principled way. The course will be broken down (unequally) into three parts: demonstration, collaboration, and oversight. Demonstration is about how machines can learn from 'observing' humans---such as learning to drive a car from data collected while humans drive. In this setting, the human is assumed to be strictly better than the machine and so the primary goal is to transmit the human's knowledge and abilities into the ML model. The second part, collaboration, is about when humans and models are near equals in performance but not in abilities. A relevant setting is AI-assisted healthcare: perhaps a human radiologist and ML model are good at diagnosising different kinds of diseases. Thus we will look at methodologies that allow machines to `ask for help' when they are either unconfident in their own performance and/or think the human can better carry out the task. Lastly, the course will close with the setting in which machines are strictly better at a task than humans are, but we still wish to monitor them to ensure safety and alignment with our goals (oversight).
Dates | Topics | Readings |
---|---|---|
31.10.23 | Logistics, introduction, building prior knowledge into models | |
02.11.23 | Crowdsourcing, Dawid-Skene model | |
07.11.23 | Active learning | |
09.11.23 | Practical on crowdsourcing platforms, overview of reinforcement learning (RL) | |
14.11.23 | Imitation learning: behavior cloning, distribution matching | |
16.11.23 | Imitation learning continued: distribution matching (cont.), inverse RL | |
21.11.23 | Inverse RL continued, RL from human feedback | |
23.11.23 | Direct preference optimization, exam review | |
30.11.23 | ||
05.12.23 | Guest lectures: large language models by Prof. Charlie Clarke (Univ. of Waterloo) and the alignment problem by Leon Lang (UvA) | |
07.12.23 | Combining human and machine predictions, learning to defer to an expert |
|
12.12.23 | The off-switch game (slides) | |
14.12.23 | Off-switch game continued, human control (causal definitions), course summary |