Decision Making and Reinforcement Learning
About this Course
This course is an introduction to sequential decision making and reinforcement learning. We start with a discussion of utility theory to learn how preferences can be represented and modeled for decision making. We first model simple decision problems as multi-armed bandit problems in and discuss several approaches to evaluate feedback. We will then model decision problems as finite Markov decision processes (MDPs), and discuss their solutions via dynamic programming algorithms. We touch on the notion of partial observability in real problems, modeled by POMDPs and then solved by online planning methods. Finally, we introduce the reinforcement learning problem and discuss two paradigms: Monte Carlo methods and temporal difference learning. We conclude the course by noting how the two paradigms lie on a spectrum of n-step temporal difference methods. An emphasis on algorithms and examples will be a key part of this course.Created by: Columbia University

Related Online Courses
This specialization is intended for aspiring DevOps professionals and IT enthusiasts. Through 8 comprehensive courses, you will cover essential topics including DevOps prerequisites, Git, Jenkins,... more
This is a self-paced lab that takes place in the Google Cloud console. In this lab, you will learn how to use grounding in Vertex AI to generate content grounded in your own documents and... more
Yunnan, abbreviated as Yun and called Dian for short, is a multi-ethnic province which has historically been situated in the southwest frontier of China. Over time, the ethnic minorities evolved... more
Knowledge of soft skills or people Skills are increasingly coming into focus and being recognized as critical for success - hence at IBM we call them \"Success Skills\", and categorize them as... more
Code and run your first Java program in minutes without installing anything! This course is designed for learners with limited coding experience, providing a solid foundation of not just Java, but... more