Decision Making and Reinforcement Learning
About this Course
This course is an introduction to sequential decision making and reinforcement learning. We start with a discussion of utility theory to learn how preferences can be represented and modeled for decision making. We first model simple decision problems as multi-armed bandit problems in and discuss several approaches to evaluate feedback. We will then model decision problems as finite Markov decision processes (MDPs), and discuss their solutions via dynamic programming algorithms. We touch on the notion of partial observability in real problems, modeled by POMDPs and then solved by online planning methods. Finally, we introduce the reinforcement learning problem and discuss two paradigms: Monte Carlo methods and temporal difference learning. We conclude the course by noting how the two paradigms lie on a spectrum of n-step temporal difference methods. An emphasis on algorithms and examples will be a key part of this course.Created by: Columbia University
Related Online Courses
This course is best suited for individuals currently in the healthcare sector, as a provider, payer, or administrator. Individuals pursuing a career change to the healthcare sector may also be... more
Azure provides a variety of ways to store data: unstructured, archival, relational, and more. In this course, you will learn the basics of storage management in Azure, how to create a Storage... more
This specialization is intended for post-graduate students seeking to develop practical machine-learning skills applicable across various domains. Through three comprehensive courses, learners will... more
This course is best suited for individuals who have a technical background in mathematics/statistics/computer science/engineering pursuing a career change to jobs or industries that are data-driven... more
Welcome to the Advanced Docker course! The \"Advanced Docker\" course aligns with the Docker Certified Associate (DCA) Certification requirements. This course is designed to deepen your... more