
Audio is streamed directly from the publisher (mcdn.podbean.com) as published in their RSS feed. Play Podcasts does not host this file. Rights-holders can request removal through the copyright & takedown page.
Show Notes
This episode explores MDPs, covering stochastic environments, transition functions, reward functions, policies, value iteration, policy iteration, expected utility, finite vs. infinite horizons, discount factors, etc.
Disclosure: This episode was generated using NotebookLM by uploading Professor Chris Callison-Burch's lecture notes and slides.