
Dear all, Our next AI seminar is scheduled to be on January 12th, 2-3 PM. It will be followed by a 30-minute Q&A session with the graduate students. Location: KEC 1001 Zoom link: https://oregonstate.zoom.us/j/98684050301?pwd=ZzhianQxUFBPUmdYVWJKOFhaVURCQT09<https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Foregonstate.zoom.us%2Fj%2F98684050301%3Fpwd%3DZzhianQxUFBPUmdYVWJKOFhaVURCQT09&data=05%7C02%7Cai%40engr.orst.edu%7C5bb5e609969e446f615808dc122fe318%7Cce6d05e13c5e4d6287a84c4a2713c113%7C0%7C0%7C638405244080170811%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=Ums5X5CWxmfGJ0%2BZQZvkUB45bf%2FK7ZmvklBGVG0ujK8%3D&reserved=0> Reinforcement Learning from a Bayesian perspective Brendan O'Donoghue Research Scientist DeepMind Abstract: Reinforcement learning (RL) involves an agent interacting with an environment over time attempting to maximize its total return. Initially the agent does not know about the environment and must learn about it from experience. As the agent navigates the environment it receives noisy observations which it can use to update its (posterior) beliefs about the environment. Therefore, the RL problem is a statistical inference problem wrapped in a control problem, and the two problems must be tackled simultaneously for good data efficiency. This is because the policy of the agent affects the data it will collect, which in turn affects the policy, and so on. This is in contrast to supervised learning, where the performance of a classifier (for instance) does not influence the data it will later observe. Failure to properly consider the statistical aspect of the RL problem will result in agents that require exponential amounts of experience for good performance. On the other hand, correctly considering the statistical inference problem and the control problem together has the potential to dramatically reduce the compute requirements to solve problems and potentially unlock new domains and capabilities far outside of the range of current agents. In this talk I will introduce these concepts and discuss how Bayesian techniques can provide principled solutions to the problem. Speaker Bio: Brendan O'Donoghue earned his PhD in 2013 from Stanford working with Stephen Boyd on optimization and control theory. Since then he has worked at DeepMind as a research scientist working on deep reinforcement learning, optimization, and (more recently) large language models. Please watch this space for future AI Seminars : https://engineering.oregonstate.edu/EECS/research/AI<https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fengineering.oregonstate.edu%2FEECS%2Fresearch%2FAI&data=05%7C02%7Cai%40engr.orst.edu%7C5bb5e609969e446f615808dc122fe318%7Cce6d05e13c5e4d6287a84c4a2713c113%7C0%7C0%7C638405244080170811%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=RjOxqfVHqIEkIaF%2FEavxS%2BzdSoDPskdgQHthgMxDQd0%3D&reserved=0> Rajesh Mangannavar, Graduate Student Oregon State University ---- AI Seminar Important Reminders: -> For graduate students in the AI program, attendance is strongly encouraged
participants (1)
-
Mangannavar, Rajesh Devaraddi