Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: A Theoretical Understanding of Gradient Bias in Meta-Reinforcement Learning
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v2-stream-010-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v2-stream-010-alpha.b-cdn.net
      • sl-yoda-v2-stream-010-beta.b-cdn.net
      • 1759419103.rsc.cdn77.org
      • 1016618226.rsc.cdn77.org
      • Subtitles
      • Off
      • English
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            A Theoretical Understanding of Gradient Bias in Meta-Reinforcement Learning
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            A Theoretical Understanding of Gradient Bias in Meta-Reinforcement Learning

            Nov 28, 2022

            Speakers

            BL
            BL

            Bo Liu

            Speaker · 1 follower

            XF

            Xidong Feng

            Speaker · 0 followers

            JR

            Jie Ren

            Speaker · 0 followers

            About

            Gradient-based Meta-RL (GMRL) refers to methods that maintain two-level optimisation procedures wherein the outer-loop meta-learner guides the inner-loop gradient-based reinforcement learner to achieve fast adaptations. In this paper, we develop a unified framework that describes variations of GMRL algorithms and points out that existing stochastic meta-gradient estimators adopted by GMRL are actually biased. Such meta-gradient bias comes from two sources: 1) the compositional bias incurred by t…

            Organizer

            N2
            N2

            NeurIPS 2022

            Account · 960 followers

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            MinVIS: A Minimal Video Instance Segmentation Framework without Video-based Training
            04:50

            MinVIS: A Minimal Video Instance Segmentation Framework without Video-based Training

            De-An Huang, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Vision-centric Autonomous Driving: from Perception to Prediction
            31:33

            Vision-centric Autonomous Driving: from Perception to Prediction

            Hang Zhao

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 1 viewers voted for saving the presentation to eternal vault which is 0.1%

            First Steps Toward Understanding the Extrapolation of Nonlinear Models to Unseen Domains
            08:34

            First Steps Toward Understanding the Extrapolation of Nonlinear Models to Unseen Domains

            Kefan Dong, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interactive Imitation Learning in Robotics
            26:07

            Interactive Imitation Learning in Robotics

            Jens Kober

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Language Models Can Teach Themselves to Program Better
            04:43

            Language Models Can Teach Themselves to Program Better

            Patrick Haluptzok, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Cache-Augmented Inbatch Importance Resampling for Training Recommender Retriever
            04:48

            Cache-Augmented Inbatch Importance Resampling for Training Recommender Retriever

            Jin Chen, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow NeurIPS 2022