Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: Provable Benefit of Multitask Representation Learning in Reinforcement Learning
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v2-stream-001-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v2-stream-001-alpha.b-cdn.net
      • sl-yoda-v2-stream-001-beta.b-cdn.net
      • 1824830694.rsc.cdn77.org
      • 1979322955.rsc.cdn77.org
      • Subtitles
      • Off
      • English
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            Provable Benefit of Multitask Representation Learning in Reinforcement Learning
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            Provable Benefit of Multitask Representation Learning in Reinforcement Learning

            Nov 28, 2022

            Speakers

            YC

            Yuan Cheng

            Speaker · 0 followers

            SF

            Songtao Feng

            Speaker · 0 followers

            JY

            Jing Yang

            Speaker · 0 followers

            About

            As representation learning becomes a powerful technique to reduce sample complexity in reinforcement learning (RL) in practice, theoretical understanding of its advantage is still limited. In this paper, we theoretically characterize the benefit of representation learning under the low-rank Markov decision process (MDP) model. We first study multitask low-rank RL (as upstream training), where all tasks share a common representation, and propose a new multitask reward-free algorithm called REFUEL…

            Organizer

            N2
            N2

            NeurIPS 2022

            Account · 954 followers

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            Multi-step Planning for Automated Hyperparameter Optimization with OptFormer
            05:27

            Multi-step Planning for Automated Hyperparameter Optimization with OptFormer

            Lucio Dery, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Scaling and Shifting Your Features: A New Baseline for Efficient Model Tuning
            05:03

            Scaling and Shifting Your Features: A New Baseline for Efficient Model Tuning

            Dongze Lian, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Efficient Speech Translation with Pre-trained Models
            08:33

            Efficient Speech Translation with Pre-trained Models

            Zhaolin Li, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Artwork Spotlight: Machine Reflections: A Self-Portrait Series
            05:50

            Artwork Spotlight: Machine Reflections: A Self-Portrait Series

            Orsolya Szantho

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Fine-Tuning without Distortion: Improving Robustness to Distribution Shifts
            37:08

            Fine-Tuning without Distortion: Improving Robustness to Distribution Shifts

            Percy Liang, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Between Stochastic and Adversarial Online Convex Optimization: Improved Regret Bounds via Smoothness
            04:54

            Between Stochastic and Adversarial Online Convex Optimization: Improved Regret Bounds via Smoothness

            Sarah Sachs, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow NeurIPS 2022