Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: Variance Double-Down: The Small Batch Size Anomaly in Multistep Deep Reinforcement Learning
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v2-stream-001-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v2-stream-001-alpha.b-cdn.net
      • sl-yoda-v2-stream-001-beta.b-cdn.net
      • 1824830694.rsc.cdn77.org
      • 1979322955.rsc.cdn77.org
      • Subtitles
      • Off
      • English
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            Variance Double-Down: The Small Batch Size Anomaly in Multistep Deep Reinforcement Learning
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            Variance Double-Down: The Small Batch Size Anomaly in Multistep Deep Reinforcement Learning

            Dec 2, 2022

            Speakers

            JOC

            Johan Obando Ceron

            Speaker · 0 followers

            MB

            Marc Bellemare

            Speaker · 3 followers

            PSC

            Pablo Samuel Castro

            Speaker · 1 follower

            About

            In deep reinforcement learning, multi-step learning is almost unavoidable to achieve state-of-the-art performance. However, the increased variance that multistep learning brings makes it difficult to increase the update horizon beyond relatively small numbers. In this paper, we report the counterintuitive finding that decreasing the batch size parameter improves the performance of many standard deep RL agents that use multi-step learning. It is well-known that gradient variance decreases with in…

            Organizer

            N2
            N2

            NeurIPS 2022

            Account · 962 followers

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            Foundation Models for History Compression in RL
            04:18

            Foundation Models for History Compression in RL

            Fabian Paischer, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Poisoning Generative Models to Promote Catastrophic Forgetting
            10:20

            Poisoning Generative Models to Promote Catastrophic Forgetting

            Siteng Kang, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees
            04:39

            Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees

            Daniil Tiapkin, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Empirical Gateaux Derivatives for Causal Inference
            05:01

            Empirical Gateaux Derivatives for Causal Inference

            Angela Zhou, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            [Re] Learning Unknown from Correlations: Graph Neural Network for Inter-novel-protein Interaction Prediction
            05:04

            [Re] Learning Unknown from Correlations: Graph Neural Network for Inter-novel-protein Interaction Prediction

            Urša Zrimšek

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Understanding the Failure of Batch Normalization for Transformers in NLP
            04:45

            Understanding the Failure of Batch Normalization for Transformers in NLP

            Jiaxi Wang, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow NeurIPS 2022