Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: Discount Factor as a Regularizer in Reinforcement Learning
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v2-stream-009-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v2-stream-009-alpha.b-cdn.net
      • sl-yoda-v2-stream-009-beta.b-cdn.net
      • 1766500541.rsc.cdn77.org
      • 1441886916.rsc.cdn77.org
      • Subtitles
      • Off
      • en
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            Discount Factor as a Regularizer in Reinforcement Learning
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            Discount Factor as a Regularizer in Reinforcement Learning

            Jul 12, 2020

            Speakers

            RA

            Ron Amit

            Speaker · 0 followers

            KC

            Kamil Ciosek

            Speaker · 0 followers

            RM

            Ron Meir

            Speaker · 0 followers

            About

            Specifying a Reinforcement Learning (RL) task involves choosing a suitable planning horizon, which is typically modeled by an evaluation discount factor. It is known that applying RL algorithms with a discount set lower than the evaluation discount factor can act as a regularizer, improving performance in the limited data regime. Yet the exact nature of this regularizer has not been investigated. In this work, we fill in this gap. For TD learning and expected SARSA, we show an explicit equivalen…

            Organizer

            I2
            I2

            ICML 2020

            Account · 2.7k followers

            Categories

            AI & Data Science

            Category · 10.8k presentations

            About ICML 2020

            The International Conference on Machine Learning (ICML) is the premier gathering of professionals dedicated to the advancement of the branch of artificial intelligence known as machine learning. ICML is globally renowned for presenting and publishing cutting-edge research on all aspects of machine learning used in closely related areas like artificial intelligence, statistics and data science, as well as important application areas such as machine vision, computational biology, speech recognition, and robotics. ICML is one of the fastest growing artificial intelligence conferences in the world. Participants at ICML span a wide range of backgrounds, from academic and industrial researchers, to entrepreneurs and engineers, to graduate students and postdocs.

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            INTERPRET: INTERACTION-Dataset-based PREdicTion Challenge
            20:30

            INTERPRET: INTERACTION-Dataset-based PREdicTion Challenge

            Wei Zhan, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Invited Talk 8

            Shanghang Zhang

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Refined bounds for algorithm configuration: The knife-edge of dual class approximability
            14:06

            Refined bounds for algorithm configuration: The knife-edge of dual class approximability

            Nina Balcan, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Coresets for Clustering in Graphs of Bounded Treewidth
            14:48

            Coresets for Clustering in Graphs of Bounded Treewidth

            Xuan Wu

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Soft Threshold Weight Reparameterization for Learnable Sparsity
            14:24

            Soft Threshold Weight Reparameterization for Learnable Sparsity

            Aditya Kusupati, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Energy-Based Models for Object-Oriented Learning
            42:02

            Energy-Based Models for Object-Oriented Learning

            Igor Mordatch

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow ICML 2020