Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: Gradient Temporal-Difference Learning with Regularized Corrections
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v3-stream-014-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v3-stream-014-alpha.b-cdn.net
      • sl-yoda-v3-stream-014-beta.b-cdn.net
      • 1978117156.rsc.cdn77.org
      • 1243944885.rsc.cdn77.org
      • Subtitles
      • Off
      • en
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            Gradient Temporal-Difference Learning with Regularized Corrections
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            Gradient Temporal-Difference Learning with Regularized Corrections

            Jul 12, 2020

            Speakers

            SG

            Sina Ghiassian

            Speaker · 0 followers

            AP

            Andrew Patterson

            Speaker · 0 followers

            SG

            Shivam Garg

            Speaker · 0 followers

            About

            Value function learning remains a critical component of many reinforcement learning systems. Many algorithms are based on temporal difference (TD) updates, which have well-documented divergence issues, even though potentially sound alternatives exist like Gradient TD. Unsound approaches like Q-learning and TD remain popular because divergence seems rare in practice and these algorithms typically perform well. However, recent work with large neural network learning systems reveals that instabilit…

            Organizer

            I2
            I2

            ICML 2020

            Account · 2.7k followers

            Categories

            AI & Data Science

            Category · 10.8k presentations

            About ICML 2020

            The International Conference on Machine Learning (ICML) is the premier gathering of professionals dedicated to the advancement of the branch of artificial intelligence known as machine learning. ICML is globally renowned for presenting and publishing cutting-edge research on all aspects of machine learning used in closely related areas like artificial intelligence, statistics and data science, as well as important application areas such as machine vision, computational biology, speech recognition, and robotics. ICML is one of the fastest growing artificial intelligence conferences in the world. Participants at ICML span a wide range of backgrounds, from academic and industrial researchers, to entrepreneurs and engineers, to graduate students and postdocs.

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            Error Analysis of Nonnegative Tensor Train Utilized for Nonnegative Canonical Polyadic Decomposition
            04:00

            Error Analysis of Nonnegative Tensor Train Utilized for Nonnegative Canonical Polyadic Decomposition

            Svetlana Kuksova, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Efficient Statistical Inference for Population Variable Importance Using Shapley Values
            14:34

            Efficient Statistical Inference for Population Variable Importance Using Shapley Values

            Brian Williamson, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Density Deconvolution with Normalizing Flows
            04:05

            Density Deconvolution with Normalizing Flows

            Tim Dockhorn, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Variational Inference for Sequential Data with Future Likelihood Estimates
            11:39

            Variational Inference for Sequential Data with Future Likelihood Estimates

            Geon-Hyeong Kim, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Certified Robustness to Label-Flipping Attacks via Randomized Smoothing
            14:57

            Certified Robustness to Label-Flipping Attacks via Randomized Smoothing

            Elan Rosenfeld, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Bridging Worlds in Reinforcement Learning with Model-Advantage
            05:14

            Bridging Worlds in Reinforcement Learning with Model-Advantage

            Nirbhay Modhe, …

            I2
            I2
            ICML 2020 5 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow ICML 2020