Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: High-dimensional Asymptotics of Feature Learning: How One Gradient Step Improves the Representation
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v2-stream-008-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v2-stream-008-alpha.b-cdn.net
      • sl-yoda-v2-stream-008-beta.b-cdn.net
      • 1159783934.rsc.cdn77.org
      • 1511376917.rsc.cdn77.org
      • Subtitles
      • Off
      • English
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            High-dimensional Asymptotics of Feature Learning: How One Gradient Step Improves the Representation
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            High-dimensional Asymptotics of Feature Learning: How One Gradient Step Improves the Representation

            Nov 28, 2022

            Speakers

            JB

            Jimmy Ba

            Sprecher:in · 2 Follower:innen

            MAE

            Murat A. Erdogdu

            Sprecher:in · 0 Follower:innen

            TS

            Taiji Suzuki

            Sprecher:in · 1 Follower:in

            About

            We study the first gradient descent step on the first-layer parameters W in a two-layer neural network: f(x) = 1/√(N)a^⊤σ(W^⊤x), where W∈ℝ^d× N, a∈ℝ^N are randomly initialized, and the training objective is the empirical MSE loss: 1/n∑_i=1^n (f(x_i)-y_i)^2. In the proportional asymptotic limit where n,d,N→∞ at the same rate, and an idealized student-teacher setting where the teacher f^* is a single-index model, we compute the prediction risk of ridge regression on the conjugate kernel after one…

            Organizer

            N2
            N2

            NeurIPS 2022

            Konto · 961 Follower:innen

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            Lagrangian Model Based Reinforcement Learning
            07:29

            Lagrangian Model Based Reinforcement Learning

            Adithya Ramesh, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Ewigspeicher-Fortschrittswert: 0 = 0.0%

            Rethinking the Reverse-engineering of Trojan Triggers
            04:48

            Rethinking the Reverse-engineering of Trojan Triggers

            Zhenting Wang, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Ewigspeicher-Fortschrittswert: 0 = 0.0%

            From Advice-Taking to Active Learnig
            31:15

            From Advice-Taking to Active Learnig

            Daniel S. Weld

            N2
            N2
            NeurIPS 2022 2 years ago

            Ewigspeicher-Fortschrittswert: 0 = 0.0%

            A Mixture Of Surprises for Unsupervised Reinforcement Learning
            04:43

            A Mixture Of Surprises for Unsupervised Reinforcement Learning

            Andrew Zhao, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Ewigspeicher-Fortschrittswert: 0 = 0.0%

            A Robust Phased Elimination Algorithm for Corruption-Tolerant Gaussian Process Bandits
            04:00

            A Robust Phased Elimination Algorithm for Corruption-Tolerant Gaussian Process Bandits

            Ilija Bogunovic, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Ewigspeicher-Fortschrittswert: 0 = 0.0%

            What is Where by Looking (WWbL) - Weakly-Supervised Open-World Phrase-Grounding without Text Inputs
            05:04

            What is Where by Looking (WWbL) - Weakly-Supervised Open-World Phrase-Grounding without Text Inputs

            Tal Shaharabany, …

            N2
            N2
            NeurIPS 2022 2 years ago

            Ewigspeicher-Fortschrittswert: 0 = 0.0%

            Interested in talks like this? Follow NeurIPS 2022