Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: A case where a spindly two-layer linear network decisively outperforms any neural network with a fully connected input layer
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v3-stream-012-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v3-stream-012-alpha.b-cdn.net
      • sl-yoda-v3-stream-012-beta.b-cdn.net
      • 1338956956.rsc.cdn77.org
      • 1656830687.rsc.cdn77.org
      • Subtitles
      • Off
      • English
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            A case where a spindly two-layer linear network decisively outperforms any neural network with a fully connected input layer
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            A case where a spindly two-layer linear network decisively outperforms any neural network with a fully connected input layer

            Mar 9, 2021

            Speakers

            MKW

            Manfred K. Warmuth

            Speaker · 0 followers

            WK

            Wojciech Kotłowski

            Speaker · 0 followers

            EA

            Ehsan Amid

            Speaker · 0 followers

            About

            It was conjectured that any neural network of any structure and arbitrary differentiable transfer functions at the nodes cannot learn the following problem sample efficiently when trained with gradient descent: The instances are the rows of a $d$-dimensional Hadamard matrix and the target is one of the features, i.e. very sparse. We essentially prove this conjecture: We show that after receiving a random training set of size $k < d$, the expected squared loss is still $1-\frac{k}{(d-1)}$. Th…

            Organizer

            A2
            A2

            ALT 2021

            Account · 1 follower

            Categories

            AI & Data Science

            Category · 10.8k presentations

            About ALT 2021

            The 32nd International Conference on Algorithmic Learning Theory

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            Asymptotically Optimal Strategies For Combinatorial Semi-Bandits in Polynomial Time
            11:27

            Asymptotically Optimal Strategies For Combinatorial Semi-Bandits in Polynomial Time

            Thibaut Cuvelier, …

            A2
            A2
            ALT 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Sample Complexity Bounds for Stochastic Shortest Path with a Generative Model
            11:48

            Sample Complexity Bounds for Stochastic Shortest Path with a Generative Model

            Jean Tarbouriech, …

            A2
            A2
            ALT 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Last Round Convergence and No-Dynamic Regret in Asymmetric Repeated Games
            10:23

            Last Round Convergence and No-Dynamic Regret in Asymmetric Repeated Games

            Le Cong Dinh, …

            A2
            A2
            ALT 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            A Deep Conditioning Treatment of Neural Networks
            11:53

            A Deep Conditioning Treatment of Neural Networks

            Naman Agarwal, …

            A2
            A2
            ALT 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Precise Minimax Regret for Logistic Regression with Categorical Feature Values
            12:10

            Precise Minimax Regret for Logistic Regression with Categorical Feature Values

            Philippe Jacquet, …

            A2
            A2
            ALT 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Descent-to-Delete: Gradient-Based Methods for Machine Unlearning
            11:21

            Descent-to-Delete: Gradient-Based Methods for Machine Unlearning

            Saeed Sharifi-Malvajerdi, …

            A2
            A2
            ALT 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow ALT 2021