Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: Fast Causal Attention with Dynamic Sparsity
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v2-stream-001-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v2-stream-001-alpha.b-cdn.net
      • sl-yoda-v2-stream-001-beta.b-cdn.net
      • 1824830694.rsc.cdn77.org
      • 1979322955.rsc.cdn77.org
      • Subtitles
      • Off
      • English
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            Fast Causal Attention with Dynamic Sparsity
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            Fast Causal Attention with Dynamic Sparsity

            Jul 28, 2023

            Speakers

            MP

            Matteo Pagliardini

            Speaker · 0 followers

            DP

            Daniele Paliotta

            Speaker · 0 followers

            MJ

            Martin Jaggi

            Speaker · 1 follower

            About

            Transformer-based language models have found many diverse applications requiring them to process sequences of increasing length. For these applications, the causal self-attention---which is the only component scaling quadratically w.r.t. the sequence length---becomes a central concern. While many works have proposed schemes to sparsify the attention patterns and reduce the computational overhead of self-attention, those are often limited by implementation concerns and end up imposing a simple an…

            Organizer

            I2
            I2

            ICML 2023

            Account · 626 followers

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            Efficient Transformed Gaussian Processes
            05:25

            Efficient Transformed Gaussian Processes

            Juan Maroñas, …

            I2
            I2
            ICML 2023 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Entity Divider with Language Grounding in Multi-Agent Reinforcement Learning
            05:15

            Entity Divider with Language Grounding in Multi-Agent Reinforcement Learning

            Ziluo Ding, …

            I2
            I2
            ICML 2023 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Formal Control Synthesis for Stochastic Neural Network Dynamic Models
            13:32

            Formal Control Synthesis for Stochastic Neural Network Dynamic Models

            Steven Adams, …

            I2
            I2
            ICML 2023 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Low Complexity Homeomorphic Projection to Ensure Neural-Network Solution Feasibility for Optimization over (Non-)Convex Set
            04:53

            Low Complexity Homeomorphic Projection to Ensure Neural-Network Solution Feasibility for Optimization over (Non-)Convex Set

            Enming Liang, …

            I2
            I2
            ICML 2023 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Data-Efficient Contrastive Self-supervised Learning
            05:14

            Data-Efficient Contrastive Self-supervised Learning

            Siddharth Joshi, …

            I2
            I2
            ICML 2023 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            UPop: Unified and Progressive Pruning for Compressing Vision-Language Transformers
            04:38

            UPop: Unified and Progressive Pruning for Compressing Vision-Language Transformers

            Dachuan Shi, …

            I2
            I2
            ICML 2023 2 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow ICML 2023