Next
Livestream will start soon!
Livestream has already ended.
Presentation has not been recorded yet!
  • title: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
      0:00 / 0:00
      • Report Issue
      • Settings
      • Playlists
      • Bookmarks
      • Subtitles Off
      • Playback rate
      • Quality
      • Settings
      • Debug information
      • Server sl-yoda-v2-stream-007-alpha.b-cdn.net
      • Subtitles size Medium
      • Bookmarks
      • Server
      • sl-yoda-v2-stream-007-alpha.b-cdn.net
      • sl-yoda-v2-stream-007-beta.b-cdn.net
      • 1678031076.rsc.cdn77.org
      • 1932936657.rsc.cdn77.org
      • Subtitles
      • Off
      • English
      • Playback rate
      • Quality
      • Subtitles size
      • Large
      • Medium
      • Small
      • Mode
      • Video Slideshow
      • Audio Slideshow
      • Slideshow
      • Video
      My playlists
        Bookmarks
          00:00:00
            An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
            • Settings
            • Sync diff
            • Quality
            • Settings
            • Server
            • Quality
            • Server

            An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

            May 3, 2021

            Speakers

            AD

            Alexey Dosovitskiy

            Speaker · 0 followers

            LB

            Lucas Beyer

            Speaker · 1 follower

            AK

            Alexander Kolesnikov

            Speaker · 0 followers

            About

            While the Transformer architecture has become the de-facto standard for natural language processing tasks, its applications to computer vision remain limited. In vision, attention is either applied in conjunction with convolutional networks, or used to replace certain components of convolutional networks while keeping their overall structure in place. We show that this reliance on CNNs is not necessary and a pure transformer applied directly to sequences of image patches can perform very well on…

            Organizer

            I2
            I2

            ICLR 2021

            Account · 887 followers

            Categories

            AI & Data Science

            Category · 10.8k presentations

            About ICLR 2021

            The International Conference on Learning Representations (ICLR) is the premier gathering of professionals dedicated to the advancement of the branch of artificial intelligence called representation learning, but generally referred to as deep learning. ICLR is globally renowned for presenting and publishing cutting-edge research on all aspects of deep learning used in the fields of artificial intelligence, statistics and data science, as well as important application areas such as machine vision, computational biology, speech recognition, text understanding, gaming, and robotics.

            Like the format? Trust SlidesLive to capture your next event!

            Professional recording and live streaming, delivered globally.

            Sharing

            Recommended Videos

            Presentations on similar topic, category or speaker

            TADPOLE: Task ADapted Pre-training via anOmaLy dEtection
            03:10

            TADPOLE: Task ADapted Pre-training via anOmaLy dEtection

            Vivek Madan, …

            I2
            I2
            ICLR 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Semantic Re-tuning with Contrastive Tension
            05:07

            Semantic Re-tuning with Contrastive Tension

            Fredrik Carlsson, …

            I2
            I2
            ICLR 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Semantic Audio-Visual Navigation
            05:48

            Semantic Audio-Visual Navigation

            Changan Chen, …

            I2
            I2
            ICLR 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Curious, cooperative, and communicative: Young children as never-ending learners
            19:22

            Curious, cooperative, and communicative: Young children as never-ending learners

            Hyowon Gweon

            I2
            I2
            ICLR 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Multi-class Uncertainty Calibration via Mutual Information Maximization-based Binning (I-Max Calibration)
            05:13

            Multi-class Uncertainty Calibration via Mutual Information Maximization-based Binning (I-Max Calibration)

            Kanil Patel, …

            I2
            I2
            ICLR 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Learning a Latent Search Space for Routing Problems using Variational Autoencoders
            04:56

            Learning a Latent Search Space for Routing Problems using Variational Autoencoders

            André Hottung, …

            I2
            I2
            ICLR 2021 4 years ago

            Total of 0 viewers voted for saving the presentation to eternal vault which is 0.0%

            Interested in talks like this? Follow ICLR 2021