Everything You Need to Know about Transformers: Architectures, Optimization, Applications, and Interpretation
Time and Location
Instructors: Andy Zeng, Boqing Gong, Chen Sun, Ellie Pavlick, and Neil Houlsby
Sessions:
- Tuesday, February 7 2023, 14:00 - 18:00 (Eastern Time)
About
The tutorial aims to share the exciting recent developments on unified neural architectures that process different input modalities and learn to solve diverse tasks, from the perspective of Transformer architectures. The goal is to equip attendees with “everything they need to know about Transformers”. The tutorial covers the basic architecture and its recent variants (Neil Houlsby), effective optimization algorithms (Boqing Gong), representative and emerging new applications in multimodal learning and robotics (Chen Sun, Andy Zeng), and the tools to probe and analyze what knowledge has been captured by a trained network (Ellie Pavlick).
We envision the underlying principles for the success of Transformers are general, and the tutorial will be beneficial for a wide range of AI researchers and practitioners. Finally, the tutorial will discuss the limitations of existing Transformer-based approaches and highlight some future research directions.
We expect the participants to have general knowledge about machine learning and deep learning, including commonly used neural architectures and learning methods. A hands-on experience with computer vision, language understanding, or robotics research or applications is helpful, but not required.
Syllabus
Lecture slides
- 1
- The Transformer Architecture and its Variants
- Neil Houlsby
- Slides
- 2
- How to Train Your (Vision) Transformer?
- Boqing Gong
- Slides
- 3
- Does Multimodal Pretraining Learn Useful Representation for Reasoning?
- Chen Sun
- Slides
- 4
- Language as Robot Middleware
- Andy Zeng
- Slides
- 5
- Probing Knowledge and Structure in Transformers
- Ellie Pavlick
- Slides