About The Seminar
The ASAP is a fully virtual seminar that, while emphasizing efficiency as its name suggests, takes a distinct approach from traditional ML systems seminars. As foundation models grow increasingly powerful through scaling of data and model size, their capabilities remain inherently limited by their architectures. This seminar explores novel architectural designs and paradigms for both training and inference, examining sequence modeling approaches from an algorithmic perspective. The seminar serves as a bridge between theoretical, algorithmic, and systems research communities, tackling fundamental challenges in Transformer models—including reasoning capabilities, generalization power, memorization, and expressiveness. By adopting first-principles approaches, it seeks solutions that address these limitations while maintaining computational efficiency on modern hardware, with the ultimate goal of motivating the next generation of foundation models that overcome current limitations.
Sources
Want to present?
1. Get in touch
Contact us via email or any other method with the topic you'd like to present and your preferred time slot.
2. We add you
We will add you to an available time slot.
3. Present
Present your paper or a topic of interest.
Event Schedule
3/27/2025
3/20/2025
3/18/2025
3/12/2025
3/5/2025
3/3/2025
2/24/2025
Organizers
This seminar is organized by

Songlin Yang
Massachusetts Institute of Technology

Malachy Yang
Carnegie Mellon University

Han Guo
Massachusetts Institute of Technology

Simran Arora
Stanford University