On the Computational Hardness of Transformers
Barna Saha, Yinzhan Xu, Christopher Ye, Hantao Yu
Symposium on Theory of Computing (STOC), 2026
Two Heads are Better than One: Simulating Large Transformers with Small Ones [arXiv]
Hantao Yu, Josh Alman
Spotlight at Neural Information processing Systems (NeurIPS), 2025
Fast Attention Mechanisms: a Tale of Parallelism [arXiv]
Jingwen Liu, Hantao Yu, Clayton Sanford, Alexandr Andoni, Daniel Hsu
Neural Information processing Systems (NeurIPS), 2025
Fundamental Limitations on Subquadratic Alternatives to Transformers* [arXiv]
Josh Alman, Hantao Yu
International Conference on Learning Representations (ICLR), 2025
Theoretical Computer Science
Improving the Leading Constant of Matrix Multiplication* [arXiv]
Josh Alman, Hantao Yu
Symposium on Discrete Algorithms (SODA), 2025
Tensor Ranks and the Fine-Grained Complexity of Dynamic Programming* [arXiv]
Josh Alman, Ethan Turok, Hantao Yu, Hengzhi Zhang
Innovations in Theoretical Computer Science (ITCS), 2024
Robust Empirical Risk Minimization with Tolerance [arXiv]
Robi Bhattacharjee, Max Hopkins, Akash Kumar, Hantao Yu, Kamalika Chaudhuri
International Conference on Algorithmic Learning Theory (ALT), 2023
Active Learning Polynomial Threshold Functions* [arXiv]
Omri Ben-Eliezer, Max Hopkins, Chutong Yang, Hantao Yu
Neural Information processing Systems (NeurIPS), 2022