Theory of Large Language Models (LLMs)
Studying the capabilities of large language models, including but not limited to Transformer-based models, from a theoretical perspective.
I am currently a PhD student at the Gaoling School of Artificial Intelligence, Renmin University of China, advised by Prof. Zhewei Wei.
Before that, I received my master's degree in Artificial Intelligence from the Gaoling School of Artificial Intelligence, Renmin University of China in June 2025, and my bachelor's degree in Computer Science and Technology from the School of Information, Renmin University of China in June 2022.
My current research focuses on theoretical aspects of artificial intelligence, machine learning, and computation.
Studying the capabilities of large language models, including but not limited to Transformer-based models, from a theoretical perspective.
Studying the capabilities of graph neural networks from a theoretical perspective, with a particular focus on their ability to perform graph algorithms.
Designing and analyzing algorithms for large-scale graphs.
Gaoling School of Artificial Intelligence, Renmin University of China
Gaoling School of Artificial Intelligence, Renmin University of China
One paper, Position: The Turing-Completeness of Real-World Autoregressive Transformers Relies Heavily on Context Management, was accepted by ICML 2026.
One paper, Mixing Time Matters: Accelerating Effective Resistance Estimation via Bidirectional Method, was accepted by KDD 2025.
One paper, MGNN: Graph Neural Networks Inspired by Distance Geometry Problem, was accepted by KDD 2023.
Authors marked with * are corresponding authors.
Preprint.
43rd International Conference on Machine Learning (ICML), 2026.
31st ACM SIGKDD Conference on Knowledge Discovery & Data Mining (KDD), 2025.
29th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (KDD), 2023.