Leo (Linxiao) Li
AI Infrastructure & LLM Systems
李麟霄
Shanghai, China
Welcome to my homepage!
I am a researcher focusing on large-scale LLM systems and AI infrastructure. Currently I work as an AI Infrastructure / Training Systems Researcher at AIGCode (蔻町科技).
My work focuses on large-scale LLM training systems. I study system-level optimizations to improve efficiency and hardware utilization, including Model FLOPs Utilization (MFU) improvement, communication–computation overlap, operator/kernel fusion, and runtime-level scheduling and memory optimization. I work closely with distributed training stacks to improve scalability, throughput, and cost-efficiency of foundation model training.
Previously, I was a Senior Researcher at Huawei Cloud Architecture Innovation Lab (Cloud Lab), working in Huawei Cloud CTO Fellow Team (Huawei Fellow- and Academician-led) on MoE inference, memory offloading, and cloud-scale AI infrastructure. I also collaborated with the Guangming Laboratory (Intelligent Recommendation & Resource Scheduling Group).
My long-term research goal is to design more efficient system architectures for large-scale AI models, improving scalability and resource efficiency of future AI infrastructure.
My research and engineering interests include:
- Efficient training and inference systems for large language models
- Model FLOPs Utilization (MFU) optimization and system efficiency
- Communication–computation overlap and operator / kernel fusion
- Distributed training systems and memory-efficient execution
- Scalable inference systems for large models (LLMs, MoE)
Check out my publications, CV, and gallery for more details.
News
| Apr 2026 | Paper Deterministic Component Mining for Multi-framework UI2Code Generation accepted to ICML 2026 as co-first author! (Acceptance rate: 26.6%) |
|---|---|
| Apr 2026 | Codex AutoResearch — my open-source project, a self-directed iterative research system inspired by Karpathy’s autoresearch concept — reached 1,000+ GitHub Stars! |
| Jan 2026 | Paper EcoThink accepted to The Web Conference 2026 (WWW’26) as first author! Average 40.4% energy savings for LLM agents. |
| Dec 2025 | Invited to give a talk at Guangming Laboratory on research methodology, internship planning, and career development for graduate students. |
| Oct 2025 | Joined AIGCode (蔻町科技) as Large Model Algorithm AI Infra Tech Leader, leading model pretraining optimization. |
| Jun 2025 | Ranked 14th out of 5200+ in the Huawei ICT Software Competition (Top 3 in Huawei Cloud, 1st in Chengdu Research Institute). |
| May 2025 | Ascend NPU User-Space Virtualization and MoE Memory Offloading solutions announced at Huawei HDC 2025 (~50 min mark) — industry-leading with <3% performance overhead. |
| May 2024 | Awarded Golden Cloud Award (金代码, Top 2%) and HCS Cloud Summit Star (云巅之星, Top 5%) by Huawei Cloud. |
| Jan 2023 | Received Huawei “Rising Star” (明日之星) Award again (Top 5%). |
| May 2022 | Received Huawei “Rising Star” (明日之星) Award (Top 5%). |
Selected Publications
More papers coming in 2026 — stay tuned!
- WWWIn Proceedings of the ACM Web Conference 2026 (WWW’26), 2026
- ICMLIn Proceedings of the 43rd International Conference on Machine Learning (ICML 2026), 2026