Leo (Linxiao) Li

AI Infrastructure & LLM Systems

prof_pic.jpg

李麟霄

Shanghai, China

Welcome to my homepage!

I am a researcher focusing on large-scale LLM systems and AI infrastructure. Currently I work as an AI Infrastructure / Training Systems Researcher at AIGCode (蔻町科技).

My work focuses on large-scale LLM training systems. I study system-level optimizations to improve efficiency and hardware utilization, including Model FLOPs Utilization (MFU) improvement, communication–computation overlap, operator/kernel fusion, and runtime-level scheduling and memory optimization. I work closely with distributed training stacks to improve scalability, throughput, and cost-efficiency of foundation model training.

Previously, I was a Senior Researcher at Huawei Cloud Architecture Innovation Lab (Cloud Lab), working in Huawei Cloud CTO Fellow Team (Huawei Fellow- and Academician-led) on MoE inference, memory offloading, and cloud-scale AI infrastructure. I also collaborated with the Guangming Laboratory (Intelligent Recommendation & Resource Scheduling Group).

My long-term research goal is to design more efficient system architectures for large-scale AI models, improving scalability and resource efficiency of future AI infrastructure.

My research and engineering interests include:

  • Efficient training and inference systems for large language models
  • Model FLOPs Utilization (MFU) optimization and system efficiency
  • Communication–computation overlap and operator / kernel fusion
  • Distributed training systems and memory-efficient execution
  • Scalable inference systems for large models (LLMs, MoE)

Check out my publications, CV, and gallery for more details.


News

Apr 2026 Paper Deterministic Component Mining for Multi-framework UI2Code Generation accepted to ICML 2026 as co-first author! (Acceptance rate: 26.6%)
Apr 2026 Codex AutoResearch — my open-source project, a self-directed iterative research system inspired by Karpathy’s autoresearch concept — reached 1,000+ GitHub Stars!
Jan 2026 Paper EcoThink accepted to The Web Conference 2026 (WWW’26) as first author! Average 40.4% energy savings for LLM agents.
Dec 2025 Invited to give a talk at Guangming Laboratory on research methodology, internship planning, and career development for graduate students.
Oct 2025 Joined AIGCode (蔻町科技) as Large Model Algorithm AI Infra Tech Leader, leading model pretraining optimization.
Jun 2025 Ranked 14th out of 5200+ in the Huawei ICT Software Competition (Top 3 in Huawei Cloud, 1st in Chengdu Research Institute).
May 2025 Ascend NPU User-Space Virtualization and MoE Memory Offloading solutions announced at Huawei HDC 2025 (~50 min mark) — industry-leading with <3% performance overhead.
May 2024 Awarded Golden Cloud Award (金代码, Top 2%) and HCS Cloud Summit Star (云巅之星, Top 5%) by Huawei Cloud.
Jan 2023 Received Huawei “Rising Star” (明日之星) Award again (Top 5%).
May 2022 Received Huawei “Rising Star” (明日之星) Award (Top 5%).

Selected Publications

More papers coming in 2026 — stay tuned!

  1. WWW
    Linxiao Li* and Zhixiang Lu*†
    In Proceedings of the ACM Web Conference 2026 (WWW’26), 2026
  2. ICML
    Zixiong Yang*, Linxiao Li*, Jiaye Lin, and 3 more authors
    In Proceedings of the 43rd International Conference on Machine Learning (ICML 2026), 2026