Systems for LLM Inference

  1. SlimInfer: Accelerating Long-Context LLM Inference via Dynamic Token Pruning. Lingkun Long, Ruibin Yang, Yushi Huang, Desheng Hui, Ao Zhou, Jianlei Yang, AAAI Conference on Artificial Intelligence (AAAI), 2026.
Zeyu Hao
Zeyu Hao
Master Student
Rubing Yang
Rubing Yang
Master Student
Lingkun Long
Lingkun Long
Master Student