I am a PhD student at Beihang University, supervised by Prof. Xianglong Liu. I am currently focused on the practical application of large AIGC models such as LLM and Diffusion, including inference acceleration and model quantization. I hope that advanced technological methods can truly bring convenience to peopleโ€™s lives.

๐Ÿ“ Publications

2024

NeurIPS 2024
sym

BiDM: Pushing the Limit of Quantization for Diffusion Models

Xingyu Zheng, Haotong Qin, Yichen Bian, Xudong Ma, Yulun Zhang, Jiakai Wang, Jinyang Guo, Xianglong Liu

Github

  • This paper proposes a novel method, namely BiDM, for fully binarizing weights and activations of DMs, pushing quantization to the 1-bit limit.
arxiv
sym

A Survey of Low-bit Large Language Models: Basics, Systems, and Algorithms

Ruihao Gong, Yifu Ding, Zining Wang, Chengtao Lv, Xingyu Zheng, Jinyang Du, Haotong Qin, Jinyang Guo, Michele Magno, Xianglong Liu

  • This paper presents a comprehensive survey of low-bit quantization methods tailored for LLMs, covering the fundamental principles, system implementations, and algorithmic strategies.
arxiv
sym

An Empirical Study of LLaMA3 Quantization: From LLMs to MLLMs

Wei Huang*, Xingyu Zheng*, Xudong Ma*, Haotong Qin, Chengtao Lv, Hong Chen, Jie Luo, Xiaojuan Qi, Xianglong Liu, Michele Magno

Github Hugging Face

  • This paper explores LLaMA3โ€™s capabilities when quantized to low bit-width, demonstrating its value in advancing future models.
arxiv
sym

BinaryDM: Accurate Weight Binarization for Efficient Diffusion Models

Xingyu Zheng, Xianglong Liu, Haotong Qin, Xudong Ma, Mingyuan Zhang, Haojie Hao, Jiakai Wang, Zixiang Zhao, Jinyang Guo, Michele Magno

Github

  • This paper proposes BinaryDM, a novel accurate quantization-aware training approach to push the weights of diffusion models towards the limit of 1-bit.
ICML 2024 (Oral)
sym

Accurate LoRA-Finetuning Quantization of LLMs via Information Retention

Haotong Qin*, Xudong Ma*, Xingyu Zheng, Xiaoyang Li, Yang Zhang, Shouda Liu, Jie Luo, Xianglong Liu, Michele Magno

Github

  • This paper proposes a novel IR-QLoRA for pushing quantized LLMs with LoRA to be highly accurate through information retention.

2023

ACM MM 2023
sym

DIsolation and Induction: Training Robust Deep Neural Networks against Model Stealing Attacks

Jun Guo, Xingyu Zheng, Aishan Liu, Siyuan Liang, Yisong Xiao, Yichao Wu, Xianglong Liu

Github

  • This paper proposes Isolation and Induction (InI), a novel and effective training framework for model stealing defenses.

Book

CCIS
sym

Generalizing from Limited Resources in the Open World

Jinyang Guo, Yuqing Ma, Yifu Ding, Ruihao Gong, Xingyu Zheng, Changyi He, Yantao Lu, Xianglong Liu

  • This book presents the Proceedings from the Second International Workshop GLOW 2024 held in conjunction with the International Joint Conference on Artificial Intelligence, IJCAI 2024, in Jeju Island, South Korea, in August 2024.

๐ŸŽ– Honors

  • 2024, Honor Student of Beihang University.
  • 2024, Excellent Undergraduate Graduation Thesis.
  • 2024, Excellent Graduate of Beihang University.
  • 2022, Lanqiao Cup, National First prize.

๐Ÿ“– Educations

  • 2024 - now, PhD student, Beihang University.
  • 2020 - 2024, Undergrad student, Beihang University.