Yushi Huang

Ph.D. candidate @ HKUST

Email  /  CV  /  Scholar  /  Twitter  /  Github

profile photo

I am a Ph.D. student at the Hong Kong University of Science and Technology (HKUST), supervised by Prof. Jun Zhang. I received my B.E. degree from Beihang University. Currently, I am interested in multimodal reinforcement learning (e.g., VLMs and diffusion models). Moreover, I focus on efficient training and inference for vision and language generative models.

News

  • 2026.01: ย ๐ŸŽ‰๐ŸŽ‰ Our QVGen is accepted to ICLR.
  • 2025.11: ย ๐ŸŽ‰๐ŸŽ‰ Our SlimInfer and LLMC+ are accepted to AAAI.
  • 2025.06: ย ๐ŸŽ‰๐ŸŽ‰ Our Temporal Feature Matters is accepted to TPAMI.
  • 2025.05: ย ๐ŸŽ‰๐ŸŽ‰ Our HarmoniCa is accepted to ICML.
  • 2024.10: ย ๐ŸŽ‰๐ŸŽ‰ Our LLMC is accepted to EMNLP Industry Track.
  • 2024.07: ย ๐ŸŽ‰๐ŸŽ‰ Our PTSBench is accepted to ACM MM.
  • 2024.02: ย ๐ŸŽ‰๐ŸŽ‰ Our TFMQ-DM is accepted to CVPR as a Highlight Poster (Top $2.8\%$).

Selected Papers

Includes preprints; * indicates equal contribution, ๐Ÿ“ง indicates corresponding author

Preprint
sym

MoDES: Accelerating Mixture-of-Experts Multimodal Large Language Models via Dynamic Expert Skipping

Yushi Huang, Zining Wang, Zhihang Yuan๐Ÿ“ง, Yifu Ding, Ruihao Gong, Jinyang Guo, Xianglong Liu, Jun Zhang๐Ÿ“ง

Preprint
sym

LinVideo: A Post-Training Framework towards $\mathcal{O}(n)$ Attention in Efficient Video Generation

Yushi Huang, Xingtong Ge, Ruihao Gong๐Ÿ“ง, Chengtao Lv, Jun Zhang๐Ÿ“ง

ICLR 2026
sym

QVGen: Pushing the Limit of Quantized Video Generative Models

Yushi Huang, Ruihao Gong๐Ÿ“ง, Jing Liu, Yifu Ding, Chengtao Lv, Haotong Qin, Jun Zhang๐Ÿ“ง

ICML 2025
sym

HarmoniCa: Harmonizing Training and Inference for Better Feature Caching in Diffusion Transformer Acceleration

Yushi Huang*, Zining Wang*, Ruihao Gong๐Ÿ“ง, Jing Liu, Xinjie Zhang, Jinyang Guo, Xianglong Liu, Jun Zhang๐Ÿ“ง

Projects

Toolkit
LightCompress

LightCompress is an off-the-shelf compression suite for AIGC models (LLMs, VLMs, diffusion, etc.) that packages SOTA quantization, sparsification, and deployment best practices to shrink models while preserving accuracy. 600+ GitHub Stars.

Services

  • Conference Reviews: NeurIPS, ICLR, ICML, COLM, AAAI, CVPR, ECCV.

Educations

  • 2025.02 - Now, Ph.D. in Electronic Computer and Engineering, the Hong Kong University of Science and Technology.
  • 2020.09 - 2024.06, B.Eng. in Computer Science and Engineering, Shenyuan Honors College, Beihang University.

Internships

  • 2025.09-2025.11, Bytedance Seed.
  • 2024.12 - 2025.02, Microsoft Research Asia.
  • 2023.05 - Now, SenseTime Research.