📜 Biography
I am a Ph.D. candidate of Beijing Institute of Technology (BIT), advised by Prof. Yuwei Wu and Prof. Yunde Jia. I received my Master degree in 2020 from Northeastern University supervised by Prof. Shukuan Lin, and Bachlor degree in 2017 from Harbin University of Science and Technology. Now, I have also interned at the Shanghai AI Lab, collaborating closely with Dr. Kaipeng Zhang and Dr. Wenqi Shao.
My research area lies at:
- vision-and-language
- multimodal large language models
- compositional generalization
🎓 Education
- 2020.09 - 2025.03,
Ph.D. in CS, Beijing Institute of Technology, Beijing, China
- 2017.09 - 2020.01,
Master in CS, Northeastern University, Shenyang, Liaoning, China
- 2013.09 - 2017.06,
Bachelor in CS, Harbin University of Science and Technology, Harbin, Heilongjiang, China
⚡ Preprint
ARMOR: Empowering Autoregressive Multimodal Understanding Model with Interleaved Multimodal Generation via Asymmetric Synergy
- Jianwen Sun, Yukang Feng,
Chuanhao Li
, Fanrui Zhang, Zizhen Li, Jiaxin Ai, Sizhuo Zhou, Pengfei Zhou, Yu Dai, Shenglin Zhang, and Kaipeng Zhang. - [arXiv 2025] [paper] [code]
📝 Publications
GATE OpenING: A Comprehensive Benchmark for Judging Open-ended Interleaved Image-Text Generation
- Pengfei Zhou, Xiaopeng Peng, Jiajun Song,
Chuanhao Li
, Zhaopan Xu, Yue Yang, Ziyao Guo, Hao Zhang, Yuqi Lin, Yefei He, Lirui Zhao, Shuo Liu, Tianhua Li, Yuxuan Xie, Xiaojun Chang, Yu Qiao, Wenqi Shao, and Kaipeng Zhang. - [CVPR 2025] [paper] [code]
MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models
- Fanqing Meng*, Jin Wang*,
Chuanhao Li*
, Quanfeng Lu, Hao Tian, Jiaqi Liao, Xizhou Zhu, Jifeng Dai, Yu Qiao, Ping Luo, Kaipeng Zhang, and Wenqi Shao. (* equal contribution) - [ICLR 2025] [paper] [code]
Consistency of Compositional Generalization across Multiple Levels
Chuanhao Li
, Zhen Li, Chenchen Jing, Xiaomeng Fan, Wenbo Ye, Yuwei Wu, and Yunde Jia.- [AAAI 2025] [paper] [code]
SearchLVLMs: A Plug-and-Play Framework for Augmenting Large Vision-Language Models by Searching Up-to-Date Internet Knowledge
Chuanhao Li
, Zhen Li, Chenchen Jing, Shuo Liu, Wenqi Shao, Yuwei Wu, Ping Luo, Yu Qiao, and Kaipeng Zhang.- [NeurIPS 2024] [paper] [code]
ConvBench: A Multi-Turn Conversation Evaluation Benchmark with Hierarchical Capability for Large Vision-Language Models
- Shuo Liu, Kaining Ying, Hao Zhang, Yue Yang, Yuqi Lin, Tianle Zhang,
Chuanhao Li
, Yu Qiao, Ping Luo, Wenqi Shao, and Kaipeng Zhang. - [NeurIPS 2024] [paper] [code]
Compositional Substitutivity of Visual Reasoning for Visual Question Answering
Chuanhao Li
, Zhen Li, Chenchen Jing, Yuwei Wu, Mingliang Zhai, and Yunde Jia.- [ECCV 2024] [paper] [code]
In-Context Compositional Generalization for Large Vision-Language Models
Chuanhao Li
, Chenchen Jing, Zhen Li, Mingliang Zhai, Yuwei Wu, and Yunde Jia.- [EMNLP 2024] [paper]
Adversarial Sample Synthesis for Visual Question Answering
Chuanhao Li
, Chenchen Jing, Zhen Li, Yuwei Wu, and Yunde Jia.- [TOMM 2024] [paper]
Exploring the Effect of Primitives for Compositional Generalization in Vision-and-Language
Learning the Dynamics of Visual Relational Reasoning via Reinforced Path Routing
- Chenchen Jing, Yunde Jia, Yuwei Wu,
Chuanhao Li
, and Qi Wu. - [AAAI 2022] [paper]
🏅 Selected Awards
- 2023.01, the second prize in the multi-modal technology innovation competition of the first “Xingzhi Cup” National Artificial Intelligence Innovation Application Competition
- 2016.05, the first prize in the CCPC Heilongjiang Collegiate Programming Contest
- 2015.05, the first prize in the CCPC Heilongjiang Collegiate Programming Contest
- 2014.07, the silver medal in the ACM-ICPC Collegiate Programming Contest Shanghai Invitational
🏛️ Academic Activities
- Conference Reviewer of MM2025, ICML 2025, IJCAI 2025, CVPR 2024, NeurIPS 2024, MM 2024, etc.
- Journal Reviewer of T-MM.
- Invited Speaker in The 3rd SMBU-BIT Machine Intelligence Graduate Student Forum.
💻 Work Experience
- 2024.01 - Present, Intern, Shanghai AI Lab, Shanghai, China
- 2019.07 - 2019.10, Intern, UISEE, Beijing, China