Biography
Dr. Kejun Wu (Faculty HomePage) is currently a Lecturer/Assistant Professor at School of Electronic Information and Communications, Huazhong University of Science and Technology, Wuhan, China. Before that, he worked as a Research Fellow researcher from 2022 to 2024 at School of Electrical and Electronic Engineering, Nanyang Technological University (work with Prof. Yap Kim-Hui and Chau Lap-Pui周立培). He obtained his Ph.D. degree from Huazhong University of Science and Technology supervised by You Yang (杨铀), Master degree from Harbin Engineering University supervised by Chengtao Cai (蔡成涛), and Bachelor degree from Shanghai University. His research interest includes Multimodal Large Language Models, Generative Signal/Image Processing, and Learned Video Compression. He has published more than 40 papers in the past 3 years. He serves as an Editorial Board Member / Executive Guest Editor of ASOC, JVCIR, and JRTIP, etc., an Area Chair / Session Chair / Program Committee in IJCNN 2025, AIGC 2025, IEEE ICASSP2024, ISCAS2024, and MMSP2023. He is an IEEE Senior Member.
🚀 Recruiting intern/ visiting/ Master/ PhD students interested in MLLM and Gen-AI. Flexible remote learning is welcome.
🚀 招收兴趣从事多模态大模型、生成式AI研究的实习生/ 访问学生/ 研究生, 欢迎线下或远程学习, 欢迎联系.
Professional Activities
- Editorial Board, Applied Soft Computing
- Lead Guest Editor, Journal of Real-Time Image Processing (Co-Guest Editor, 蔡成涛,程明明,王亮)
- Lead Guest Editor, Journal of Visual Communication and Image Representation
- Experts, UWA Technical Planning Expert Group, UHD World Association (UWA, 世界超高清视频产业联盟)
- Member, 应急态势感知与应急通信技术决策咨询专家团队, 中国科协
- Program Committee, AIGC 2025, Hangzhou, China
- Area Chair, IJCNN 2025, Rome, Italy
- Session Chair, IEEE ISCAS 2024, Singapore
- Session Chair, IEEE ICASSP 2024, Korea
- Session Chair, IEEE MMSP 2023, France
News
- 2025.07: 2 papers are accepted by 33rd ACM International Conference on Multimedia (ACM Multimedia 2025), Dublin, Ireland
- 2025.06: 1 paper is accepted by IEEE Transactions on Multimedia (Corresponding Author)
- 2025.06: Invited to be Program Committee for International Conference on AI-Generated Content (AIGC 2025), Hangzhou
- 2024.04: Research on Gen-AI based image restoration is Under Revision at IEEE TPAMI
- 2024.12: 1 paper is accepted by ACM Transactions on Multim. Comput. Commun. Appl. (JCR Q1)
- 2024.11: 2 papers are accepted by IEEE Transactions on Multimedia (中科院一区TOP,JCR Q1)
- 2024.08: 1 paper is accepted by Computer Vision and Image Understanding(JCR Q1)
- 2024.05: I am appointed as Special Session Chair by IEEE ISCAS 2024, Singapore
- 2024.04: I am appointed as Special Session Chair by IEEE ICASSP 2024, Korea
- 2024.01: 2 papers are accepted by Optics Express and Optics Letters
- 2023.12: I am awarded the 全国金奖 in 中国国际大学生创新大赛
- 2023.11: I am awarded the 全国铜奖 in 中国博士后创新创业大赛
- 2023.10: I am awarded the 优胜奖 in 春晖杯中国留学人员创新创业大赛
- 2023.09: I am appointed as Special Session Chair by IEEE MMSP 2023, France
- 2023.09: 1 paper has been accepted by NeurIPS 2023, New Orleans, USA
Selected Publications
- T. Liu, K. Wu, C. Cai, Y. Wang, K. Yap, L. Chau, "Towards Blind Bitstream-corrupted Video Recovery: A Visual Foundation Model-driven Framework," in ACM MM 2025
- C. Cai, T. Liu, J. Gao, W. Liu, K. Wu, R. Wang, Y. Wang, S. Liew, "From Semantics, Scene to Instance-awareness: Distilling Foundation Model for Open-vocabulary Grounded Situation Recognition," in ACM MM 2025 (Corresponding Author)
- K. Wu, Z. Li, Y. Yang, Q. Liu, and X. Zhang, “End-to-end Deep Video Compression Based on Hierarchical Temporal Context Learning,” in IEEE Transactions on Multimedia, 2025.
- W. Liu, C. Cai, J. Gao, K. Wu**, Y. Wang, K. Yap, and L. Chau, “PromptSR: Cascade Prompting for Lightweight Image Super-Resolution,” in IEEE Transactions on Multimedia, 2025. (Corresponding Author)
- K. Wu, Y. Yang, G. Jiang, and X. Zhang, “Hierarchical Independent Coding Scheme for Varifocal Multiview Images based on Angular-focal Joint Prediction,” IEEE Transactions on Multimedia, 26:2993-3006, 2024. $\color{Tomato} {ESI Highly Cited Papers}$
- W. Liu, K. Wu**, T. Liu, Y. Wang, K. Yap, and L. Chau, “ByteNet: Rethinking Multimedia File Fragment Classification through Visual Perspectives,” in IEEE Transactions on Multimedia, 2024. (Corresponding Author)
- K. Wu, Z. Li, Y. Yang, and Q. Liu, “Deep Video Compression based on Long-range Temporal Context Learning,” in Computer Vision and Image Understanding, 248(2024): 104127.
- K. Wu, Q. Liu, and X. Zhang, “Focal Stack Image Compression Based on Basis-Quadtree Representation,” in IEEE Transactions on Multimedia, 25:3975-3988, 2023.
- K. Wu, Q. Liu, Y. Yang, and X. Zhang, “Gaussian-Wiener Representation and Hierarchical Coding Scheme for Focal Stack Images,” in IEEE Transactions on Circuits and Systems for Video Technology, 32(2):523-537, 2022.
- X. Yu, K. Wu**, Y. Yang, and Q. Liu, “WaRENet: A Novel Urban Waterlogging Risk Evaluation Network,” in ACM Transactions on Multimedia Computing, Communications, and Applications, 2024, 20(7):1–28. (Equal Contribution)
- K. Wu, Q. Liu, K. Yap, and Y. Yang, “Multifocal Multiview Imaging and Data Compression based on Angular-Focal-Spatial Representation,” in Optics Letters, 2024.
- K. Wu, Q. Liu, K. Yap, and Y. Yang, “High Dimensional Optical Data Varifocal Multiview Imaging, Compression and Evaluation,” in Optics Express, 2023.
- K. Wu, Q. Liu, Y. Wang, and Y. Yang, “End-to-end Varifocal Multiview Images Coding Framework from Data Acquisition End to Vision Application End,” in Optics Express, 31(7): 11659-11679, 2023.
- T. Liu, K. Wu**, Y. Wang, W. Liu, K. Yap, and L. Chau, “Bitstream corrupted Video Recovery: A Novel Benchmark Dataset and Method,” NeurIPS 2023. (Equal Contribution)
- J. Gao, K. Yap, K. Wu, D. Phan, and K. Garg, "Contextual Human Object Interaction Understanding from Pre-Trained Large Language Model," 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Korea
- C. Cai, R. Zhang, J. Gao, K. Wu, K. Yap, Y. Wang, "Temporal Sentence Grounding with Temporally Global Textual Knowledge," 2024 IEEE International Conference on Multimedia and Expo (ICME), Niagra Falls, Canada
- K. Wu, Y. Wang, W. Liu, K. Yap, and L. Chau, “A Spatial-Focal Error Concealment Scheme for Corrupted Focal Stack Video,” 2023 Data Compression Conference (DCC), Snowbird, UT, USA, Mar 2023.
Honors & Awards
- China International College Students’ Innovation Competition (formerly Internet +, Gold Award)
- IET Excellence and Innovation Awards International Awards (Silver Award)
- National Postdoctoral Innovation and Entrepreneurship Competition (Bronze Award, ranked first)
- Chunhui Cup Innovation and Entrepreneurship Competition for Chinese Overseas Students (ranked first)
- IET Impact in Society Awards (Shortlist)
- Letter of Appreciation from University of Oxford
Educations
- Exchange Ph.D Student, School of Electrical and Electronic Engineering, Nanyang Technological University
- Ph.D Student, School of Electronic Information and Communications, Huazhong University of Science and Technology
- Master Student, College of Intelligent Science and Engineering, Harbin Engineering University
- Undergraduate Student, School of Mechatronic Engineering and Automation, Shanghai University
Students
I’m honored to have participated in the supervision of these excellent students:
- J. Liang, Intern Student, Huazhong University of Science and Technology (LLMs on multimedia understanding)
- F. Li, Master Student, Huazhong University of Science and Technology (LLMs on Byte-domain understanding)
- C. Zhang, Ph.D Student, Huazhong University of Science and Technology (LLMs on signal processing)
- R. Wang, Ph.D Student, Harbin Engineering University (LLMs on fine-grained recognition)
- S. Wang, Ph.D Student, Harbin Engineering University (LLMs on SAR image anomaly detection)
- J. Gao, Ph.D Student, Nanyang Technological University (LLMs on HOI)
- W. Liu, Ph.D Student, Nanyang Technological University (Gen-AI on image restoration)
- T. Liu, Ph.D Student, Nanyang Technological University (Gen-AI on image restoration)
- C. Cai, Ph.D, National University of Singapore (Scientist, NUS, LLMs on remote sensing)