Zhipeng Zhang 张志鹏
Dr. Zhipeng Zhang is currently a Tenure-Track Assistant Professor in the School of Artificial Intelligence at Shanghai Jiao Tong University since April 2025. Prior to joining SJTU, he served as a Senior Researcher at KargBot from July 2022 to March 2025, leading key AI projects in autonomous driving.
He earned his Ph.D. from the National Laboratory of Pattern Recognition (NLPR) at the Chinese Academy of Sciences (CASIA) in 2022 under the supervision of Prof. Weiming Hu. During his doctoral studies, he completed a research internship at MSRA working closely with Dr. Houwen Peng, focusing on advanced computer vision and deep learning. Upon graduation, he was awarded the "Huawei Genius Young Talent".
His recent research interests include: (1) Multimodal Perception & Video Understanding; (2) Autonomous Driving (BEV, E2E); (3) Vision-Language-Action (VLA) in Embodied AI; (4) 3D Scene Reconstruction; (5) Generative AI; (6) Model Quantization for Edge Devices.
Email /
Google Scholar
|
|
Hiring
We are recruiting undergraduate or graduate research assistants for students applying to Ph.D./master programs starting in Fall 2026. (2026年保研直博/硕士欢迎联系,另长期招聘科研实习生,可远程或来上交) If you are passionate about AI and eager to contribute to cutting-edge research, don’t hesitate to contact me (see email above). Although I just joined SJTU this year, one of my Ph.D. whom I mentored for four years recently also obtained Huawei’s Genius Young Talent Program (华为天才少年). I am committed to supporting exceptional individuals like you and helping you achieve your academic and professional goals.
|
News
- [2025.02] 🎉 3 papers are accepted by CVPR2025 (2 corresponding/1 co-first author).
- [2025.02] 🎉 1 paper is accepted by ICRA2025 (first author).
- [2024.09] 🎉 1 paper is accepted by NeurIPS2024.
- [2024.07] 🎉 1 paper is accepted by ECCV2024.
- [2024.07] 🎉 1 paper is accepted by IJCV (corresponding author).
- [2024.06] 🎉 1 paper is accepted by T-PAMI (corresponding author).
- [2024.02] 🎉 1 paper is accepted by CVPR2024 (corresponding author).
- [2023.09] 🎉 1 paper is accepted as Oral by BMVC.
- [2023.02] 🎉 1 paper is accepted by CVPR2023 (co-first author).
- [2022.09] 🎉 2 papers are accepted by NeurIPS2022 (1 co-first author).
- [2022.09] 🎉 Our method won the runner-up of task 1-1 and 2nd runner-up of task 1-2 in 1st Learning and Mining with Noisy Labels Challenge.
- [2022.03] 🎉 1 paper is accepted as Oral by IJCAI2022 (co-first author).
- [2021.12] 🎉 1 paper is accepted by AAAI2022 (co-first author).
- [2021.10] 🎉 1 paper is accepted by IEEE T-IP (first author).
- [2021.09] 🎉 1 paper is accepted by IEEE T-IP (co-first author).
- [2021.07] 🎉 1 paper is accepted by ICCV2021 (first author).
- [2021.02] 🎉 1 paper is accepted by CVPR2021.
- [2020.12] 🎉 1 paper is accepted by IEEE T-IP.
- [2020.09] 🎉 Our method ranks at 2nd in short-term and real-time/1st in RGBT tracks of VOT Challenge.
- [2020.07] 🎉 1 paper is accepted by ECCV2020 (first author).
- [2019.09] 🎉 Our method ranks at 2nd in RGBT tracks of VOT Challenge.
- [2019.02] 🎉 1 paper is accepted as Oral by CVP2019 (first author).
|
|
CorrBEV: Multi-View 3D Object Detection by Correlation Learning with Multi-modal Prototypes
Ziteng Xue, Mingzhe Guo, Heng Fan, Shihui Zhang, Zhipeng Zhang✉
CVPR, 2025
CorrBEV improves BEV detection methods in autonomous driving by introducing vision-language multimodal prototypes.
|
|
Evolving High-Quality Rendering and Reconstruction in a Unified Framework with Contribution-Adaptive Regularization
You Shen, Zhipeng Zhang*, Xinyang Li, Yansong Qu, Yu Lin, Shengchuan Zhang, Liujuan Cao
CVPR, 2025
CarGS simultaneously achieves promising performances in both scene reconstruction and novel view synthesis with a unified model, improving the quality of 3DGS.
|
|
DreamTrack: Dreaming the Future for Multimodal Visual Object Tracking
Mingzhe Guo, Weiping Tan, Wenyu Ran, Liping Jing, Zhipeng Zhang✉
CVPR, 2025
DreamTrack shows the best performances in visual tracking by dreaming the future presentation with latent world model.
|
|
Cyclic Refiner: Object-Aware Temporal Representation Learning for Multi-View 3D Detection and Tracking
Mingzhe Guo, Zhipeng Zhang*✉, Liping Jing, Yuan He, Ke Wang, Heng Fan
IJCV
Cycer reduces false positives in BEV detection of autonomous driving by propagating results of t - 1 frame to t, which generates a mask to filter distractors in BEV representation.
|
|
A-Teacher: Asymmetric Network for 3D Semi-Supervised Object Detection
Hanshi Wang, Zhipeng Zhang✉, Jin Gao, Weiming Hu
CVPR, 2023
A-Teachers proposes the first online asymmetric framework for semi-supervised 3D LiDAR detection.
|
|
End-to-End Autonomous Driving without Costly Modularization and 3D Manual Annotation
Mingzhe Guo, Zhipeng Zhang✉, Yuan He, Ke Wang, Liping Jing, Haibin Ling
Arxiv
UAD proposes the first work demonstrating that an unsupervised model can outperform supervised End-to- End autonomous driving method.
|
|
VastTrack: Vast Category Visual Object Tracking
Liang Peng, Junyuan Gao, Xinran Liu, Weihong Li, Shaohua Dong, Zhipeng Zhang, Heng Fan, Libo Zhang
NeurIPS, 2024
VAST is the largest visual tracking benchmark to date.
|
|
AUNet: Learning Relations Between Action Units for Face Forgery Detection
Weiming Bai, Yufan Liu*, Zhipeng Zhang*, Bing Li, Weiming Hu
CVPR, 2023
AUNet proposes the Action-Units Relation Learning framework to improve the generality of forgery (deepfake) detection.
|
|
Ocean: Object-aware Anchor-free Tracking
Zhipeng Zhang, Houwen Peng, Jianlong Fu, Bing Li, Weiming Hu
ECCV, 2020 (Cite 900+)
Ocean explores an efficient anchor-free framework to improve object tracking robustness.
|
|
Deeper and Wider Siamese Networks for Real-Time Visual Tracking
Zhipeng Zhang, Houwen Peng
CVPR, 2019 (Oral, Cite 1200+)
SiamDW is the first work to solve the performance degradation in the Siamese tracking framework when using a deeper network.
|
- Conference Reviewer: CVPR, ICCV, ECCV, ICML, ICLR, NeurIPS, AAAI.
- Journal Reviewer: T-PAMI, IJCV, TIP, et.al.
- Invited Talk: SiamDW in CVPR2019 (极市平台)
- Invited Talk: Ocean in ECCV2020 (极市平台)
- Award: National Scholarship for Ph.D.
- Award: National Scholarship for undergraduate.
|