Zeming Chen

Zeming Chen

Researcher in Multi-view Generation & Autonomous Driving

Email: czm369gg@gmail.com

GitHub: github.com/Czm369

Biography

I work on generative modeling and 3D perception for autonomous driving. My research focuses on using AIGC techniques such as diffusion models and autoencoders to construct world models from multi-view images. I am also interested in vision-language models (VLM) and vision-language-action (VLA) models for end-to-end autonomous driving, aiming to bridge perception, reasoning, and planning.

Education

Experience

OpenMMLab, Shanghai AI Lab (2021.11–2023.03)

Research

BEV-VAE: Multi-view Image Generation with Spatial Consistency for Autonomous Driving

BEV-VAE structure
Overview of the BEV-VAE framework with spatially consistent generation.

MixPL: Semi-supervised Object Detection with Mixed Pseudo Labels

BEV-VAE structure
The framework of DetMeanTeacher and MixPL for Semi-supervised Object Detection.

Contact

I'm open to collaborations in world model and end-to-end autonomous driving. Feel free to reach out via email or connect through GitHub.