Zechuan Zhang (张泽川)

I am a third-year PhD student at Zhejiang University, supervised by Prof. Yi Yang (IEEE Fellow, 90,000+ citations). My research focuses on Computer Vision and Machine Learning, with an emphasis on 3D Vision, Multimodal LLMs, and AIGC. My open-source projects have attracted over 2,500 GitHub stars.

My academic excellence has been recognized through several honors, including the CHU KOCHEN Scholarship (Zhejiang University's highest honor), three China National Scholarships, and the First Prize in the China Graduate AI Competition (Ranked 4/2,228). I have consistently published at top-tier venues such as NeurIPS and CVPR, where my research was selected as a Highlight (top 3%).

Currently, I am a Visiting Research Scholar at the Broad Institute of Harvard and MIT and the University of Chicago, collaborating with Prof. Benjamin Neale (200,000+ citations) and Prof. Siwei Chen. I am dedicated to developing scalable ML frameworks for high-dimensional multimodal data, aiming to bridge cutting-edge AI research with complex real-world applications.

Email  |  CV  |  GitHub  |  Google Scholar  |  LinkedIn

headshot
Zhejiang University Broad Institute Harvard Medical School University of Chicago
News
  • 02/2026 Our paper IF-Edit is accepted by CVPR 2026!
  • 10/2025 I was selected as a TOP Reviewer in NeurIPS 2025!
  • 09/2025 Our paper ICEdit is accepted by NeurIPS 2025!
  • 06/2025 I've joined the Broad Institute of Harvard and MIT as a visiting scholar, working with Prof. Siwei Chen and Prof. Benjamin Neale.
  • 04/2025 We released ICEdit, a novel framework for instruction-guided image editing.
  • 10/2024 Achieved First Prize (Top 0.2%) in the China Graduate Artificial Intelligence Competition, ranking 4th among 2,228 teams across the country.
  • 04/2024 Our paper SIFU is selected as Highlight (top 12%) by CVPR2024!
  • 04/2024 Invited to give a talk (in Chinese) on Real-world Usable Clothed Human Reconstruction at SHUZIHUANYU.
  • 03/2024 One paper was accepted by CVPR 2024!
  • 09/2023 One paper was accepted by NeurIPS 2023!
  • ---- show more ----
Research
TRACE: High-Fidelity 3D Scene Editing via Tangible Reconstruction and Geometry-Aligned Contextual Video Masking
Jiyuan Hu, Zechuan Zhang, Zongxin Yang, Yi Yang
arXiv, 2026
paper | project page | code

We introduce TRACE, a mesh-guided 3DGS editing framework for high-fidelity scene transformation with tangible reconstruction and geometry-aligned contextual video masking.

Are Image-to-video Models good Zero-shot Image Editors?
Zechuan Zhang, Zhenyuan Chen, Zongxin Yang, Yi Yang
CVPR, 2026
paper

We introduce IF-Edit, a tuning-free framework that repurposes pretrained image-to-video diffusion models for instruction-driven image editing.

Enabling Instructional Image Editing with In-Context Generation in Large Scale Diffusion Transformer
Zechuan Zhang, Ji Xie, Yu Lu, Zongxin Yang, Yi Yang
Conference on Neural Information Processing Systems (NeurIPS), 2025
paper | arXiv | project page | code | Semantic Scholar | GitHub stars

We introduce ICEdit, an efficient and effective framework for instruction-based image editing. With only 1% trainable parameters (200M) and 0.1% training data (50k) compared to previous methods, ICEdit demonstrates strong generalization capabilities and is capable of handling diverse editing tasks. Compared with commercial models such as Gemini, GPT4o, we are more open-source, with lower costs, faster speed (it takes about 9 seconds to process one image), and powerful performance.

3D object manipulation in a single image using generative models
Ruisi Zhao, Zechuan Zhang, Zongxin Yang, Yi Yang
ArXiv, 2025
paper | project page

We introduce OMG3D, a novel framework that combines precise geometric control with the generative power of diffusion models, thus achieving significant improvements in visual performance.

SIFU: Side-view Conditioned Implicit Function for Real-world Usable Clothed Human Reconstruction
Zechuan Zhang, Zongxin Yang, Yi Yang,
Conference on Computer Vision and Pattern Recognition (CVPR), 2024 (Highlight, top 12%)
paper | project page | poster | code | media (AK) | [新智元]| GitHub stars

With just a single image, SIFU is capable of reconstructing a high-quality 3D clothed human model, making it well-suited for practical applications such as scene creation and 3D printing.

Global-correlated 3D-decoupling Transformer for Clothed Avatar Reconstruction
Zechuan Zhang, Li Sun, Zongxin Yang, Ling Chen, Yi Yang
Conference on Neural Information Processing Systems (NeurIPS), 2023
paper | project page | code | media (CVer)| GitHub stars

A transformer-based implicit function for single image clothed human reconstruction.

Academic Services
  • Conference Reviewer: ACM MM, NeurIPS, Siggraph Asia, CVPR
  • Journal Reviewer: TPAMI
Awards
  • Scholar Award of NeurIPS 2025 and Top Reviewer of NeurIPS25.
  • First Prize in the China Graduate Artificial Intelligence Competition (4/2228).
  • BYD Scholarship in the session of 2024.
  • CHU KOCHEN Scholarship in the session of 2022 (Highest Honor at Zhejiang University).
  • China National Scholarship for undergraduate in the session of 2020, 2021 and 2022.
  • First Class Scholarship of Zhejiang University in the session of 2020, 2021 and 2022.
  • Outstanding Winner in National Application of GIS Skill Competition in 2021.

template adapted from this awesome website