He was born in 1996 in Henan, China.
He graduated from Dalian University of Technology (DUT) in 2018 with a bachelor's degree in computer science
and technology (Outstanding Graduates of Liaoning Province), supervised by Prof. YANG, Xin.
He obtained his Ph.D. degree in Computer Science and Engineering from The Chinese University of Hong Kong
(CUHK) in 2022, supervised by Prof. WONG, Tien-Tsin.
From 2022 to 2023, he was a researcher and developer at
PICO Mixed Reality, Bytedance.
Since November 2024, he joined Tencent as a senior researcher, leading an effort to Generative World
Model, including 3D from Images/Videos, Novel View Synthesis, and Video Generation.
His work has been selected as Best Paper Finalist in ICCV'2023.
He received the CCF Elite Collegiate Award in 2017.
He has served as a reviewer for top-tier conferences and journals, including SIGGRAPH, SIGGRAPH Asia, CVPR,
ICCV, ECCV, NeurIPS, ICML, EG, TVCG, IJCV, etc.
We have several open positions for research interns and full-time researchers. Feel free to drop me an email if you are interested.
News!
[03/2025]
Invited talk by AnySyn3D about "Video Diffusion for 3D".
[03/2025]
We released code and models for TrajectoryCrafter.
[02/2025]
Three papers accepted to CVPR'25.
[10/2024]
Invited talks by Chinagraph about:
[DepthCrafter, StereoCrafter,
ViewCrafter] and
[Tri-MipRF, Rip-NeRF, Analytic-Splatting].
[09/2024]
One paper accepted to NeurIPS'24.
[07/2024]
Four papers accepted to ECCV'24.
[05/2024]
Invited talk by NeRF/GS & Beyond about Anti-Aliasing in Neural Rendering. [Slides]
[04/2024]
Invited talk by China3DV 2024 about Anti-Aliasing in Neural Rendering.
[03/2024]
One paper conditionally accepted to SIGGRAPH'24.
[02/2024]
One paper accepted to CVPR'24.
[10/2023]
Our Tri-MipRF was selected in ICCV'23 Best Paper Finalist!
[08/2023]
Invited talk by GAMES about ourTri-MipRF.
[07/2023]
Our Tri-MipRF was accepted to ICCV'23 as ORAL presentation.
[07/2023]
One paper on invertile image downscaling accepted to TIP'23.
[08/2022]
One paper conditionally accepted to SIGGRAPH Asia'22 (Journal Track).
[05/2022]
Passed the oral defense and became a Dr.
[08/2021]
Invited talk by 智东西about ourBPNet.
[07/2021]
One paper on lighting estimation accepted to TIP'22.
[07/2021]
Two papers accepted to ICCV'21.
[07/2021]
One paper on 3D human pose estimation accepted to ACM MM'21.
[06/2021]
Invited talk at ScanNet CVPR'21 Workshop about
our BPNet (Recording).
[06/2021]
Code of BPNet is released now.
[03/2021]
One paper conditionally accepted to CVPR 2021 (Oral).
[08/2020]
One paper conditionally accepted to SIGGRAPH Asia 2020.
[08/2018]
I start my Ph.D study at CUHK.
[01/2018] I will work as a research intern at SenseTime.
Selected, full list on Google
Scholar
(*equal contribution, †corresponding author)


Wenbo Hu* †, Xiangjun Gao*, Xiaoyu Li* †, Sijie Zhao, Xiaodong Cun, Yong Zhang, Long Quan, Ying Shan
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2025. [ Project ]   [ Paper ]   [ Code ]   [ Demo ]  


Wangbo Yu*, Jinbo Xing*, Li Yuan*, Wenbo Hu†, Xiaoyu Li, Zhipeng Huang, Xiangjun Gao, Tien-Tsin Wong, Ying Shan, Yonghong Tian†
arXiv preprint, 2024. [ Project ]   [ Paper ]   [ Video ]   [ Code ]   [ Demo ]  






Wenbo Hu, Yuling Wang, Lin Ma, Bangbang Yang, Lin Gao, Xiao Liu, Yuewen Ma
IEEE International Conference on Computer Vision (ICCV), 2023.   Oral Presentation [ Project ]   [ Paper ]   [ Code ]   Best paper finalist (17/8260)





Education

Aug. 2018 - May. 2022
 
The Chinese University of Hong Kong
 
Dept of Computer Science & Engineering
Ph.D.

Sep. 2014 - Jun. 2018
 
Dalian University of Technology
 
Dept of Computer Science and Technology
Bachelor (Rank: 3/105, GPA: 90.3/100)
Work Experience

Dec. 2023 - Present
Tencent AI Lab
Researcher at CVC

Sep. 2022 - Nov. 2023
PICO Mixed Reality, ByteDance
Researcher & Developer

Nov. 2021 - May. 2022
Tencent AI Lab
Research Intern

Sep. 2020 - Nov. 2021
DAMO Academy, Alibaba Group
Research Intern

Jan. 2018 - Jul. 2018
SenseTime
Research Intern

Jul. 2016 - Sep. 2016
DJI
Summer Intern
- CN108827302A (In process), “Multi-rotor aerocraft air navigation aid based on rotor tachometric survey”.
- CN107656227B (Issued Oct. 2019), “Magnetometer calibration method based on Levenberg-Marquardt algorithm”.
- CN107655463B (Issued Oct. 2019), “Electronic compass calibration method based on simulated annealing”.
- CN207400517U (Issued May. 2018), “Intelligent toothbrush system ”.
Miscellaneous
Excited to share our #TrajectoryCrafter, a diffusion model for Redirecting Camera Trajectory in Monocular Videos!
— HU, Wenbo (@wbhu_cuhk) March 10, 2025
Try to explore the world underlying your videos~
Page: https://t.co/hWuDRDcv10
Demo: https://t.co/e3JF0SSXIC
Code: https://t.co/Y84MN6D2iO pic.twitter.com/WAS9Vbbosu
Introducing 𝚅𝚒𝚎𝚠𝙲𝚛𝚊𝚏𝚝𝚎𝚛 🥳. 𝚅𝚒𝚎𝚠𝙲𝚛𝚊𝚏𝚝𝚎𝚛 can generate high-fidelity novel views from single or sparse input images with accurate camera pose control!
— Jinbo Xing (@Double47685693) September 5, 2024
✨Paper: https://t.co/dH4Dw0Eb1e
🎯Code: https://t.co/53ai21Px99
🥁Demo: https://t.co/9xCyqtsFco pic.twitter.com/R5ZZpYfdM3
I ported DepthCrafter to ComfyUI! 🔥
— akatz (@akatz_ai) October 18, 2024
Now you can generate super stable depthmap videos from any input video.
The VRAM requirement is pretty high (>16GB) if you want to render long videos in high res (768p and up)
It pairs well with Depthflow!
Repo link in comments below👇 pic.twitter.com/dlS5wr7mRU
Introducing StereoCrafter: Transforming monocular videos into high-fidelity 3D movies, compatible with various depth estimation methods and currently performing best with DepthCrafter. Feel free to download and experience the 3D results on our project page using Vision Pros,… pic.twitter.com/6YEHisStz8
— Ying Shan (@yshan2u) September 12, 2024
Excited to share our DepthCrafter, a super consistent video depth model for long open-world videos!
— HU, Wenbo (@wbhu_cuhk) September 4, 2024
Project webpage: https://t.co/9SiMUv4hoW https://t.co/qy55L7cm44 pic.twitter.com/TZBWnjEGmk
— Kosta Derpanis (@CSProfKGD) October 4, 2023