Zhi Li (PhD student)

MSc Zhi Li
- Address
- Max-Planck-Institut für Informatik
Saarland Informatics Campus
Campus E1 4
66123 Saarbrücken - Standort
- E1 4 - 626
- Telefon
- +49 681 9325 2143
- Fax
- +49 681 9325 2099
- Get email via email
Personal Information
Homepage | Github | Google Scholar | LinkedIn
About Me
I am currently a PhD student in Department of Computer Vision and Machine Learning at Max Planck Institute for Informatics, advised by Prof. Dr. Bernt Schiele. My research interests include computer vision for autonomous driving especially image perception under domain shift, and 3D computer vision.
Education/ Research Experience
- Jun. 2021 - Present: PhD student in Computer Vision and Machine Learning, Max Planck Institute for Informatics, Germany (Advisor: Prof. Dr. Bernt Schiele)
- Dec. 2021 - May 2021: Research intern in Computer Vision and Machine Learning, Max Planck Institute for Informatics, Germany (Advisors: Prof. Dr. Bernt Schiele and Dr. Dengxin Dai)
- Dec. 2020 - Nov. 2021: Research intern in Computer Vision and Machine Learning, Max Planck Institute for Informatics, Germany (Advisors: Prof. Dr. Bernt Schiele and Prof. Dr. Christian Theobalt)
- Sep. 2017 - Jun. 2020: MSc in Software Engineering, Xi'an Jiaotong University, China
- Sep. 2013 - Jun. 2017: BA in English Literature, Xi'an Jiaotong University, China
Publications (for a full list of publications please click)
2022
HULC: 3D HUman Motion Capture with Pose Manifold SampLing and Dense Contact Guidance
S. Shimada, V. Golyanik, Z. Li, P. Pérez, W. Xu and C. Theobalt
Computer Vision -- ECCV 2022, 2022
S. Shimada, V. Golyanik, Z. Li, P. Pérez, W. Xu and C. Theobalt
Computer Vision -- ECCV 2022, 2022
MoCapDeform: Monocular 3D Human Motion Capture in Deformable Scenes
Z. Li, S. Shimada, B. Schiele, C. Theobalt and V. Golyanik
International Conference on 3D Vision, 2022
Z. Li, S. Shimada, B. Schiele, C. Theobalt and V. Golyanik
International Conference on 3D Vision, 2022
Abstract
3D human motion capture from monocular RGB images respecting interactions of<br>a subject with complex and possibly deformable environments is a very<br>challenging, ill-posed and under-explored problem. Existing methods address it<br>only weakly and do not model possible surface deformations often occurring when<br>humans interact with scene surfaces. In contrast, this paper proposes<br>MoCapDeform, i.e., a new framework for monocular 3D human motion capture that<br>is the first to explicitly model non-rigid deformations of a 3D scene for<br>improved 3D human pose estimation and deformable environment reconstruction.<br>MoCapDeform accepts a monocular RGB video and a 3D scene mesh aligned in the<br>camera space. It first localises a subject in the input monocular video along<br>with dense contact labels using a new raycasting based strategy. Next, our<br>human-environment interaction constraints are leveraged to jointly optimise<br>global 3D human poses and non-rigid surface deformations. MoCapDeform achieves<br>superior accuracy than competing methods on several datasets, including our<br>newly recorded one with deforming background scenes.<br>
2021