Talk

Relightable and Animatable, Neural Avatar from Sparse-View Video

avatar
Zhen Xu
2023/9/6
Neural Avatar
NeRF
main image for this paper

Abstract

Reconstructing relightable and animatable neural avatar from sparse-view (or monocular) video. Our method takes only a sparse-view (or monocular) video as input and reconstructs a relightable and animatable neural avatar under unknown illumination, which can then be relit with arbitrary environment lights and animated with arbitrary motion sequences. Note that our method successfully captures the shininess of the skin and pants as well as the specular highlights on the t-shirt’s plastisol printings.

Speaker

Zhen Xu is a second-year Ph.D. candidate at Zhejiang University, where he is deeply engaged in the study of Computer Graphics/Vision, with a particular interest in 3D technologies. Despite the vastness of the internet, Xu is delighted to connect with those who find and take an interest in his work, emphasizing a mutual care and appreciation for his audience.

His passion for the field is evident through his active involvement in managing private repositories on GitHub, which mainly consist of course and research projects. Currently, Xu is dedicating his research efforts towards 3D/4D neural reconstruction and rendering, as well as digital humans, demonstrating a commitment to advancing knowledge in these cutting-edge areas.

Video

Extra Details

Paper Link