ShanghaiTech University Knowledge Management System
Video-driven Neural Physically-based Facial Asset for Production | |
2022-12-01 | |
Source Publication | ACM TRANSACTIONS ON GRAPHICS
![]() |
ISSN | 0730-0301 |
EISSN | 1557-7368 |
Volume | 41Issue:6 |
Status | 已发表 |
DOI | 10.1145/3550454.3555445 |
Abstract | Production-level workflows for producing convincing 3D dynamic human faces have long relied on an assortment of labor-intensive tools for geometry and texture generation, motion capture and rigging, and expression synthesis. Recent neural approaches automate individual components but the corresponding latent representations cannot provide artists with explicit controls as in conventional tools. In this paper, we present a new learning-based, video-driven approach for generating dynamic facial geometries with high-quality physically-based assets. For data collection, we construct a hybrid multiview-photometric capture stage, coupling with ultra-fast video cameras to obtain raw 3D facial assets. We then set out to model the facial expression, geometry and physically-based textures using separate VAEs where we impose a global MLP based expression mapping across the latent spaces of respective networks, to preserve characteristics across respective attributes. We also model the delta information as wrinkle maps for the physically-based textures, achieving high-quality 4K dynamic textures. We demonstrate our approach in high-fidelity performer-specific facial capture and cross-identity facial motion retargeting. In addition, our multi-VAE-based neural asset, along with the fast adaptation schemes, can also be deployed to handle in-the-wild videos. Besides, we motivate the utility of our explicit facial disentangling strategy by providing various promising physically-based editing results with high realism. Comprehensive experiments show that our technique provides higher accuracy and visual fidelity than previous video-driven facial reconstruction and animation methods. |
Keyword | Physically-Based Face Rendering Facial Modeling Digital Human Video-Driven Animation |
URL | 查看原文 |
Indexed By | SCI ; EI |
Language | 英语 |
Funding Project | Shanghai YangFan Program[21YF1429500] ; Shanghai Local College Capacity Building Program[22010502800] ; NSFC programs[ |
WOS Research Area | Computer Science |
WOS Subject | Computer Science, Software Engineering |
WOS ID | WOS:000891651900028 |
Publisher | ASSOC COMPUTING MACHINERY |
Citation statistics | |
Document Type | 期刊论文 |
Identifier | https://kms.shanghaitech.edu.cn/handle/2MSLDSTB/266605 |
Collection | 信息科学与技术学院_博士生 信息科学与技术学院_PI研究组_虞晶怡组 信息科学与技术学院_硕士生 信息科学与技术学院_本科生 信息科学与技术学院_PI研究组_许岚组 |
Co-First Author | Zeng, Chuxiao; Zhang, Qixuan |
Corresponding Author | Xu, Lan; Yu, Jingyi |
Affiliation | 1.ShanghaiTech Univ, Shanghai, Peoples R China 2.Deemos Technol Co Ltd, Shanghai, Peoples R China 3.Huazhong Univ Sci & Technol, Wuhan, Peoples R China 4.Shanghai Engn Res Ctr Intelligent Vis & Imaging, Shanghai, Peoples R China |
First Author Affilication | ShanghaiTech University |
Corresponding Author Affilication | ShanghaiTech University |
First Signature Affilication | ShanghaiTech University |
Recommended Citation GB/T 7714 | Zhang, Longwen,Zeng, Chuxiao,Zhang, Qixuan,et al. Video-driven Neural Physically-based Facial Asset for Production[J]. ACM TRANSACTIONS ON GRAPHICS,2022,41(6). |
APA | Zhang, Longwen.,Zeng, Chuxiao.,Zhang, Qixuan.,Lin, Hongyang.,Cao, Ruixiang.,...&Yu, Jingyi.(2022).Video-driven Neural Physically-based Facial Asset for Production.ACM TRANSACTIONS ON GRAPHICS,41(6). |
MLA | Zhang, Longwen,et al."Video-driven Neural Physically-based Facial Asset for Production".ACM TRANSACTIONS ON GRAPHICS 41.6(2022). |
Files in This Item: | Download All | |||||
File Name/Size | DocType | Version | Access | License |
Edit Comment
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.