BEGIN:VCALENDAR VERSION:2.0 PRODID:Linklings LLC BEGIN:VTIMEZONE TZID:Asia/Tokyo X-LIC-LOCATION:Asia/Tokyo BEGIN:STANDARD TZOFFSETFROM:+0900 TZOFFSETTO:+0900 TZNAME:JST DTSTART:18871231T000000 END:STANDARD END:VTIMEZONE BEGIN:VEVENT DTSTAMP:20250110T023313Z LOCATION:Hall B5 (2)\, B Block\, Level 5 DTSTART;TZID=Asia/Tokyo:20241206T113100 DTEND;TZID=Asia/Tokyo:20241206T114300 UID:siggraphasia_SIGGRAPH Asia 2024_sess143_papers_1101@linklings.com SUMMARY:NPGA: Neural Parametric Gaussian Avatars DESCRIPTION:Technical Papers\n\nSimon Giebenhain and Tobias Kirschstein (T echnical University of Munich); Martin Rünz (Synthesia); Lourdes Agapito ( University College London (UCL), Synthesia); and Matthias Nießner (Technic al University of Munich, Synthesia)\n\nThe creation of high-fidelity, digi tal versions of human heads is an important stepping stone in the process of further integrating virtual components into our everyday lives. Constru cting such avatars is a challenging research problem, due to a high demand for photo-realism and real-time rendering performance. In this work, we p ropose Neural Parametric Gaussian Avatars (NPGA), a data-driven approach t o create high-fidelity, controllable avatars from multi-view video recordi ngs. We build our method around 3D Gaussian splatting for its highly effic ient rendering and to inherit the topological flexibility of point clouds. In contrast to previous work, we condition our avatars’ dynamics on the r ich expression space of neural parametric head models (NPHM), instead of m esh-based 3DMMs. To this end, we distill the backward deformation field of our underlying NPHM into forward deformations which are compatible with r asterization-based rendering. All remaining fine-scale, expression-depende nt details are learned from the multi-view videos. For increased represent ational capacity of our avatars, we propose per-Gaussian latent features t hat condition each primitives dynamic behavior. To regularize this increas ed dynamic expressivity, we propose Laplacian terms on the latent features and predicted dynamics. We evaluate our method on the public NeRSemble da taset, demonstrating that NPGA significantly outperforms the previous stat e-of-the-art avatars on the self-reenactment task by 2.6 PSNR. Furthermore , we demonstrate accurate animation capabilities from real-world monocular videos.\n\nRegistration Category: Full Access, Full Access Supporter\n\nL anguage Format: English Language\n\nSession Chair: Iain Matthews (Epic Gam es, Carnegie Mellon University) URL:https://asia.siggraph.org/2024/program/?id=papers_1101&sess=sess143 END:VEVENT END:VCALENDAR