BEGIN:VCALENDAR
VERSION:2.0
PRODID:Linklings LLC
BEGIN:VTIMEZONE
TZID:Asia/Tokyo
X-LIC-LOCATION:Asia/Tokyo
BEGIN:STANDARD
TZOFFSETFROM:+0900
TZOFFSETTO:+0900
TZNAME:JST
DTSTART:18871231T000000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTAMP:20250110T023312Z
LOCATION:Hall B5 (2)\, B Block\, Level 5
DTSTART;TZID=Asia/Tokyo:20241204T110800
DTEND;TZID=Asia/Tokyo:20241204T111900
UID:siggraphasia_SIGGRAPH Asia 2024_sess113_papers_253@linklings.com
SUMMARY:Cafca: High-quality Novel View Synthesis of Expressive Faces from 
 Casual Few-shot Captures
DESCRIPTION:Technical Papers\n\nMarcel C. Buehler and Gengyan Li (ETH Züri
 ch, Google VR); Erroll Wood, Leonhard Helminger, Xu Chen, Tanmay Shah, Dao
 ye Wang, Stephan Garbin, and Sergio Orts Escolano (Google VR); Otmar Hilli
 ges (ETH Zürich); and Dmitry Lagun, Jérémy Riviere, Paulo Gotardo, Thabo B
 eeler, Abhimitra Meka, and Kripasindhu Sarkar (Google VR)\n\nVolumetric mo
 deling and neural radiance field representations have revolutionized 3D fa
 ce capture and photorealistic novel view synthesis. However, these methods
  often require hundreds of multi-view input images and are thus inapplicab
 le to cases with less than a handful of inputs.\nWe present a novel volume
 tric prior on human faces that allows for high-fidelity expressive face mo
 deling from as few as three input views captured in the wild. Our key insi
 ght is that an implicit prior trained on synthetic data alone can generali
 ze to extremely challenging real-world identities and expressions and rend
 er novel views with fine idiosyncratic details like wrinkles and eyelashes
 .\nWe leverage a 3D Morphable Face Model to synthesize a large training se
 t, rendering each identity with different expressions, hair, clothing, and
  other assets. We then train a conditional Neural Radiance Field prior on 
 this synthetic dataset and, at inference time, fine-tune the model on a ve
 ry sparse set of real images of a single subject. On average, the fine-tun
 ing requires only three inputs to cross the synthetic-to-real domain gap. 
 The resulting personalized 3D model reconstructs strong idiosyncratic faci
 al expressions and outperforms the state-of-the-art in high-quality novel 
 view synthesis of faces from sparse inputs in terms of perceptual and phot
 o-metric quality.\n\nRegistration Category: Full Access, Full Access Suppo
 rter\n\nLanguage Format: English Language\n\nSession Chair: Forrester Cole
  (Google)
URL:https://asia.siggraph.org/2024/program/?id=papers_253&sess=sess113
END:VEVENT
END:VCALENDAR
