BEGIN:VCALENDAR VERSION:2.0 PRODID:Linklings LLC BEGIN:VTIMEZONE TZID:Asia/Tokyo X-LIC-LOCATION:Asia/Tokyo BEGIN:STANDARD TZOFFSETFROM:+0900 TZOFFSETTO:+0900 TZNAME:JST DTSTART:18871231T000000 END:STANDARD END:VTIMEZONE BEGIN:VEVENT DTSTAMP:20250110T023313Z LOCATION:Hall B7 (1)\, B Block\, Level 7 DTSTART;TZID=Asia/Tokyo:20241206T130000 DTEND;TZID=Asia/Tokyo:20241206T141000 UID:siggraphasia_SIGGRAPH Asia 2024_sess147@linklings.com SUMMARY:(Do) Make Some Noise DESCRIPTION:Technical Papers\n\nEach Paper gives a 10 minute presentation. \n\nDance-to-Music Generation with Encoder-based Textual Inversion\n\nThe seamless integration of music with dance movements is essential for commun icating the artistic intent of a dance piece. This alignment also signific antly improves the immersive quality of gaming experiences and animation p roductions. Although there has been remarkable advancement in creating hig ...\n\n\nSifei Li, Weiming Dong, and Yuxin Zhang (MAIS, Institute of Autom ation, Chinese Academy of Sciences; School of Artificial Intelligence, Uni versity of Chinese Academy of Sciences); Fan Tang (University of Chinese A cademy of Sciences); Chongyang Ma (Kuaishou Technology); Oliver Deussen (U niversity of Konstanz); Tong-Yee Lee (National Cheng-Kung University); and Changsheng Xu (MAIS, Institute of Automation, Chinese Academy of Sciences ; School of Artificial Intelligence, University of Chinese Academy of Scie nces)\n---------------------\nSketching With Your Voice: "Non-Phonorealist ic" Rendering of Sounds via Vocal Imitation\n\nWe present a method for aut omatically producing human-like vocal imitations of sounds: the equivalent of ``sketching,'' but for auditory rather than visual representation. Sta rting with a simulated model of the human vocal tract, we first try genera ting vocal imitations by tuning the model's control...\n\n\nMatthew Caren, Kartik Chandra, Joshua Tenenbaum, Jonathan Ragan-Kelley, and Karima Ma (M assachusetts Institute of Technology)\n---------------------\nSIGGesture: Generalized Co-Speech Gesture Synthesis via Semantic Injection with Large- Scale Pre-Training Diffusion Models\n\nThe automated synthesis of high-qua lity 3D gestures from speech holds significant value for virtual humans an d gaming. Previous methods primarily focus on synchronizing gestures with speech rhythm, often neglecting semantic gestures. These semantic gestures are sparse and follow a long-tailed distri...\n\n\nQingrong Cheng (Tencen t AI Lab, Tencent TIMI L1 Studio) and Xu Li and Xinghui Fu (Tencent AI Lab )\n---------------------\nWaveBlender: Practical Sound-Source Animation in Blended Domains\n\nSynthesizing plausible sound sources for modern physic s-based animation is exceptionally challenging due to complex animated phe nomena that form rapidly moving, deforming, and vibrating interfaces which produce acoustic waves within the air domain. Not only must the methods s ynthesize sounds that ar...\n\n\nKangrui Xue (Stanford University); Jui-Hs ien Wang and Timothy Langlois (Adobe Research); and Doug James (Stanford U niversity, NVIDIA)\n---------------------\nSpeed-Aware Audio-Driven Speech Animation using Adaptive Windows\n\nWe present a novel method that can ge nerate realistic speech animations of a 3D face from audio using multiple adaptive windows. In contrast to previous studies that use a fixed size au dio window, our method accepts an adaptive audio window as input, reflecti ng the audio speaking rate to use consist...\n\n\nSunjin Jung (KAIST, Visu al Media Lab); Yeongho Seol (NVIDIA); Kwanggyoon Seo and Hyeonho Na (KAIST , Visual Media Lab); Seonghyeon Kim (KAIST, Visual Media Lab; Anigma Techn ologies); and Vanessa Tan and Junyong Noh (KAIST, Visual Media Lab)\n\nReg istration Category: Full Access, Full Access Supporter\n\nLanguage Format: English Language\n\nSession Chair: Yi Zhou (Adobe) END:VEVENT END:VCALENDAR