BEGIN:VCALENDAR VERSION:2.0 PRODID:Linklings LLC BEGIN:VTIMEZONE TZID:Asia/Tokyo X-LIC-LOCATION:Asia/Tokyo BEGIN:STANDARD TZOFFSETFROM:+0900 TZOFFSETTO:+0900 TZNAME:JST DTSTART:18871231T000000 END:STANDARD END:VTIMEZONE BEGIN:VEVENT DTSTAMP:20250110T023311Z LOCATION:Hall B7 (1)\, B Block\, Level 7 DTSTART;TZID=Asia/Tokyo:20241203T153100 DTEND;TZID=Asia/Tokyo:20241203T154300 UID:siggraphasia_SIGGRAPH Asia 2024_sess108_papers_373@linklings.com SUMMARY:LVCD: Reference-based Lineart Video Colorization with Diffusion Mo dels DESCRIPTION:Technical Papers\n\nZhitong Huang (City University of Hong Kon g); Mohan Zhang (Wechat, Tencent Inc.); and Jing Liao (City University of Hong Kong)\n\nWe propose the first video diffusion framework for reference -based lineart video colorization. Unlike previous works that rely solely on image generative models to colorize lineart frame by frame, our approac h leverages a large-scale pretrained video diffusion model to generate col orized animation videos. This approach leads to more temporally consistent results and is better equipped to handle large motions. Firstly, we intro duce Sketch-guided ControlNet which provides additional control to finetun e an image-to-video diffusion model for controllable video synthesis, enab ling the generation of animation videos conditioned on lineart. We then pr opose Reference Attention to facilitate the transfer of colors from the re ference frame to other frames containing fast and expansive motions. Final ly, we present a novel scheme for sequential sampling, incorporating the O verlapped Blending Module and Prev-Reference Attention, to extend the vide o diffusion model beyond its original fixed-length limitation for long vid eo colorization. Both qualitative and quantitative results demonstrate tha t our method significantly outperforms state-of-the-art techniques in term s of frame and video quality, as well as temporal consistency. Moreover, o ur method is capable of generating high-quality, long temporal-consistent animation videos with large motions, which is not achievable in previous w orks. Our code and model are available at https://luckyhzt.github.io/lvcd. \n\nRegistration Category: Full Access, Full Access Supporter\n\nLanguage Format: English Language\n\nSession Chair: I-Chao Shen (The University of Tokyo) URL:https://asia.siggraph.org/2024/program/?id=papers_373&sess=sess108 END:VEVENT END:VCALENDAR