BEGIN:VCALENDAR VERSION:2.0 PRODID:Linklings LLC BEGIN:VTIMEZONE TZID:Asia/Tokyo X-LIC-LOCATION:Asia/Tokyo BEGIN:STANDARD TZOFFSETFROM:+0900 TZOFFSETTO:+0900 TZNAME:JST DTSTART:18871231T000000 END:STANDARD END:VTIMEZONE BEGIN:VEVENT DTSTAMP:20250110T023312Z LOCATION:Hall B5 (2)\, B Block\, Level 5 DTSTART;TZID=Asia/Tokyo:20241204T165300 DTEND;TZID=Asia/Tokyo:20241204T170500 UID:siggraphasia_SIGGRAPH Asia 2024_sess122_papers_507@linklings.com SUMMARY:StyleTex: Style Image-Guided Texture Generation for 3D Models DESCRIPTION:Technical Papers\n\nZhiyu Xie, Yuqing Zhang, Xiangjun Tang, Yi qian Wu, and Dehan Chen (State Key Laboratory of CAD&CG, Zhejiang Universi ty); Gongsheng Li (Zhejiang University); and Xiaogang Jin (State Key Labor atory of CAD&CG, Zhejiang University)\n\nStyle-guided texture generation a ims to generate a texture that is harmonious with both the style of the re ference image and the geometry of the input mesh, given a reference style image and a 3D mesh with its text description. \nAlthough diffusion-based 3D texture generation methods, such as distillation sampling, have numero us promising applications in stylized games and films, it requires address ing two challenges: 1) decouple style and content completely from the refe rence image for 3D models, and 2) align the generated texture with the col or tone, style of the reference image, and the given text prompt.\nTo this end, we introduce StyleTex, an innovative diffusion-model-based framework for creating stylized textures for 3D models. Our key insight is to decou ple style information from the reference image while disregarding content in diffusion-based distillation sampling.\nSpecifically, given a reference image, we first decompose its style feature from the image CLIP embedding by subtracting the embedding's orthogonal projection in the direction of the content feature, which is represented by a text CLIP embedding. \nOur novel approach to disentangling the reference image's style and content in formation allows us to generate distinct style and content features. \nWe then inject the style feature into the cross-attention mechanism to incorp orate it into the generation process, while utilizing the content feature as a negative prompt to further dissociate content information. \nFinally, we incorporate these strategies into StyleTex to obtain stylized textures . We utilize Interval Score Matching to address over-smoothness and over-s aturation, in combination with a geometry-aware ControlNet that ensures co nsistent geometry throughout the generative process. The resulting texture s generated by StyleTex retain the style of the reference image, while als o aligning with the text prompts and intrinsic details of the given 3D mes h.\nQuantitative and qualitative experiments show that our method outperfo rms existing baseline methods by a significant margin.\n\nRegistration Cat egory: Full Access, Full Access Supporter\n\nLanguage Format: English Lang uage\n\nSession Chair: Minhyuk Sung (Korea Advanced Institute of Science a nd Technology (KAIST)) URL:https://asia.siggraph.org/2024/program/?id=papers_507&sess=sess122 END:VEVENT END:VCALENDAR