논문 상세보기

Artistic character generation technique using a controllable diffusion model KCI 등재

  • 언어ENG
  • URLhttps://db.koreascholar.com/Article/Detail/427714
구독 기관 인증 시 무료 이용이 가능합니다. 4,000원
한국컴퓨터게임학회 논문지 (Journal of The Korean Society for Computer Game)
한국컴퓨터게임학회 (Korean Society for Computer Game)
초록

With the recent advent of Metaverse, the character industry that reflects the characteristics of users' faces is drawing attention. there is a hassle that users have to select face components such as eyes, nose, and mouth one by one. In this paper, we propose a diffusion-based model that automatically generates characters from content human photographs. Our model generates user artistic characters by reflecting content information such as face angle, direction, and shape of a content human photo. In particular, our model automatically analyzes detailed information such as glasses and whiskers from content photo images and reflects them in artistic characters generated. Our network generates the final character through a three-step: diffusion process, UNet, and denoising processes. We use image encoders and CLIP encoders for the connection between style and input data. In the diffusion process, a collection of noise vectors is gradually added to a style vector to enable lossless learning of the detailed styles. All input values except for the style images are vectorized with CLIP encoders and then learned with noise style vectors in the UNet. Subsequently, noise is removed from the vectors through the UNet to obtain the artistic character image. We demonstrate our performance by comparing the results of other models with our results. Our method reflects content information without loss and generates natural high-definition characters.

목차
1. Introduction
    1.2 Related Resarch
2. Body
    2.1 OverView
    2.2 Methode
3. Conclusion
    3.1 Conclusion
    3.1 compare
저자
  • Hyemin YANG(Dept. of Computer Science, Sangmyung Univ., 20, Hongjimun 2-gil, Jongno-gu, Seoul, Republic of Korea)
  • Heekyung YANG(Dept. of Software, Sangmyung Univ., 31, Sangmyeongdae-gil, Dongnam-gu, Cheonan 31066, Republic of Korea) Corresponding author
  • Kyungha MIN(Dept. of Computer Science, Sangmyung Univ., 20, Hongjimun 2-gil, Jongno-gu, Seoul, Republic of Korea) Corresponding author