“AniDepth : Anime In-between Diffusion using Depth-guided Warped Line-art” by Koga, Kubo, Shinagawa, Fujimura, Kitano, et al. … – ACM SIGGRAPH HISTORY ARCHIVES

“AniDepth : Anime In-between Diffusion using Depth-guided Warped Line-art” by Koga, Kubo, Shinagawa, Fujimura, Kitano, et al. …

  • 2025 Talks_Koga_AniDepth_Anime In-between Diffusion using Depth-guided Warped Line-art

Conference:


Type(s):


Title:

    AniDepth : Anime In-between Diffusion using Depth-guided Warped Line-art

Session/Category Title:

    ML in Production

Presenter(s)/Author(s):


Moderator(s):



Abstract:


    We propose AniDepth, a novel anime in-betweening method using a video diffusion model. Even if the model is fine-tuned on an anime dataset, it still suffers from the domain gap between anime and the natural image domain, due to the strong priors of the base model. Therefore, we convert anime illustration into depth map, which is a modality filling the gap between anime and realistic domains to fully exploit the model’s prior knowledge. In addition, by using line-arts as guide during the in-betweening, we enhance the fidelity of the generated line-arts details. Our approach first interpolates converted depth maps, warps line-arts based on the depth maps, then interpolates colored images with using the line-arts as the conditions, preserving complex line-art and flat coloring even under large movements. Experiments show that our method outperforms the existing diffusion method in quantitative evaluation, improves temporal smoothness, and reduces line-art distortion. Moreover, because it requires no extra training, it can be easily integrated into current production pipelines at low operational cost.

References:


    [1] G. L. Moing, J. Ponce, and C. Schmid. 2024. Dense Optical Tracking: Connecting the Dots. In 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE/CVF, 19187–19197.
    [2] Jinbo Xing, Hanyuan Liu, Menghan Xia, Yong Zhang, Xintao Wang, Ying Shan, and Tien-Tsin Wong. 2024. ToonCrafter: Generative Cartoon Interpolation. ACM Trans. Graph. 43, 6, Article 245 (2024), 11 pages.
    [3] Lihe Yang, Bingyi Kang, Zilong Huang, Zhen Zhao, Xiaogang Xu, Jiashi Feng, and Hengshuang Zhao. 2024. Depth Anything V2. 37 (2024), 21875–21911.


ACM Digital Library Publication:



Overview Page:



Submit a story:

If you would like to submit a story about this presentation, please contact us: historyarchives@siggraph.org