Neural Puppet: Generative Layered Cartoon Characters

Research output: Contribution to journalConference articleResearchpeer-review

Standard

Neural Puppet : Generative Layered Cartoon Characters. / Poursaeed, Omid; Kim, Vladimir G.; Shechtman, Eli; Saito, Jun; Belongie, Serge.

In: Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020, 03.2020, p. 3335-3345.

Research output: Contribution to journalConference articleResearchpeer-review

Harvard

Poursaeed, O, Kim, VG, Shechtman, E, Saito, J & Belongie, S 2020, 'Neural Puppet: Generative Layered Cartoon Characters', Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020, pp. 3335-3345. https://doi.org/10.1109/WACV45572.2020.9093346

APA

Poursaeed, O., Kim, V. G., Shechtman, E., Saito, J., & Belongie, S. (2020). Neural Puppet: Generative Layered Cartoon Characters. Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020, 3335-3345. https://doi.org/10.1109/WACV45572.2020.9093346

Vancouver

Poursaeed O, Kim VG, Shechtman E, Saito J, Belongie S. Neural Puppet: Generative Layered Cartoon Characters. Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020. 2020 Mar;3335-3345. https://doi.org/10.1109/WACV45572.2020.9093346

Author

Poursaeed, Omid ; Kim, Vladimir G. ; Shechtman, Eli ; Saito, Jun ; Belongie, Serge. / Neural Puppet : Generative Layered Cartoon Characters. In: Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020. 2020 ; pp. 3335-3345.

Bibtex

@inproceedings{73979ecfd8ad42e1a23c8c150d6d8f21,
title = "Neural Puppet: Generative Layered Cartoon Characters",
abstract = "We propose a learning based method for generating new animations of a cartoon character given a few example images. Our method is designed to learn from a traditionally animated sequence, where each frame is drawn by an artist, and thus the input images lack any common structure, correspondences, or labels. We express pose changes as a deformation of a layered 2.5D template mesh, and devise a novel architecture that learns to predict mesh deformations matching the template to a target image. This enables us to extract a common low-dimensional structure from a diverse set of character poses. We combine recent advances in differentiable rendering as well as mesh-aware models to successfully align common template even if only a few character images are available during training. In addition to coarse poses, character appearance also varies due to shading, out-of-plane motions, and artistic effects. We capture these subtle changes by applying an image translation network to refine the mesh rendering, providing an end-to-end model to generate new animations of a character with high visual quality. We demonstrate that our generative model can be used to synthesize in-between frames and to create data-driven deformation. Our template fitting procedure outperforms state-of-the-art generic techniques for detecting image correspondences.",
author = "Omid Poursaeed and Kim, {Vladimir G.} and Eli Shechtman and Jun Saito and Serge Belongie",
note = "Publisher Copyright: {\textcopyright} 2020 IEEE.; 2020 IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2020 ; Conference date: 01-03-2020 Through 05-03-2020",
year = "2020",
month = mar,
doi = "10.1109/WACV45572.2020.9093346",
language = "English",
pages = "3335--3345",
journal = "Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020",

}

RIS

TY - GEN

T1 - Neural Puppet

T2 - 2020 IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2020

AU - Poursaeed, Omid

AU - Kim, Vladimir G.

AU - Shechtman, Eli

AU - Saito, Jun

AU - Belongie, Serge

N1 - Publisher Copyright: © 2020 IEEE.

PY - 2020/3

Y1 - 2020/3

N2 - We propose a learning based method for generating new animations of a cartoon character given a few example images. Our method is designed to learn from a traditionally animated sequence, where each frame is drawn by an artist, and thus the input images lack any common structure, correspondences, or labels. We express pose changes as a deformation of a layered 2.5D template mesh, and devise a novel architecture that learns to predict mesh deformations matching the template to a target image. This enables us to extract a common low-dimensional structure from a diverse set of character poses. We combine recent advances in differentiable rendering as well as mesh-aware models to successfully align common template even if only a few character images are available during training. In addition to coarse poses, character appearance also varies due to shading, out-of-plane motions, and artistic effects. We capture these subtle changes by applying an image translation network to refine the mesh rendering, providing an end-to-end model to generate new animations of a character with high visual quality. We demonstrate that our generative model can be used to synthesize in-between frames and to create data-driven deformation. Our template fitting procedure outperforms state-of-the-art generic techniques for detecting image correspondences.

AB - We propose a learning based method for generating new animations of a cartoon character given a few example images. Our method is designed to learn from a traditionally animated sequence, where each frame is drawn by an artist, and thus the input images lack any common structure, correspondences, or labels. We express pose changes as a deformation of a layered 2.5D template mesh, and devise a novel architecture that learns to predict mesh deformations matching the template to a target image. This enables us to extract a common low-dimensional structure from a diverse set of character poses. We combine recent advances in differentiable rendering as well as mesh-aware models to successfully align common template even if only a few character images are available during training. In addition to coarse poses, character appearance also varies due to shading, out-of-plane motions, and artistic effects. We capture these subtle changes by applying an image translation network to refine the mesh rendering, providing an end-to-end model to generate new animations of a character with high visual quality. We demonstrate that our generative model can be used to synthesize in-between frames and to create data-driven deformation. Our template fitting procedure outperforms state-of-the-art generic techniques for detecting image correspondences.

UR - http://www.scopus.com/inward/record.url?scp=85085497976&partnerID=8YFLogxK

U2 - 10.1109/WACV45572.2020.9093346

DO - 10.1109/WACV45572.2020.9093346

M3 - Conference article

AN - SCOPUS:85085497976

SP - 3335

EP - 3345

JO - Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020

JF - Proceedings - 2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020

Y2 - 1 March 2020 through 5 March 2020

ER -

ID: 301823072