Multimodal dance style transfer

Research output: Contribution to journalJournal articleResearchpeer-review

Standard

Multimodal dance style transfer. / Yin, Wenjie; Yin, Hang; Baraka, Kim; Kragic, Danica; Björkman, Mårten.

In: Machine Vision and Applications, Vol. 34, No. 4, 48, 2023.

Research output: Contribution to journalJournal articleResearchpeer-review

Harvard

Yin, W, Yin, H, Baraka, K, Kragic, D & Björkman, M 2023, 'Multimodal dance style transfer', Machine Vision and Applications, vol. 34, no. 4, 48. https://doi.org/10.1007/s00138-023-01399-x

APA

Yin, W., Yin, H., Baraka, K., Kragic, D., & Björkman, M. (2023). Multimodal dance style transfer. Machine Vision and Applications, 34(4), [48]. https://doi.org/10.1007/s00138-023-01399-x

Vancouver

Yin W, Yin H, Baraka K, Kragic D, Björkman M. Multimodal dance style transfer. Machine Vision and Applications. 2023;34(4). 48. https://doi.org/10.1007/s00138-023-01399-x

Author

Yin, Wenjie ; Yin, Hang ; Baraka, Kim ; Kragic, Danica ; Björkman, Mårten. / Multimodal dance style transfer. In: Machine Vision and Applications. 2023 ; Vol. 34, No. 4.

Bibtex

@article{0ae51959c0c64a03bab031dcc3c04671,
title = "Multimodal dance style transfer",
abstract = "This paper first presents CycleDance, a novel dance style transfer system that transforms an existing motion clip in one dance style into a motion clip in another dance style while attempting to preserve the motion context of the dance. CycleDance extends existing CycleGAN architectures with multimodal transformer encoders to account for the music context. We adopt a sequence length-based curriculum learning strategy to stabilize training. Our approach captures rich and long-term intra-relations between motion frames, which is a common challenge in motion transfer and synthesis work. Building upon CycleDance, we further propose StarDance, which enables many-to-many mappings across different styles using a single generator network. Additionally, we introduce new metrics for gauging transfer strength and content preservation in the context of dance movements. To evaluate the performance of our approach, we perform an extensive ablation study and a human study with 30 participants, each with 5 or more years of dance experience. Our experimental results show that our approach can generate realistic movements with the target style, outperforming the baseline CycleGAN and its variants on naturalness, transfer strength, and content preservation. Our proposed approach has potential applications in choreography, gaming, animation, and tool development for artistic and scientific innovations in the field of dance.",
keywords = "Dance motion, Generative models, Multimodal learning, Style transfer",
author = "Wenjie Yin and Hang Yin and Kim Baraka and Danica Kragic and M{\aa}rten Bj{\"o}rkman",
note = "Publisher Copyright: {\textcopyright} 2023, The Author(s).",
year = "2023",
doi = "10.1007/s00138-023-01399-x",
language = "English",
volume = "34",
journal = "Machine Vision and Applications",
issn = "0932-8092",
publisher = "Springer",
number = "4",

}

RIS

TY - JOUR

T1 - Multimodal dance style transfer

AU - Yin, Wenjie

AU - Yin, Hang

AU - Baraka, Kim

AU - Kragic, Danica

AU - Björkman, Mårten

N1 - Publisher Copyright: © 2023, The Author(s).

PY - 2023

Y1 - 2023

N2 - This paper first presents CycleDance, a novel dance style transfer system that transforms an existing motion clip in one dance style into a motion clip in another dance style while attempting to preserve the motion context of the dance. CycleDance extends existing CycleGAN architectures with multimodal transformer encoders to account for the music context. We adopt a sequence length-based curriculum learning strategy to stabilize training. Our approach captures rich and long-term intra-relations between motion frames, which is a common challenge in motion transfer and synthesis work. Building upon CycleDance, we further propose StarDance, which enables many-to-many mappings across different styles using a single generator network. Additionally, we introduce new metrics for gauging transfer strength and content preservation in the context of dance movements. To evaluate the performance of our approach, we perform an extensive ablation study and a human study with 30 participants, each with 5 or more years of dance experience. Our experimental results show that our approach can generate realistic movements with the target style, outperforming the baseline CycleGAN and its variants on naturalness, transfer strength, and content preservation. Our proposed approach has potential applications in choreography, gaming, animation, and tool development for artistic and scientific innovations in the field of dance.

AB - This paper first presents CycleDance, a novel dance style transfer system that transforms an existing motion clip in one dance style into a motion clip in another dance style while attempting to preserve the motion context of the dance. CycleDance extends existing CycleGAN architectures with multimodal transformer encoders to account for the music context. We adopt a sequence length-based curriculum learning strategy to stabilize training. Our approach captures rich and long-term intra-relations between motion frames, which is a common challenge in motion transfer and synthesis work. Building upon CycleDance, we further propose StarDance, which enables many-to-many mappings across different styles using a single generator network. Additionally, we introduce new metrics for gauging transfer strength and content preservation in the context of dance movements. To evaluate the performance of our approach, we perform an extensive ablation study and a human study with 30 participants, each with 5 or more years of dance experience. Our experimental results show that our approach can generate realistic movements with the target style, outperforming the baseline CycleGAN and its variants on naturalness, transfer strength, and content preservation. Our proposed approach has potential applications in choreography, gaming, animation, and tool development for artistic and scientific innovations in the field of dance.

KW - Dance motion

KW - Generative models

KW - Multimodal learning

KW - Style transfer

U2 - 10.1007/s00138-023-01399-x

DO - 10.1007/s00138-023-01399-x

M3 - Journal article

AN - SCOPUS:85158999932

VL - 34

JO - Machine Vision and Applications

JF - Machine Vision and Applications

SN - 0932-8092

IS - 4

M1 - 48

ER -

ID: 370580421