AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (2.1 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Research Article | Open Access

Dance to the beat: Synchronizing motion to audio

Tel Aviv University, Tel Aviv 6997801, Israel.
Show Author Information

Abstract

In this paper we introduce a video post-processing method that enhances the rhythm of a dancing performance, in the sense that the dancing movements are more in time to the beat of the music. The dancing performance as observed in a video is analyzed and segmented into motion intervals delimited by motion beats. We present an image-space method to extract the motion beats of a video by detecting frames at which there is a significant change in direction or motion stops. The motion beats are then synchronized with the music beats such that as many beats as possible are matched with as little as possible time-warping distortion to the video. We show two applications for this cross-media synchronization: one where a given dance performance is enhanced to be better synchronized with its original music, and one where a given dance video is automatically adapted to be synchronized with different music.

Electronic Supplementary Material

Video
41095_2018_115_MOESM1_ESM.mp4

References

[1]
Repp, B. H. Musical synchronization. In: Music, Motor Control and the Brain. Altenmuller, E.; Wiesendanger, M.; Keselring, J. Eds. Oxford University Press, 55-76, 2006.
[2]
Kim, T.-h.; Park, S. I.; Shin, S. Y. Rhythmic-motion synthesis based on motion-beat analysis. ACM Transactions on Graphics Vol. 22, No. 3, 392-401, 2003.
[3]
Shiratori, T.; Nakazawa, A.; Ikeuchi, K. Dancing-to-music character animation. Computer Graphics Forum Vol. 25, No. 3, 449-458, 2006.
[4]
Chu, W.-T.; Tsai, S.-Y. Rhythm of motion extraction and rhythm-based cross-media alignment for dance videos. IEEE Transactions on Multimedia Vol. 14, No. 1, 129-141, 2012.
[5]
Flash, T.; Hogan, N. The coordination of arm movements: An experimentally confirmed mathematical model. Journal of Neuroscience Vol. 5, No. 7, 1688-1703, 1985.
[6]
Jones, M. R.; Boltz, M. Dynamic attending and responses to time. Psychological Review Vol. 96, No. 3, 459-491, 1989.
[7]
Leyvand, T.; Cohen-Or, D.; Dror, G.; Lischinski, D. Digital face beautification. In: Proceedings of the ACM SIGGRAPH 2006 Sketches, Article No. 169, 2006.
[8]
Zhou, S.; Fu, H.; Liu, L.; Cohen-Or, D.; Han, X. Parametric reshaping of human bodies in images. ACM Transactions on Graphics Vol. 29, No. 4, Article No. 126, 2010.
[9]
Zhou, F.; Torre, F. Canonical time warping for alignment of human behavior. In: Proceedings of the Advances in Neural Information Processing Systems, 2286-2294, 2009.
[10]
Zhou, F.; De la Torre, F. Generalized time warping for multi-modal alignment of human motion. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1282-1289, 2012.
[11]
Yoon, J.-C.; Lee, I.-K.; Byun, S. Automated music video generation using multi-level feature-based segmentation. In: Handbook of Multimedia for Digital Entertainment and Arts. Furht, B. Ed. Springer, 385-401, 2009.
[12]
Yoon, J.-C.; Lee, I.-K.; Lee, H.-C. Feature-based synchronization of video and background music. In: Advances in Machine Vision, Image Processing, and Pattern Analysis. Lecture Notes in Computer Science, Vol. 4153. Zheng, N.; Jiang, X.; Lan, X. Eds. Springer Berlin Heidelberg, 205-214, 2006.
[13]
Jehan, T.; Lew, M.; Vaucelle, C. Cati dance: Self-edited, self-synchronized music video. In: Proceedings of the ACM SIGGRAPH 2003 Sketches & Applications, 1-1, 2003.
[14]
Suwajanakorn, S.; Seitz, S. M.; Kemelmacher-Shlizerman, I. Synthesizing Obama: Learning lip sync from audio. ACM Transactions on Graphics Vol. 36, No. 4, Article No. 95, 2017.
[15]
Caspi, Y.; Irani, M. Spatio-temporal alignment of sequences. IEEE Transactions on Pattern Analysis and Machine Intelligence Vol. 24, No. 11, 1409-1424, 2002.
[16]
Slaney, M.; Covell, M. FaceSync: A linear operator for measuring synchronization of video facial images and audio tracks. In: Proceedings of the 13th International Conference on Neural Information Processing Systems, 784-790, 2000.
[17]
Wang, O.; Schroers, C.; Zimmer, H.; Gross, M.; Sorkine-Hornung, A. VideoSnapping: Interactive synchronization of multiple videos. ACM Transactions on Graphics Vol. 33, No. 4, Article No. 77, 2014.
[18]
Lu, S.-P.; Zhang, S.-H.; Wei, J.; Hu, S.-M.; Martin, R. R. Timeline editing of objects in video. IEEE Transactions on Visualization and Computer Graphics Vol. 19, No. 7, 1218-1227, 2013.
[19]
Shiratori, T.; Nakazawa, A.; Ikeuchi, K. Detecting dance motion structure through music analysis. In: Proceedings of the 6th IEEE International Conference on Automatic Face and Gesture Recognition, 857-862, 2004.
[20]
Denman, H.; Doyle, E.; Kokaram, A.; Lennon, D.; Dahyot, R.; Fuller, R. Exploiting temporal discontinuities for event detection and manipulation in video streams. In: Proceedings of the 7th ACM SIGMM International Workshop on Multimedia Information Retrieval, 183-192, 2005.
[21]
McKinney, M. F.; Moelants, D.; Davies, M. E. P.; Klapuri, A. Evaluation of audio beat tracking and music tempo extraction algorithms. Journal of New Music Research Vol. 36, No. 1, 1-16, 2007.
[22]
Ellis, D. P. W. Beat tracking by dynamic programming. Journal of New Music Research Vol. 36, No. 1, 51-60, 2007.
[23]
Patel, A. D.; Iversen, J. R.; Bregman, M. R.; Schulz, I. Experimental evidence for synchronization to a musical beat in a nonhuman animal. Current Biology Vol. 19, No. 10, 827-830, 2009.
[24]
Patel, A. D.; Iversen, J. R.; Bregman, M. R.; Schulz, I. Studying synchronization to a musical beat in nonhuman animals. Annals of the New York Academy of Sciences Vol. 1169, No. 1, 459-469, 2009.
[25]
Patel, A. D.; Iversen, J. R.; Bregman, M. R.; Schulz, I.; Schulz, C. Investigating the human-specificity of synchronization to music. In: Proceedings of the 10th International Conference on Music Perception and Cognition, 100-104, 2008.
Computational Visual Media
Pages 197-208
Cite this article:
Bellini R, Kleiman Y, Cohen-Or D. Dance to the beat: Synchronizing motion to audio. Computational Visual Media, 2018, 4(3): 197-208. https://doi.org/10.1007/s41095-018-0115-y

793

Views

17

Downloads

9

Crossref

N/A

Web of Science

10

Scopus

0

CSCD

Altmetrics

Revised: 03 February 2018
Accepted: 27 March 2018
Published: 06 August 2018
© The Author(s) 2018

This article is published with open access at Springerlink.com

The articles published in this journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http:// creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Other papers from this open access journal are available free of charge from http://www.springer.com/journal/41095. To submit a manuscript, please go to https://www. editorialmanager.com/cvmj.

Return