DEEP LEARNING MODELS FOR CHOREOGRAPHY GENERATION

Authors

  • Dr. Afroj Alam Assistant Professor, Department of Computer Science & Engineering, Presidency University, Bangalore, Karnataka, India
  • Sadhana Sargam Assistant Professor, School of Business Management, Noida International University, India
  • Dr. Jyoti Rani Assistant Professor, Department of Fashion Design, Parul Institute of Design, Parul University, Vadodara, Gujarat, India
  • Pavas Saini Centre of Research Impact and Outcome, Chitkara University, Rajpura 140417, Punjab, India
  • Amol Bhilare Department of Computer Engineering, Vishwakarma Institute of Technology, Pune 411037, Maharashtra, India
  • S. Balakrishnan Professor and Head, Department of Computer Science and Engineering, Aarupadai Veedu Institute of Technology, Vinayaka Mission’s Research Foundation (DU), Tamil Nadu, India

DOI:

https://doi.org/10.29121/shodhkosh.v6.i5s.2025.6908

Keywords:

Deep Learning, Choreography Generation, Motion Synthesis, LSTM and Transformers, Generative Adversarial Networks, Diffusion Models

Abstract [English]

Due to the rapid development of deep learning, the new opportunities of computational production of human movement have been opened, especially in the field of dance choreography. The paper discusses deep learning choreography generators that combine movement information, music framework, and time in order to create expressive and sensible sequences of dances. Conventional choreography models tend to have a handmade regulation or professionalized composition where flexibility and creative variety is restricted. Conversely, deep learning methods that are data driven can directly learn complex spatio-temporal patterns using big datasets of motion and video. The suggested framework uses pose representations, video frames and rhythmic information based on music to simulate the inherent relationship between motion and sound. Recurrent neural networks as LSTM and GRU models are used to learn long-lasting temporal dependencies in dance sequences whereas transformer-based models are used to improve global context awareness and sequence coherence. Also, generative adversarial networks, diffusion-based networks are explored to achieve motion synthesis which provides smooth transitions, stylistic variability and a sense of realistic continuity in movement. A modular system architecture is structured in such a way that it can allow multimodal inputs, convolutional feature extraction and temporal sequence generation. The evaluation of the experimental results is carried out on standard choreography and motion datasets and performance is measured by quantitative evaluation measures including mean absolute error, Fréchet Inception Distance, and a smoothness index specific to the movement evaluation.

References

Chen, K., Tan, Z., Lei, J., Zhang, S. H., Guo, Y. C., Zhang, W., and Hu, S. M. (2021). ChoreoMaster: Choreography-Oriented Music-Driven Dance Synthesis. ACM Transactions on Graphics, 40, 1–13. https://doi.org/10.1145/3450626.3459849 DOI: https://doi.org/10.1145/3450626.3459932

Chen, Y., Wang, H., Yu, K., and Zhou, R. (2024). Artificial Intelligence Methods in Natural Language Processing: A Comprehensive Review. Highlights in Science, Engineering and Technology, 85, 545–550. DOI: https://doi.org/10.54097/vfwgas09

Copet, J., Kreuk, F., Gat, I., Remez, T., Kant, D., Synnaeve, G., Adi, Y., and Défossez, A. (2023). Simple and Controllable Music Generation. arXiv.

Croitoru, F. A., Hondru, V., Ionescu, R. T., and Shah, M. (2023). Diffusion Models in Vision: A Survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45, 10850–10869. https://doi.org/10.1109/TPAMI.2023.3288679 DOI: https://doi.org/10.1109/TPAMI.2023.3261988

Hong, C. (2024). Application of Virtual Digital People in the Inheritance and Development of Intangible Cultural Heritage. People’s Forum, 6, 103–105.

Hou, C. (2024). Artificial Intelligence Technology Drives Intelligent Transformation of Music Education. Applied Mathematics and Nonlinear Sciences, 9, 21–23. https://doi.org/10.2478/amns-2024-0021 DOI: https://doi.org/10.2478/amns-2024-1947

Lauriola, I., Lavelli, A., and Aiolli, F. (2022). An Introduction to Deep Learning in Natural Language Processing: Models, Techniques, and Tools. Neurocomputing, 470, 443–456. https://doi.org/10.1016/j.neucom.2021.10.040 DOI: https://doi.org/10.1016/j.neucom.2021.05.103

Lund, B. D., Wang, T., Mannuru, N. R., Nie, B., Shimray, S., and Wang, Z. (2023). ChatGPT and a New Academic Reality: Artificial Intelligence-Written Research Papers and the Ethics of the Large Language Models in Scholarly Publishing. Journal of the Association for Information Science and Technology, 74, 570–581. https://doi.org/10.1002/asi.24750 DOI: https://doi.org/10.1002/asi.24750

Pleshakova, E., Osipov, A., Gataullin, S., Gataullin, T., and Vasilakos, A. (2024). Next Gen Cybersecurity Paradigm Towards Artificial General Intelligence: Russian Market Challenges and Future Global Technological Trends. Journal of Computer Virology and Hacking Techniques, 7, Article 23. DOI: https://doi.org/10.1007/s11416-024-00529-x

Sun, Y., Dong, L., Huang, S., Ma, S., Xia, Y., Xue, J., Wang, J., and Wei, F. (2023). Retentive Network: A Successor to Transformer for Large Language Models. arXiv.

Tay, Y., Dehghani, M., Bahri, D., and Metzler, D. (2023). Efficient Transformers: A Survey. ACM Computing Surveys, 55, 16–21. https://doi.org/10.1145/3530811 DOI: https://doi.org/10.1145/3530811

Wu, J., Gan, W., Chen, Z., Wan, S., and Lin, H. (2023). AI-Generated Content (AIGC): A Survey. arXiv.

Zeng, D. (2025). AI-Powered Choreography Using a Multilayer Perceptron Model for Music-Driven Dance Generation. Informatica, 49, 137–148. DOI: https://doi.org/10.31449/inf.v49i20.8103

Zhang, L., and Zhang, L. (2022). Artificial Intelligence for Remote Sensing Data Analysis: A Review of Challenges and Opportunities. IEEE Geoscience and Remote Sensing Magazine, 10, 270–294. https://doi.org/10.1109/MGRS.2021.3132935 DOI: https://doi.org/10.1109/MGRS.2022.3145854

Downloads

Published

2025-12-28

How to Cite

Alam, A., Sargam, S., Rani, J., Saini, P., Bhilare, A., & S. Balakrishnan. (2025). DEEP LEARNING MODELS FOR CHOREOGRAPHY GENERATION. ShodhKosh: Journal of Visual and Performing Arts, 6(5s), 525–535. https://doi.org/10.29121/shodhkosh.v6.i5s.2025.6908