AFFECTIVE COMPUTING IN MODERN ART EDUCATION

Authors

  • Manikandan Jagarajan Assistant Professor, Department of Computer Science and Engineering, Aarupadai Veedu Institute of Technology, Vinayaka Mission’s Research Foundation (DU), Tamil Nadu, India
  • Ananta Narayana Assistant Professor, School of Business Management, Noida International University, Greater Noida, Uttar Pradesh, India
  • Saksham Sood Centre of Research Impact and Outcome, Chitkara University, Rajpura- 140417, Punjab, India
  • Dr. Abhishek Upadhyay Assistant Professor, Department of Management, ARKA JAIN University Jamshedpur, Jharkhand, India
  • Dinesh Shravan Datar Department of Artificial Intelligence and Data Science, Vishwakarma Institute of Technology, Pune, Maharashtra, 411037 India
  • Kalpana Munjal Associate Professor, Department of Design, Vivekananda Global University, Jaipur, India

DOI:

https://doi.org/10.29121/shodhkosh.v6.i4s.2025.6850

Keywords:

Affective Computing, Emotion Recognition, Art Education, Creative Learning, Emotional Intelligence, Adaptive Pedagogy

Abstract [English]

Affective computing Affective computing, or more strictly speaking, its recognition and interpretation of human emotions, has emerged as an influential model in the redesigning of art education. This paper will examine how emotion-sensitive technologies can be applied to contemporary art schools to improve creativity, participation, and self-directed learning. Conventional pedagogy of art tend to focus on technical achievement and abstract exploration but does not acknowledge the emotional relations that is the foundation of art. With the introduction of affective computing (recognition of emotions based on facial expressions, tonal, and physiological elements), the educator will be able to more effectively understand the emotional state of the learners and incorporate this knowledge in the instructional approach. The study applies multimodal data gathering in academic arts, including emotion, performance and interaction records to come up with affect-adaptive systems. Emotion classification methods based on machine learning, including CNNs, LSTMs, and multimodal fusion networks, are applied, whereas the performance of the systems is measured by such metrics as accuracy, F1-score, and emotional congruence index. The suggested affect-aware art education platform gives real time emotional feedback, adaptive learning and the use of digital art tools in order to facilitate expressive development. The experimental applications show that emotion-augmented assignments and feedback loops can enhance motivation and quality of the work done by learners.

References

Agung, E. S., Rifai, A. P., and Wijayanto, T. (2024). Image-Based Facial Emotion Recognition Using Convolutional Neural Network on Emognition Dataset. Scientific Reports, 14, Article 14429. https://doi.org/10.1038/s41598-024-65276-x

Canal, F. Z., Müller, T. R., Matias, J. C., Scotton, G. G., de Sá Junior, A. R., Pozzebon, E., and Sobieranski, A. C. (2022). A Survey on Facial Emotion Recognition Techniques: A State-Of-The-Art Literature Review. Information Sciences, 582, 593–617. https://doi.org/10.1016/j.ins.2021.10.005

El Bahri, N., Itahriouan, Z., Abtoy, A., and Belhaouari, S. B. (2023). Using Convolutional Neural Networks to Detect Learner’s Personality Based on the Five Factor Model. Computers and Education: Artificial Intelligence, 5, Article 100163. https://doi.org/10.1016/j.caeai.2023.100163

Elansary, L., Taha, Z., and Gad, W. (2024). Survey on Emotion Recognition Through Posture Detection and the Possibility of its Application in Virtual Reality (arXiv preprint arXiv:2408.01728).

Elsheikh, R. A., Mohamed, M. A., Abou-Taleb, A. M., and Ata, M. M. (2024). Improved Facial Emotion Recognition Model Based on a Novel Deep Convolutional Structure. Scientific Reports, 14, Article 29050. https://doi.org/10.1038/s41598-024-79167-8

Khare, S. K., Blanes-Vidal, V., Nadimi, E. S., and Acharya, U. R. (2024). Emotion Recognition and Artificial Intelligence: A Systematic Review (2014–2023) and Research Recommendations. Information Fusion, 102, Article 102019. https://doi.org/10.1016/j.inffus.2023.102019

Lin, W., and Li, C. (2023). Review of Studies on Emotion Recognition and Judgment Based on Physiological Signals. Applied Sciences, 13(4), Article 2573. https://doi.org/10.3390/app13042573

Mohana, M., and Subashini, P. (2024). Facial Expression Recognition Using Machine Learning and Deep Learning Techniques: A Systematic Review. SN Computer Science, 5, Article 432. https://doi.org/10.1007/s42979-024-02792-7

Saganowski, S., Perz, B., Polak, A. G., and Kazienko, P. (2022). Emotion Recognition for Everyday Life Using Physiological Signals from Wearables: A Systematic Literature Review. IEEE Transactions on Affective Computing, 14, 1876–1897. https://doi.org/10.1109/TAFFC.2022.3176135

Schuller, D. M., and Schuller, B. W. (2021). A Review on Five Recent and Near-Future Developments in Computational Processing of Emotion in the Human Voice. Emotion Review, 13(1), 44–50. https://doi.org/10.1177/1754073919898526

Singh, Y. B., and Goel, S. (2022). A Systematic Literature Review of Speech Emotion Recognition Approaches. Neurocomputing, 492, 245–263. https://doi.org/10.1016/j.neucom.2022.04.028

Trinh Van, L., Dao Thi Le, T., Le Xuan, T., and Castelli, E. (2022). Emotional Speech Recognition Using Deep Neural Networks. Sensors, 22(4), Article 1414. https://doi.org/10.3390/s22041414

Wu, J., Zhang, Y., Sun, S., Li, Q., and Zhao, X. (2022). Generalized Zero-Shot Emotion Recognition from Body Gestures. Applied Intelligence, 52, 8616–8634. https://doi.org/10.1007/s10489-021-02927-w

Zhao, Y., and Shu, X. (2023). Speech Emotion Analysis Using Convolutional Neural Network (CNN) and Gamma Classifier-Based Error Correcting Output codes (ECOC). Scientific Reports, 13, Article 20398. https://doi.org/10.1038/s41598-023-47118-4

Downloads

Published

2025-12-25

How to Cite

Jagarajan, M., Narayana, A., Sood, S., Upadhyay, A., Datar, D. S., & Munjal, K. (2025). AFFECTIVE COMPUTING IN MODERN ART EDUCATION. ShodhKosh: Journal of Visual and Performing Arts, 6(4s), 171–180. https://doi.org/10.29121/shodhkosh.v6.i4s.2025.6850