|
ShodhKosh: Journal of Visual and Performing ArtsISSN (Online): 2582-7472
Automated Editing Tools for Media Students: A Comparative Study Dr. Ragini Kunal Jadhav 1 1 Assistant
Professor, Prin. L. N. Welingkar Institute of Management Development and
Research, (PGDM), India 2 Assistant
Professor, Meenakshi College of Arts and Science, Meenakshi Academy of Higher
Education and Research, Chennai, Tamil Nadu, India 3 Assistant Professor, Department of Civil Engineering Vishwakarma
Institute of Technology, Pune, Maharashtra, India 4 Associate Professor, School of Business Management, Noida
International University, Greater Noida, India 5 Department of E and TC Engineering, Bharati Vidyapeeth's College of
Engineering, Lavale, Pune, Maharashtra, India 6 Assistant Professor, Department of Journalism and Mass Communication,
Mangalayatan University, Aligarh, India
1. INTRODUCTION The speed of media production going digital has radically changed the process of producing and editing visual and audio material and distributing it. Conventionally, media editing was time-consuming and involved a lot of manual addition, technical knowledge, and long-term learning on professional software settings. Nevertheless, with the recent incorporation of artificial intelligence (AI) and machine learning (ML) into creative software, automated editing tools now exist that considerably lower the technical barriers, as well as the speed of production processes Zhai et al. (2024). Among media students (who must often juggle between conceptual creativity and the acquisition of technical expertise) these tools constitute both the opportunity and the pedagogical challenge. Computer vision, natural language processing, audio signal processing, and deep learning are some of the AI-based methods used to solve tasks that were previously done by humans through automated editing tools Licht et al. (2024). These operations are scene detection, color grading, audio enhancement, subtitle generation, motion tracking and content summarization. Such pop culture tools as Adobe premier Pro (with its Sensei AI), DaVinci Resolve, Runway ML, Descript, and Capcut have introduced automation into regular editing programs, meaning advanced post production is now available to users with minimal expertise Alier et al. (2024). Consequently, media education is undergoing a transformation in terms of tool-based learning towards concept-based learning and workflow-based learning Almassaad et al. (2024). Figure 1 |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
Table 1 Evolution of Media Editing Tools and Automation Capabilities |
||||
|
Era |
Representative Tools |
Dominant Technologies |
Level of Automation |
Educational
Implications |
|
Early Digital Editing
(1990s–2000s) |
Adobe Premiere Pro
(early versions), Final Cut Pro Almulla et al. (2023) |
Manual non-linear
editing, timeline-based interfaces |
Very Low |
Strong technical skill
development; steep learning curve for students |
|
Assisted Digital
Editing (2005–2015) |
Adobe Premiere
(presets), Avid Media Composer Tiernan et al. (2023) |
Rule-based automation,
templates, presets |
Low–Moderate |
Improved efficiency;
still tool-centric pedagogy |
|
Intelligent Editing
Systems (2016–2020) |
DaVinci Resolve Munro et al.
(2023) |
Computer vision,
ML-based color grading, facial recognition |
Moderate |
Shift toward creative
decision-making and visual aesthetics |
|
AI-Native Editing
Platforms (2020–Present) |
Runway ML, Descript Brako et al. (2023) |
Deep learning, NLP,
multimodal AI, generative models |
High |
Rapid prototyping,
accessibility for beginners, focus on storytelling |
|
Mobile & Cloud
Automation Era (2022–Present) |
CapCut, Canva Susanti et al. (2023) |
Cloud AI, template
intelligence, real-time automation |
Very High |
Democratized editing;
raises concerns on skill depth and originality |
In more recent times, mobile, and cloud-based editing system brought automation to a larger group of learners. CapCut and Canva are programs designed to provide almost instant editing output through the use of AI real-time assistance, templates, and cloud computation Matsiola et al. (2022). Although such platforms make the media creation more democratic, and facilitate access to it, they also create pedagogical issues of originality, the depth of skills, and excessive dependence on ready-made templates.
3. Categories of Automated Editing Tools
The modern media education automated editing tools can be categorized based on the extent to which they are automated, the extent to which they require human input and the level of pedagogical interest. In contrast to conventional editing software in which the main role falls on the manual control, the automated software redistributes the creative and technical tasks to users and to intelligent systems O'Connor et al. (2022). This kind of categorization is necessary to the comparative analysis because it helps differentiate between the tools that only facilitate editing and those that, on the contrary, affect the creative results. Thus, it is possible to classify automated editing tools into four categories: assistive automation tools, intelligent enhancement tools, AI-native automation platforms, and template-driven cloud-based tools Desai and Kulkarni (2022). Assistive automation systems integrate automation into more traditional editing processes in a way that does not change creative control. Auto-align, stabilization, and preset corrections are some of the features that facilitate manual editing activities, such as in Adobe Premiere Pro and Avid Media Composer. They are specifically appropriate in the basic media education, as they expose the students to industry-standard settings and minimize the mechanical load.
Table 2
|
Table 2 Classification of Automated Editing Tools in Media Education |
||||
|
Category |
Representative Tools |
Core Technologies |
Automation Role |
Educational
Suitability |
|
Assistive Automation
Tools Chaudhry et al. (2022) |
Adobe Premiere Pro,
Avid Media Composer |
Rule-based logic,
basic ML |
Supports manual
workflows |
Strong for
foundational and industry training |
|
Intelligent
Enhancement Tools Williamson et al. (2020) |
DaVinci Resolve |
Computer vision, ML |
Enhances aesthetics
selectively |
Supports visual
literacy and quality benchmarking |
|
AI-Native Automation
Platforms Luan et al. (2020) |
Runway ML, Descript |
Deep learning, NLP,
multimodal AI |
Drives editing
decisions |
Enables rapid
prototyping and conceptual learning |
|
Template-Driven Cloud
Tools |
CapCut, Canva |
Cloud AI, template
intelligence |
Automates structure
and style |
Useful for beginners;
limited creative depth |
Semantic, text-based and generative editing Runway ML and Descript tools allow semantic, text-based, and generative editing workflows with the help of deep learning and natural language processing. Cloud based tools based on templates put more emphasis on accessibility and speed by use of predefined layouts and automated processing. Content creation applications like CapCut and Canva are easy to use and may reduce the originality because of the dependency on templates. Their learning value is mainly in introductory learning scenarios and should be well framed to guarantee significant creative interaction.
4. Comparative Evaluation Framework
A framework of systematic and open comparison of automated editing tools is necessary in the context of media education, since they are not only technologically advanced but also in pedagogical terms. Compared to comparisons that are purely industry-oriented, where the central focus is considered to be efficiency or output quality, an educational evaluation framework should consider learning outcomes, creative agency, usability, and skill development. This paper also assumes a multi-dimensional comparative paradigm that is based on a combination of technical performance indicators with educational appropriateness indicators, which allow the balanced evaluation of media student automated editing tools.
E=i=1∑5wi×Si
The suggested framework analyzes tools on five major dimensions including level of automation, usability and learning curve, creative control and transparency, pedagogical alignment, and workflow efficiency. These dimensions have been chosen with references to the curricula of the media education, in which the students are supposed to achieve both conceptual and practical competence. All the dimensions are conceptualized using quantifiable standards so that the qualitative observations can be turned into quantifiable comparative ratings.
i=1∑5wi=1
The level of automation dimension evaluates how much a tool is involved in the editing process like cutting, enhancement, synchronization and reorganization of contents that it does by itself without human involvement. The scores are high in automation, meaning more decision-making is there to rely on AI, and low scores indicate the presence of manual or supportive workflows. Even though the high level of automation might increase the speed of production and lessen the cognitive load, it can also hide the underlying criteria of the editing. Thus, automation is not considered an absolute benefit but as a situational characteristic whereby the educational worth will be determined by the design of instructions.
Figure 2

Figure 2 Multi-Dimensional Comparative Evaluation Framework for Automated Editing Tools in Media Education, Integrating Technical, Creative, And Pedagogical Criteria
The dimension of usability and learning curve will include the simplicity with which learners will be able to adapt and use an instrument. They are interface intuitiveness, onboarding, tutoring accessibility, and technical pre-knowledge level. Simplified interfaces and guided workflows are rated higher on usability especially on beginner-level media course as shown in Figure 2 But over simplification can act as a constraint to exposure to professional editing paradigms, which is taken into account when analysing results. The creative control and transparency dimension measures how students are able to comprehend, customize and surpass automated decisions. A creative control is a focal idea in pedagogical issue because media education lays stress on authorship, intentionality, as well as aesthetic rationale. The higher the score in this dimension, the more explainable AI recommendations, adjustable settings, and the ability to refine manually the tools are compared to the systems that operate like black boxes. Pedagogical alignment dimension is the dimension that analyzes the alignment of a tool with learning goals like storytelling, critical reflection, iterative experimentation, and transferable of skills. This involves support of the assessment practices, feedback and revision support and curriculum outcomes. Those tools that have been developed with education-friendly workflows (transcript based editing or high speed prototyping) are more pedagogically aligned than those that are just developed with speed or social media output in mind. Lastly, the workflow efficiency will determine how automation saves time and complexity of operation without diminishing the quality of creativity. This can be applied especially in the academic environment where students operate within time limits. Measures comprise time taken to complete a task, a decrease in manual operations, and simplicity of iteration. Nevertheless, the efficiency scores are understood together with creative control in order not to overestimate speed at the cost of the depth of learning. In order to operationalise this framework, the scores of each dimension are assigned on a five-point Likert scale, with 1 referring to little support and 5 referring to excellent support. The total score of every tool is calculated as a weighted aggregate of single dimension scores. In this paper, all dimensions are given equal weights so that no one can lean towards completely technical or completely pedagogical standards. This scoring model offers a clear point of comparison, and can be modified in future researches, by varying the weights based on course objectives or learner profile. Pawlaszczyk et al. (2025)
5. Case Study: Student Task-Based Evaluation and Observations
A small-scale case study was carried out to augment the quantitative comparison that was carried out in Section 5.2 by investigating the interaction between media students and automated editing tools in authentic learning activities. The aim of the case study was to monitor the behavior at the workflow, mode of decision making, and learning outcomes in different types of automated editing tools under the same task conditions in the presence of the students. This qualitative layer becomes a more potent part of the comparative framework because it can show pedagogical implications that cannot be well measured alone with the help of numerical scores.
5.1. Case Study Design
The case study entailed 24 media undergraduate students who took up a course in digital media production. The participants were split into five groups with each group having a single automated editing tool: Adobe Premiere Pro, DaVinci Resolve, Runway ML, Descript, and CapCut. Each of the groups was presented with the identical raw material in the form of a short documentary video, an interview audio tape and a written script outline. The students were asked to do three tasks under a time constraints:
1) Assemble a coherent 2–3-minute edited video,
2) Enhance visual and audio quality,
3) Prepare one revised version based on self-reflection.
The evaluation dimensions outlined in Section 4 corresponded to the screen recordings, instructor rubrics, student reflection notes, and post-task questionnaires that were used to gather the observations. Certain trends were observed in types of tools. The time spent by students on timeline organization, manual trimming and parameter tuning was higher when using Adobe Premiere Pro and DaVinci Resolve. Though because they are slower in first steps, these students showed more interest in the editing principles, including the shot continuity, color balance, and audio layering. Based on reflection notes, there was better conceptual grasp, although at times it felt like getting frustrated by the complexity of the interface.
Figure 3

Figure 3
Radar Plot Comparing
Tool Performance Across Evaluation Dimensions (Mean Likert Scores, N=24)
Students who used Runway ML and Descript, in contrast, wrote first drafts quickly. The text-based editing in Descript allowed students to work on narrative structure instead of on timeline editing, whereas Runway ML allowed experimenting creatively by removing background with automatic background removal and visual effects. Nonetheless, there were transparency limitations as some students indicated that they could not understand the reason why some of the automated edits were made. Those students who used CapCut took the shortest time, and were very satisfied at the initial stages. Nonetheless, assessment by instructors revealed that there was a high visual consistency of the outputs because of the use of templates, and there was a minimal difference in styles.
Table 3
|
Table 3 Summary of Student Task Observations Across Tools |
|||
|
Tool |
Observed Strengths |
Observed Limitations |
Learning Impact |
|
Adobe Premiere Pro |
High creative control,
deep skill engagement |
Steep learning curve,
slower completion |
Strong conceptual
understanding |
|
DaVinci Resolve |
Professional visual
quality, color literacy |
Interface complexity |
Advanced aesthetic
awareness |
|
Runway ML |
Rapid prototyping,
creative exploration |
Opaque automation
decisions |
High ideation,
moderate skill clarity |
|
Descript |
Narrative-focused
editing, usability |
Limited fine-grain
visual control |
Strong storytelling
skills |
|
CapCut |
Speed, accessibility,
confidence boost |
Template dependence,
shallow reasoning |
High engagement, lower
skill depth |
As exhibited in the case study, the level of automation has a greater effect on learning behavior, and not efficiency of task. Notably, reflections of students showed that creative understanding was enhanced where automation facilitated and not substituted in making decisions.
Figure 4

Figure 4 Distribution of Completion Time Across Tools (Minutes), Showing Variability Among Student Groups
These results supplement the quantitative results of Section 5.2 and affirm the thesis that choice of tools in teaching media must be pedagogically sequenced. First courses can be supported by tools based on AI-natives or templates to develop confidence, but intermediate and advanced courses can incorporate assistive and intelligent enhancement tools to develop critical editing skills. Even in situations where the efficiency of workflow was high, the creative control and pedagogical alignment came out as determinant factors that contributed to the quality of the learning process. This proves the inability to predict educational appropriateness based on the performance of automation solely, hence the necessity of a multi-dimensional and integrated evaluation method.
6. Comparative Analysis and Results
To complete the research, the tools representing each of the categories were tested: Adobe Premiere Pro (assistive automation), DaVinci Resolve (intelligent enhancement), Runway ML and Descript (AI-native platforms), and CapCut (template-driven tools). The assessment of all tools was made through the same datasets, tasks and scoring rubrics to compare them with each other. AI-native and template-driven tools scored the most in automation and efficiency, and were particularly more successful in auto-cutting and removing the background, but tended to decrease the student involvement in creative decision-making. Conversely, more creative control and transparency was provided by assistive and intelligent enhancement tools that facilitated more insight into the principles of editing even at higher learning curves. Beginner-oriented tools like CapCut and Descript were more usable, whereas the professional tools were favored when the skill was developed over the long term.
Table 4
|
Table 4 Comparative Scores of Automated Editing Tools (Mean Values) |
||||||
|
Tool |
Automation |
Usability |
Creative Control |
Pedagogical Alignment |
Efficiency |
Overall Score |
|
Adobe Premiere Pro |
2.5 |
3 |
4.5 |
4 |
3 |
3.4 |
|
DaVinci Resolve |
3.5 |
2.8 |
4.3 |
3.8 |
3.2 |
3.5 |
|
Runway ML |
4.6 |
4.2 |
2.9 |
4.1 |
4.5 |
4.1 |
|
Descript |
4.2 |
4.5 |
3.2 |
4.4 |
4.3 |
4.1 |
|
Cap Cut |
4.8 |
4.7 |
2.4 |
3 |
4.8 |
4 |
The outcomes of the comparative analysis reveal that there is no tool that fits all teaching and technical requirements. In lieu, all groups of automated editing tools have different pedagogical applications. Professional assistive tools are highly creative controlled and transferable in skills and hence can be applied in advanced courses in media. AI-native platforms are more nuanced strikes between automation and creativity and promote exploration and concept-based learning. Tools based on templates are as accessible and efficient as possible, but need to be carefully framed in the instructional approach to prevent superficial learning.
Figure 5

Figure 5 Observation Rubric Scores Across Learning-Impact Constructs (mean, 1–5)
The results indicate that blended-tool approach might be the most promising in media-education whereby, students would be introduced to various editing paradigms at various levels of learning. When educators choose the tools according to the objectives of the course and the level of proficiency in students, automation can be used to promote creativity without undermining the base of skills.
7. Conclusion
This research paper reviewed automated editing tools to media students based on a comparative, education-based paradigm, which combines technical performance with the pedagogical relevance. The analysis developed three categories of tools, namely, assistive, intelligent enhancement, AI-native and template-driven, and proved that educational effectiveness is not solely based on automation. Although extremely automated tools made it more efficient and more accessible, they tended to decrease the transparency of creativity and ability to articulate skills. By contrast, professional assistive aid helped to learn underlying concepts and creative control, but at a greater cognitive and time cost. The observations made in case-studies ensured that learning outcomes are enhanced when automation aids student decision-making as opposed to substituting it. The results support the significance of the relevance of the tools chosen to the course goals and the learner level. On the whole, automated editing tools have a great potential to improve the process of media education when incorporated with a blended, stage-appropriate pedagogical approach that will balance the efficiency, creativity, and development of the skills.
8. Limitations and Future Work
Although this study has aided, it has some limitations. To begin with, the sample size used in the case study was relatively small and the study was done in one academic setting and this may not represent the generalizability of the results to various other institutions and learner profiles. Second, the assessment was based more on the performance of short-term tasks and observation rubrics; no assessment was done on long-term learning outcomes, skill retention, and creative transferability. Third, the research was done on a selected list of popular automated editing tools, and new platforms or proprietary AI systems were not covered. Also, the conceptual but not empirically quantified ethical aspects were algorithmic bias, attribution of authorship, and data privacy.
These limitations can be overcome in the future by carrying out longitudinal studies with many institutions, larger and more heterogeneous samples of students, and adding objective learning metrics like revision depth, skill transfer also creative originality. Additional work can also expand the system of assessment to encompass ethical disclosure, explainable AI markers, and evaluation fairness, enhancing the practices related to the responsible use of AI-aided editing tools in the field of media education.
CONFLICT OF INTERESTS
None.
ACKNOWLEDGMENTS
None.
REFERENCES
Alier, M., Camba, J., and García‑Peñalvo, F. (2024). Generative Artificial Intelligence in Education: From Deceptive to Disruptive. International Journal of Interactive Multimedia and Artificial Intelligence, 5(1), 5-14. https://doi.org/10.9781/ijimai.2024.02.011
Almassaad, A., Alajlan, H., and Alebaikan, R. (2024). Student Perceptions of Generative Artificial Intelligence: Investigating Utilization, Benefits, and Challenges in Higher Education. Systems, 12(10), 385. https://doi.org/10.3390/systems12100385
Almulla, M. A. (2023). Constructivism Learning Theory: A Paradigm for Students' Critical Thinking, Creativity, and Problem Solving to Affect Academic Performance in Higher Education. Cogent Education, 10(1), 2172929. https://doi.org/10.1080/2331186X.2023.2172929
Al‑Zahrani, A. M. (2024). Unveiling the Shadows: Beyond the Hype of AI in Education. Heliyon, 10, e30696. https://doi.org/10.1016/j.heliyon.2024.e30696
Brako, D. K., and Mensah, A. K. (2023). Robots Over Humans? The Place of Artificial Intelligence in the Pedagogy of Art Direction in Film Education. Emerging Technologies, 3(1), 51-59. https://doi.org/10.57040/jet.v3i2.484
Bulut, O., Cutumisu, M., and Yilmaz, M. (2024). Ethical Challenges in AI‑Based Educational Measurement (arXiv:2406.18900). arXiv. https://arxiv.org/abs/2406.18900
Chaudhry, A., Wolf, J., and Gervasio, M. (2022). A Transparency Index for AI in Education (arXiv:2206.03220). arXiv. https://arxiv.org/abs/2206.03220
Chinta, R., Qazi, M., and Raza, M. (2024). FairAIED: A Review of Fairness, Accountability, and Bias in Educational AI systems (arXiv:2407.18745). arXiv. https://arxiv.org/abs/2407.18745
Deng, R., Yang, Y., and Shen, S. (2024). Impact of Question Presence and Interactivity in Instructional Videos on Student Learning. Education and Information Technologies, 1-29.
Desai, T. S., and Kulkarni, D. C. (2022). Assessment of Interactive Video to Enhance Learning Experience: A Case Study. Journal of Engineering Education Transformations, 35(1), 74-80. https://doi.org/10.16920/jeet/2022/v35is1/22011
Gouseti, A., James, F., Fallin, L., and Burden, K. (2024). The Ethics of Using AI in K-12 Education: A Systematic Literature Review. Technology, Pedagogy and Education, 34(2), 161-182. https://doi.org/10.1080/1475939X.2024.2428601
Licht, F. (2024). Generative Artificial Intelligence in Higher Education: Why the "Banning Approach" to Student use is Sometimes Morally Justified. Philosophy & Technology, 37(1), 113. https://doi.org/10.1007/s13347-024-00799-9
Luan, H., et al. (2020). Challenges and Future Directions of Big Data and Artificial Intelligence in Education. Frontiers in Psychology, 11, 580820. https://doi.org/10.3389/fpsyg.2020.580820
Matsiola, M., Spiliopoulos, P., and Tsigilis, N. (2022). Digital Storytelling in Sports Narrations: Employing Audiovisual tools in Sport Journalism Higher Education Course. Education Sciences, 12(1), 51. https://doi.org/10.3390/educsci12010051
Munro, R. (2023). Co‑Creating Film Education: Moments of Divergence and Convergence on Queen Margaret University's Introduction to Film Education Course. Film Education Journal, 6(1), 32-46. https://doi.org/10.14324/FEJ.06.1.04
O'Connor. (2022). Constructivism, Curriculum and the Knowledge Question: Tensions and Challenges for Higher Education. Studies in Higher Education, 47(2), 412-422. https://doi.org/10.1080/03075079.2020.1750585
Pawlaszczyk, D., Engler, P., Bodach, R., Michel, M., and Zimmermann, R. (2025). Automated Data Population for Ios Devices with Autopodmobile. Journal of Digital Security and Forensics, 2(1), 58–66. https://doi.org/10.29121/digisecforensics.v2.i1.2025.46
Susanti, A., Suyatna, A., and Herlina, K. (2023). Development of H5P Moodle‑Based Interactive STEM‑loaded Videos to Grow Performance Skills as an Effort to Overcome Learning Loss in Electrical Measuring Materials. Jurnal Penelitian Pendidikan IPA, 9(4), 6974-6984. https://doi.org/10.29303/jppipa.v9i9.3546
Tiernan, P., Costello, E., Donlon, E., Parysz, M., and Scriney, M. (2023). Information and Media Literacy in the Age of AI: Options for the future. Education Sciences, 13(9), 906. https://doi.org/10.3390/educsci13090906
Williamson, B., Eynon, R., and Potter, J. (2020). Pandemic Politics, Pedagogies and Practices: Digital Technologies and Distance Education During the Coronavirus Emergency. Learning, Media and Technology, 45(2), 107-114. https://doi.org/10.1080/17439884.2020.1761641
Zhai. (2024). AI‑Generated Content and its Impact on Student Learning and Critical Thinking: A Pedagogical Reflection. Smart Learning Environments, 11(1), 12.
|
|
This work is licensed under a: Creative Commons Attribution 4.0 International License
© ShodhKosh 2026. All Rights Reserved.