AI for Education: A Living Catalog of Emerging Practice
What is the Living Catalog?
The Living Catalog of Generative AI for Education is an evolving resource designed to support both researchers and educators in exploring how AI is being used in teaching and learning. Developed by the Center for Learning Sciences at EPFL, the catalog maps a range of applications of generative AI in education.
Our goal is to point users to relevant scientific literature and examples of real-world practice. By doing so, we aim to provide both a starting point and practical guidance for those seeking to implement generative AI in pedagogically sound ways.
The SAMR Framework
The SAMR framework (Puentedura, 2009; Belkina et al., 2025) provides a useful lens for interpreting the spectrum of AI applications in education.
How the Catalog is Built
Critical Use Encouraged – We do not verify the quality of the referenced studies. Users are encouraged to read the original articles and apply their own critical judgment.
AI for Education
Explore each practice in its dedicated space, featuring a Synthesis with key applications and a filterable Table of real-world implementations. Select a practice below.
Applications
Generative AI is being widely adopted across disciplines to substitute, augment, modify, and redefine various facets of feedback provision in education. At the substitution level, it primarily reduces teacher workload by supporting qualitative feedback delivery for assignments and exercises, while augmentation can enhance feedback quality through alignment with established pedagogical frameworks. Modification transforms feedback administration through automation, enabling immediate delivery and possible integration with learning management systems. Redefinition more profoundly reimagines feedback provision, encompassing innovative approaches such as feedback within roleplay scenarios and simulation-based learning environments.
Substitution: Generating Feedback
Given generative AI’s sophisticated linguistic capabilities, the most extensive application has been in providing feedback on written assignments, including for example essays, reports, and translation tasks, where AI systems offer linguistic corrections, structural improvements, and guidance on content quality and argumentation [2], [6], [8], [19], [32], [40], [60], [81], [84], [95], [97], [98], [111], [124], [128], [129], [132], [167], [170], [192], [194], [196], [206], [213], [218], [220], [224], [226], [133], [142], [23], [7], [33], [22], [160], [93], [5], [11], [45], [88].
Feedback on programming assignments represents a natural extension of these linguistic abilities, as programming languages themselves consist of syntax, semantics, and structural rules. Generative AI can identify syntactic or logic errors, and provide feedback on code efficiency, style, readability and documentation [12], [49], [50], [53], [54], [70], [61], [62], [63], [64], [76], [77], [83], [157], [160], [166], [180], [191].
Beyond purely linguistic feedback on natural or programming languages, AI can be employed to provide feedback that draws upon knowledge domains. This involves evaluating students’ understanding of subject matter, assessing the quality of their reasoning and problem-solving approaches, and their disciplinary performance [36], [91], [99], [117], [122], [28], [71], [46], [48], [3], [110], [195], [161], [56], [15], [101], [18], [82]
Augmentation: Enhanced Feedback Design
Generative AI can enhance educational feedback in multiple ways: alignment with specific learning objectives, adherence to established pedagogical frameworks, application of evidence-based feedback principles, incorporation of domain-specific knowledge of errors, implementation of thoughtful design principles, and feedback delivery through diverse modalities and styles [19], [168], [85], [9], [109], [230], [3], [8], [27], [172], [197], [65], [63], [82], [171], [73].
Modification: Automatic Integrated Feedback
The affordance of generative AI to generate adapted content on the fly enables modification of feedback delivery in several ways, including automation of feedback provision and it’s integration with learning platforms, and realtime or immediate delivery of feedback that is personalized or adapted to individual learner needs by representing a dynamic student state. These applications are prevalent in the computer science domain [157], [166], [53], [69], [63], [228], [85], [74], [190], [66], [67], [79], and in the educational field more generally [173], [225], [15], [176], [3], [28], [156], [170], [82], [171], [222]. Realtime delivery of feedback is particularly relevant for language learning, where students can get feedback on speaking performance during realtime language practice [149], [108], [9], [137], [24], [230].
Redefinition: Interactive Feedback and Simulations
Finally, generative AI can redefine feedback integration by creating roleplay and simulation-based learning environments where feedback is either embedded within the interactive experience or delivered immediately afterward. These applications are most prevalent in the medical field [158], [25], [26], [43], [89], [94], [115], [118], [238], [29], [99], with some exceptions in physics [59], [202] and teacher education [13]
Measures and Outcomes
Feedback Quality
A first important outcome is the actual quality of AI-generated feedback itself, and how it compares to feedback created by human experts. Several studies have argued that AI-generated feedback can be effective as a support tool for teachers [22], [160], [5], [73], [220] and/or be comparable to or exceed human feedback [155], [11], [23], [53], [28], [3], [61], [84], [122]. However, other studies have reported mixed results, noting that AI-generated feedback can be ineffective or less effective in some areas [12], [98], [191], [192], [13] and/or low comparability with human feedback [41], [6], [86], [18], or inconsistent and unreliable [156], [62], [76], [50].
Feedback Perception
Beyond the objective quality of AI-generated feedback, students’ subjective experience of this feedback represents a crucial factor in determining its ultimate educational impact. Research reveals that students’ pre-existing attitudes toward AI technology influence their perception of feedback quality beyond the actual content of the feedback itself [169], [168]. Interestingly, students often cannot reliably distinguish between AI-generated and human-generated feedback, frequently rating AI feedback quite favorably when they are unaware of its source [83], [46]. However, student preferences vary considerably when the source is disclosed. Some studies report favorable attitudes toward AI feedback or relatively even preferences between AI and human-generated responses [97] [93], [151]. Conversely, other research indicates that students prefer human feedback and express skepticism toward AI-generated responses [52], [4]. This perception deficit can be partially addressed through the design of natural, flexible interaction experiences[57].
Engagement
Several studies have investigated the impact of AI-generated feedback on student engagement, with many reporting positive effects across cognitive, affective, and behavioral dimensions [141], [194], [44], [222], [176], [15], [59]. This enhanced engagement with feedback has been associated with improved subsequent revision behaviors [66], [72], [167], [32]. However, research also reveals potential drawbacks. AI-generated feedback can overwhelm students cognitively and prove time-consuming to process effectively [7], [15], [40]. Additionally, excessive usage patterns have been linked to decreased performance outcomes [157]. More broadly, some studies suggest that student engagement with feedback remains generally low regardless of source, showing little improvement when AI-generated feedback is provided compared to human feedback [109], [81].
Educational Outcomes
Research examining the actual educational impact of AI-generated feedback has mostly painted an encouraging picture. In writing and language learning contexts, AI feedback has been associated with improved skills and performance [95], [196], [45], [114], [128], [193], [226], [154], [140], [142], [184], [108], [171]. Similar positive outcomes extend beyond language domains, with studies documenting enhanced learning performance across various educational disciplines [89], [201], [85], [15] as well as improved academic achievement [222], [172], [79], [97], sometimes even suggesting that AI feedback may help reduce disparities in learning outcomes [176]. These benefits are often mediated by increased student engagement [235], [123], [59]. However, these positive findings are not universal. Some studies have explicitly reported a lack of significant learning gains from AI-generated feedback [56], [157].
Beyond domain-specific learning outcomes, AI-generated feedback has been linked to various psychological and metacognitive benefits. These include increased motivation [154], [85], [94], reduced anxiety [42], [144], enhanced autonomy and perceived competence or self-efficacy [44], [165], [219], [137], and improvements in self-regulated learning [9], [172].
Limitations and Recommendations
Enhancing AI Feedback
A critical challenge is countering generic AI-generated feedback that lacks both domain-specificity and pedagogical effectiveness. To address this limitation, researchers have explored several promising approaches. Some studies implement retrieval-arugmented generation (RAG) to ensure that feedback is contextually appropriate by drawing from relevant educational resources and domain-specific knowledge bases[48], [49], [230]. Other research has focused on fine-tuning language models using specialized datasets tailored to particular academic disciplines, or employing reinforcement learning techniques to guide models toward producing feedback that aligns with educator objectives [207], [230], [3]. These approaches aim to move beyond one-size-fits-all feedback toward more specialized, contextually appropriate responses.
Multi-step AI systems can further enhance feedback quality by decomposing the complex feedback generation process into distinct, specialized stages. Rather than attempting to evaluate student work and generate pedagogical guidance simultaneously, these systems can separate submission assessment from instructional feedback delivery. For instance, one component might focus exclusively on evaluating content quality and another on writing quality, while a subsequent stage incorporates pedagogical principles to craft appropriate feedback. A third validation layer can then assess the generated feedback against established quality criteria to ensure accuracy and educational effectiveness. This modular approach can be implemented through various architectural strategies. Multi-component systems may combine LLM-based modules with traditional rule-based components, capitalizing on the strengths of each approach. Alternatively, multi-agent frameworks can deploy several instances of a language model, each guided by distinct system prompts that define specific roles in the feedback pipeline. Some implementations strategically match different LLMs to particular tasks based on their relative strengths [110], [60], [68], [163], [53], [99], [230].
Designing Effective Human-AI Interaction
Studies consistently concur that hybrid human-AI workflows, where teachers curate and moderate LLM output, are preferable to fully autonomous tutors. These approaches combine scalable automation with pedagogical expertise, offering complementary strengths where teachers can for example address higher-order concerns while AI handles surface-level feedback [173], [230], [167], [161].
Furthermore, effective design requires balancing educational goals with practical usability while navigating inherent tensions between learning effectiveness and user satisfaction. Key principles include promoting transparency about AI limitations, maintaining instructor control, and creating natural, flexible interaction experiences [63], [57]. Systems must also leverage generative AI’s capabilities to promote inclusion and ensure equitable access for diverse learner populations [225], [208].
However, preventing over-reliance on AI tools remains crucial to avoid diminishing critical thinking and creativity. Both educators and students require comprehensive training in AI literacy and feedback literacy to ensure responsible and effective use of these technologies. These efforts should operate within regulatory frameworks that prioritize responsible use and support ethically grounded pedagogical implementations rather than purely technical solutions [198], [203], [159], [223], [232], [214], [174], [153].
System Implementation Recommendations
Successful deployment requires attention to system scalability and performance optimization to handle educational workloads effectively [54], [20], [228], [171]. Furthermore, privacy compliance represents a critical concern, as educational data requires special protection measures [230]. To address privacy concerns while maintaining functionality, institutions should consider implementing open-source LLMs that allow for local deployment and greater data control [70], [77], [160].
Conclusion and Future Directions
Generative AI can transform educational feedback by enabling immediate feedback that is aligned with pedagogical principles, and supporting entirely new formats such as feedback within simulations and interactive feedback. Future work should focus on advancing feedback quality through refined approaches such as retrieval-augmented generation and multi-step systems, with consistent expert validation as an essential component. Research should also emphasize AI’s potential for creating innovative feedback delivery methods that incorporate simulation and real-time interaction.
Note on data verification: Not all papers in this table have been double-checked by a human expert. Papers highlighted (marked with a reference number in the Ref column) have been cited in the synthesis and verified. Please consult original publications and apply your own critical judgment.
About the Catalog Creation
This catalog is built through a systematic, multi-stage process including data retrieval and processing, combining AI-powered analysis and essential human oversight. Our goal is to create a dynamic resource that is regularly updated.
We do not evaluate the scientific quality, methodological rigor, or validity of the studies referenced in this catalog. Our role is to point to relevant literature and synthesize available information—not to certify its credibility. Users must consult the original publications and apply their own critical judgment when interpreting the findings.
Select a topic below to learn more about our methodology.
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
How to Cite This Catalog
If you use this catalog in your research or work, please use the citation information below. Click on a style to expand it and copy the citation to your clipboard.
Uittenhove, K. (2025). AI for Education: A Living Catalog of Emerging Practice (Version v1.0.0). Zenodo. https://doi.org/10.5281/zenodo.15633017
Uittenhove, Kim. AI for Education: A Living Catalog of Emerging Practice. Version v1.0.0, Zenodo, 10 June 2025, doi:10.5281/zenodo.15633017.
Uittenhove, Kim. 2025. "AI for Education: A Living Catalog of Emerging Practice." Zenodo. https://doi.org/10.5281/zenodo.15633017.
Uittenhove, K. (2025) 'AI for Education: A Living Catalog of Emerging Practice', Zenodo. doi: 10.5281/zenodo.15633017.
1. Uittenhove K. AI for Education: A Living Catalog of Emerging Practice [Internet]. Zenodo; 2025 Jun 10. Available from: https://doi.org/10.5281/zenodo.15633017
[1] K. Uittenhove, "AI for Education: A Living Catalog of Emerging Practice," Zenodo, Jun. 10, 2025. doi: 10.5281/zenodo.15633017.@misc{uittenhove_2025,
author = {Uittenhove, Kim},
title = {{AI for Education: A Living Catalog of Emerging Practice}},
month = {June},
year = {2025},
publisher = {Zenodo},
version = {v1.0.0},
doi = {10.5281/zenodo.15633017},
url = {[https://doi.org/10.5281/zenodo.15633017](https://doi.org/10.5281/zenodo.15633017)}
}