Generative Artificial Intelligence as an aid for the assessment of Early Modern History student work in higher education
DOI:
https://doi.org/10.1590/1983-3652.2026.59410Palabras clave:
Early Modern History, Hybrid assessment model, Generative Artificial Intelligence, Higher education, Analytical rubricsResumen
This article examines the use of generative artificial intelligence (AI) as a supportive instrument in the assessment of undergraduate academic work in Early Modern American History. Drawing on a hybrid evaluation model that integrates human judgment with AI-assisted assessment, the study implemented three successive stages of correction: an initial assessment based on a basic rubric, a second evaluation using an advanced analytical rubric, and a final phase involving the critical recalibration of prior results. The corpus consisted of 21 research-based assignments produced by fourth-year History students and evaluated according to technical, historiographical, and critical-thinking criteria. The findings show that, once properly calibrated, the AI system was able to discriminate effectively between different levels of academic quality and to identify recurring patterns of historical reasoning in student work. A comparison between human-generated and AI-generated grades revealed a high degree of convergence, alongside significant divergences. While the AI demonstrated greater sensitivity to methodological rigor and critical engagement, human evaluators tended to prioritize formal presentation and technical aspects of writing. Rather than replacing instructor judgment, the proposed model reframes assessment as a more rigorous, equitable, and reflexive pedagogical process. The study suggests that, when embedded within demanding, transparent, and reviewable pedagogical frameworks, generative artificial intelligence can operate as an epistemic agent within the Humanities, contributing meaningfully to the evaluation of complex historical learning outcomes.
Descargas
Referencias
ALVAREZ, Humberto. Evaluación del pensamiento histórico de estudiantes de secundaria a través de la construcción de narrativas históricas sobre los pueblos originarios de Chile. Años 90, n. 28, p. 1–28, 2021. Available from: https://doi.org/10.22456/1983-201X.111650.
ALVAREZ, Humberto. La inteligencia artificial como catalizador en la enseñanza de la Historia: Retos y posibilidades pedagógicas. Revista Docentes 2.0, v. 16, n. 2, p. 318–325, 2023. Available from: https://doi.org/10.37843/rted.v16i2.426.
ANDRADE, Heidi. A critical review of research on student self-assessment. Frontiers in Education, v. 4, p. 87, 2019. Available from: https://doi.org/10.3389/feduc.2019.00087.
ASHBY, Rosalyn; LEE, Peter. Progression in historical understanding among students ages 7-14. In: STEARNS, Peter; SEIXAS, Peter; WINEBURG, Sam (eds.). Knowing, teaching and learning history: National and international perspectives. New York: New York University Press, 2000. p. 199–222.
BAUER, Elisabeth et al. AI-based adaptive feedback in simulations for teacher education: An experimental replication in the field. Journal of Computer Assisted Learning, v. 41, n. 1, e13123, 2025. Available from: https://doi.org/10.1111/jcal.13123.
CARRASCO, Antonio. Reinventing the teaching of early Modern History in secondary school: The use of ChatGPT to enhance learning and educational innovation. Studia Historica: Historia Moderna, v. 45, n. 1, p. 101–145, 2023. Available from: https://doi.org/10.14201/shhmo2023451101146.
CARRASCO, Antonio. Perceptions of generative artificial intelligence among university early Modern History students. Tiempos Modernos, v. 14, n. 49, p. 269–285, 2024.
CARRETERO, Mario; GARTNER, Elisa. Artificial intelligence and historical thinking: A dialogic exploration of ChatGPT. Studies in Psychology, v. 45, n. 1, p. 80–102, 2024. Available from: https://doi.org/10.1177/02109395241241379.
CHEN, Chih-Ming; WITT, Barbara; LIN, Chun-Yu. A knowledge graph analysis tool of people and organizations to facilitate digital humanities research. Data Technologies and Applications, v. 59, n. 1, p. 82–110, 2025. Available from: https://doi.org/10.1108/dta-01-2024-0009.
EVANGELISTA, Edmund De Leon. Ensuring academic integrity in the age of ChatGPT: Rethinking exam design, assessment strategies, and ethical AI policies in higher education. Contemporary Educational Technology, v. 17, n. 1, ep559, 2025. Available from: https://doi.org/10.30935/cedtech/15775.
GAMMOH, Leen. ChatGPT risks in academia: Examining university educators’ challenges in Jordan. Education and Information Technologies, v. 30, n. 3, p. 3645–3667, 2025. Available from: https://doi.org/10.1007/s10639-024-13009-y.
HOLMES, Wayne; BIALIK, Maya; FADEL, Charles. Artificial intelligence in education: Promises and implications for teaching and learning. [S. l.]: Center for Curriculum Redesign, 2019.
KALDARAS, Leonora; YOSHIDA, Nicholas; HAUDEK, Kevin. Rubric development for AI-enabled scoring of three-dimensional constructed-response assessment aligned to NGSS learning progression. Frontiers in Education, v. 7, p. 983055, 2022. Available from: https://doi.org/10.3389/feduc.2022.983055.
KANSTEINER, Wulf. Digital doping for historians: Can history, memory, and historical theory be rendered artificially intelligent? History and Theory, v. 61, n. 4, p. 119–133, 2022. Available from: https://doi.org/10.1111/hith.12282.
KNOX, Jeremy. Artificial intelligence and education in China. Learning, Media and Technology, v. 45, n. 3, p. 298–311, 2020. Available from: https://doi.org/10.1080/17439884.2020.1754236.
LEME, André. Artificial history? Inquiring ChatGPT on historiography. Rethinking History, v. 27, n. 4, p. 709–749, 2023.
LUCKIN, Rosemary; HOLMES, Wayne. Intelligence unleashed: An argument for AI in education. [S. l.]: Pearson, 2016.
LUO, Shuqiong; ZOU, Di. University learners’ readiness for ChatGPT‐assisted English learning: Scale development and validation. European Journal of Education, v. 60, n. 1, e12886, 2025. Available from: https://doi.org/10.1111/ejed.12886.
MARTIN, Paul; KRANZ, David; GRAULICH, Nicole. Revealing rubric relations: Investigating the interdependence of a research-informed and a machine learning-based rubric in assessing student reasoning in chemistry. International Journal of Artificial Intelligence in Education, p. 1–39, 2024. Available from: https://doi.org/10.1007/s40593-024-00440-y.
NICOL, David; MACFARLANE, Debra. Formative assessment and self-regulated learning: A model and seven principles of good feedback practice. Studies in Higher Education, v. 31, n. 2, p. 199–218, 2006. Available from: https://doi.org/10.1080/03075070600572090.
OFEM, Usani et al. Students’ perceptions, attitudes and utilisation of ChatGPT for academic dishonesty: Multigroup analyses via PLS-SEM. Education and Information Technologies, v. 30, n. 1, p. 159–187, 2025. Available from: https://doi.org/10.1007/s10639-024-12850-5.
PANADERO, Ernesto; JONSSON, Anders. The use of scoring rubrics for formative assessment purposes revisited: A review. Educational Research Review, n. 9, p. 129–144, 2013. Available from: https://doi.org/10.1016/j.edurev.2013.01.002.
RÜSEN, Jörn. History: Narration, interpretation, orientation. New York: Berghahn, 2005.
RÜSEN, Jörn. Jörn Rüsen e o ensino de história. Curitiba: Editora UFPR, 2010.
SCOTT, Howard; SMITH, Matthew. Innovation from necessity: Digital technologies, teacher development and reciprocity with organisational innovation. Open Learning: The Journal of Open, Distance and e-Learning, v. 39, n. 2, p. 170–187, 2024. Available from: https://doi.org/10.1080/02680513.2024.2307627.
SEIXAS, Peter; MORTON, Tom. The big six historical thinking concepts. Toronto: Nelson Education, 2013.
SELWYN, Neil. Should robots replace teachers? AI and the future of education. Cambridge: Polity Press, 2021.
SVENSSON, Patrik. Beyond the big tent. In: GOLD, Matthew (ed.). Debates in the digital humanities. Minneapolis: University of Minnesota Press, 2012. p. 36–49. Available from: https://doi.org/10.5749/minnesota/9780816677948.003.0004.
TIRADO, Sergio et al. From human to machine: Investigating the effectiveness of the conversational AI ChatGPT in historical thinking. Education Sciences, v. 13, n. 8, p. 803, 2023. Available from: https://doi.org/10.3390/educsci13080803.
TOKTAS, Elif. Future scenarios of digital humanities and post-humanist education. Journal of Foresight and Public Health, v. 2, n. 1, p. 21–31, 2025.
TONDEUR, Jo et al. Preparing preservice teachers to teach with digital technologies: An update of effective SQD-strategies. Computers Education, p. 105262, 2025. Available from: https://doi.org/10.1016/j.compedu.2025.105262.
TOPPING, Keith et al. Enhancing peer assessment with artificial intelligence. International Journal of Educational Technology in Higher Education, v. 22, n. 1, p. 3, 2025. Available from: https://doi.org/10.1186/s41239-024-00501-1.
WASHBURN, Jeffrey; MCCUTCHEN, Jennifer. AI meets AI: ChatGPT as a pedagogical tool to teach American Indian history. Critical Humanities, v. 2, n. 2, p. 2, 2024. Available from: https://doi.org/10.33470/2836-3140.1037.
WINEBURG, Sam. Historical thinking and other unnatural acts: Charting the future of teaching the past. Philadelphia: Temple University Press, 2001.
ZHOU, Xue; SMITH, Christopher; AL-SAMARRAIE, Hosam. Digital technology adaptation and initiatives: A systematic review of teaching and learning during COVID-19. Journal of Computing in Higher Education, v. 36, n. 3, p. 813–834, 2024. Available from: https://doi.org/10.1007/s12528-023-09376-z.
Descargas
Publicado
Declaración de disponibilidad de datos
Research data is available in the repository
Número
Sección
Licencia
Derechos de autor 2026 Antonio Carrasco-Rodríguez, Humberto Álvarez Sepúlveda

Esta obra está bajo una licencia internacional Creative Commons Atribución 4.0.
Este es un artículo de acceso abierto que permite su uso, distribución y reproducción sin restricciones en cualquier medio siempre que se cite correctamente el artículo original.








