Lohr D, Berges MP, Chugh A, Kohlhase M, Müller D (2024)
Publication Language: English
Publication Type: Journal article, Original article
Publication year: 2024
DOI: 10.1111/jcal.13101
Background: Over the past few decades, the process and methodology of automated question generation (AQG) have undergone significant transformations. Recent progress in generative natural language models has opened up new potential in the generation of educational content.
Objectives: This paper explores the potential of large language models (LLMs) for generating computer science questions that are sufficiently annotated for automatic learner model updates, are fully situated in the context of a particular course, and address the cognitive dimension understand.
Methods: Unlike previous attempts that might use basic methods like ChatGPT, our approach involves more targeted strategies such as retrieval-augmented generation (RAG) to produce contextually relevant and pedagogically meaningful learning objects.
Results and Conclusions: Our results show that generating structural, semantic annotations works well. However, this success was not reflected in the case of relational annotations. The quality of the generated questions often did not meet educational standards, highlighting that although LLMs can contribute to the pool of learning materials, their current level of performance requires significant human intervention to refine and validate the generated content.
APA:
Lohr, D., Berges, M.-P., Chugh, A., Kohlhase, M., & Müller, D. (2024). Leveraging Large Language Models to Generate Course-specific Semantically Annotated Learning Objects. Journal of Computer Assisted Learning. https://doi.org/10.1111/jcal.13101
MLA:
Lohr, Dominic, et al. "Leveraging Large Language Models to Generate Course-specific Semantically Annotated Learning Objects." Journal of Computer Assisted Learning (2024).
BibTeX: Download