A comparative study of methods for a priori prediction of MCQ difficulty

Ghader Kurdi*, Jared Leo, Nicolas Matentzoglu, Bijan Parsia, Uli Sattler, Sophie Forge, Gina Donato, Will Dowling, Dagmar Gromann

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Successful exams require a balance of easy, medium, and difficult questions. Question difficulty is generally either estimated by an expert or determined after an exam is taken. The latter provides no utility for the generation of new questions and the former is expensive both in terms of time and cost. Additionally, it is not known whether expert prediction is indeed a good proxy for estimating question difficulty. In this paper, we analyse and compare two ontology-based measures for difficulty prediction of multiple choice questions, as well as comparing each measure with expert prediction (by 15 experts) against the exam performance of 12 residents over a corpus of 231 medical case-based questions that are in multiple choice format. We find one ontology-based measure (relation strength indicativeness) to be of comparable performance (accuracy = 47%) to expert prediction (average accuracy = 49%).

Original languageEnglish
Pages (from-to)449-465
Number of pages17
JournalSemantic Web
Volume12
Issue number3
DOIs
Publication statusPublished - 2021

Keywords

  • automatic question generation
  • difficulty modelling
  • difficulty prediction
  • multiple choice questions
  • Ontologies
  • semantic web
  • student assessment

Fingerprint

Dive into the research topics of 'A comparative study of methods for a priori prediction of MCQ difficulty'. Together they form a unique fingerprint.

Cite this