Automating Exam Generation for Assessment
One of the most significant and difficult tasks in education is the creation of valid, reliable, and useful assessments esp. exams. If we consider the question level, we generally trade of ease of creation (where so called subjective questions dominate: it’s very easy to write essay questions) to ease of marking (where so-called objective questions such as multiple choice questions (MCQs) dominate: it’s very hard to mark essay questions). Even if we succeed in question generation, a valid, reliable, and useful exam is not merely a set of individually good questions, but must also provide for good coverage of material and an appropriate balance of difficulty, among other properties.
Recently there has been some progress in terms of question generation (either from text or from structured sources such as ontologies or linked data), but the field is still immature and comparatively little work has been done on exam generation. The goal of this thesis is to significantly advance both areas. There is the strong possibility of industrial collaboration on this project.
School and University funding is available on a competitive basis.
For further details, please see our funding pages here: http://www.cs.manchester.ac.uk/study/postgraduate-research/programmes/phd/funding/
The minimum requirements to get a place in our PhD programme are available from:
How good is research at University of Manchester in Computer Science and Informatics?
FTE Category A staff submitted: 44.86
Research output data provided by the Research Excellence Framework (REF)
Click here to see the results for all UK universities