James Rennie Christie
Automated assessment of non-objective textual submissions.
Christie, James Rennie
Authors
Contributors
H.I. Ellington
Supervisor
S. Earl
Supervisor
R. Butler
Supervisor
Abstract
The author sought to explore the methodology of automated marking of non-objective textual submissions. In so doing he tasked himself to discover how far automated assessment of text could be applied in the range of submissions from the single word to a multi-sentence, multi-paragraph piece of text. In the course of his research, the author had to determine the extent and severity of any problems associated with manual marking of such submissions. In fact, the thesis informs the reader of the myriad problems associated with, and generated by, the subjectivity inherent in the manual marking process and the author indicates to what extent automated marking can remove or reduce these problems. In parallel, the author had to determine the existence, if any, of automated assessment of such type of submissions. The literature survey did show that some work had been done in this subject area, but that most of the work dates from the 1960's for style marking and only recently has some work on content marking been published. The author devised algorithms and implemented them in a software system, called SEAR (Schema Extract Assess and Report), which was intended to process word-processed submissions and award marks. Lack of suitable marked data sets prevented the full development of the style algorithm. However, the author was able to demonstrate that content marking is possible with a range of submissions. The proposed style algorithm is based on a novel idea of determining a set of metrics that could be applied to all textual submissions. The data structure that was developed for the marking of content is unique. The author also compiled a set of potential criteria for use in the evaluation of the methodology of automated marking of textual submissions. These criteria were applied to his software system.
Citation
CHRISTIE, J.R. 2003. Automated assessment of non-objective textual submissions. Robert Gordon University, PhD thesis. Hosted on OpenAIR [online]. Available from: https://doi.org/10.48526/rgu-wt-1585045
Thesis Type | Thesis |
---|---|
Deposit Date | Feb 3, 2022 |
Publicly Available Date | Feb 3, 2022 |
Keywords | Text recognition; Automation; Artificial intelligence; Marking; Examinations; Assessments |
Public URL | https://rgu-repository.worktribe.com/output/1585045 |
External URL | https://doi.org/10.48526/rgu-wt-1585045 |
Award Date | Feb 28, 2003 |
Files
CHRISTIE 2003 Automated assessment of non-objective
(15.1 Mb)
PDF
Licence
https://creativecommons.org/licenses/by-nc/4.0/
Copyright Statement
© The Author.
Downloadable Citations
About OpenAIR@RGU
Administrator e-mail: publications@rgu.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search