Questions with answers are rarely used as e-learning assessment tools because of the resulting high workload for the teacher/tutor that should grade them, which can be mitigated by doing peer-assessment. In OpenAnswer we modeled peer-assessment as a Bayesian network connecting the sub-networks representing any participating student to the corresponding answers of her graded peers. The model has shown good ability to predict (without further info from the teacher) the exact teacher’s grade (ground truth) from the peer grades only, and a very good ability to predict it within 1 mark from the right one. In this paper we explore changes to the OpenAnswer model to improve its predictions. The experimental results, obtained by simulating teacher’s grading on real datasets, show the improved predictions.

Improving peer assessment modeling of teacher’s grades, the case of OpenAnswer / DE MARSICO, Maria; Sterbini, Andrea; Temperini, Marco. - 10108:(2017), pp. 601-613. (Intervento presentato al convegno 1st International Symposium on Emerging Technologies for Education, SETE 2016 Held in Conjunction with ICWL 2016 tenutosi a Rome; Italy nel 2016) [10.1007/978-3-319-52836-6_64].

Improving peer assessment modeling of teacher’s grades, the case of OpenAnswer

DE MARSICO, Maria;STERBINI, Andrea
;
TEMPERINI, Marco
2017

Abstract

Questions with answers are rarely used as e-learning assessment tools because of the resulting high workload for the teacher/tutor that should grade them, which can be mitigated by doing peer-assessment. In OpenAnswer we modeled peer-assessment as a Bayesian network connecting the sub-networks representing any participating student to the corresponding answers of her graded peers. The model has shown good ability to predict (without further info from the teacher) the exact teacher’s grade (ground truth) from the peer grades only, and a very good ability to predict it within 1 mark from the right one. In this paper we explore changes to the OpenAnswer model to improve its predictions. The experimental results, obtained by simulating teacher’s grading on real datasets, show the improved predictions.
2017
1st International Symposium on Emerging Technologies for Education, SETE 2016 Held in Conjunction with ICWL 2016
Automatic correction of OpenAnswers; Bayesian networks; Peer assessment modeling; Theoretical Computer Science; Computer Science (all)
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Improving peer assessment modeling of teacher’s grades, the case of OpenAnswer / DE MARSICO, Maria; Sterbini, Andrea; Temperini, Marco. - 10108:(2017), pp. 601-613. (Intervento presentato al convegno 1st International Symposium on Emerging Technologies for Education, SETE 2016 Held in Conjunction with ICWL 2016 tenutosi a Rome; Italy nel 2016) [10.1007/978-3-319-52836-6_64].
File allegati a questo prodotto
File Dimensione Formato  
Marsico_Improving-Peer-Assessment_2017.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 553.18 kB
Formato Adobe PDF
553.18 kB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/948301
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
social impact