Loading [a11y]/accessibility-menu.js
Appropriate Number of Raters for IRT Based Peer Assessment Evaluation of Programming Skills | IEEE Conference Publication | IEEE Xplore

Appropriate Number of Raters for IRT Based Peer Assessment Evaluation of Programming Skills


Abstract:

In order to examine the reliability of peer assessment settings for the evaluation of programming skills using peer assessment and the item response theory (IRT) techniqu...Show More

Abstract:

In order to examine the reliability of peer assessment settings for the evaluation of programming skills using peer assessment and the item response theory (IRT) technique in a small classes, the optimal conditions such as the numbers of peers and the number of tasks are investigated using parameters extracted from the surveyed data. The survey data consisted of 31 students whose partial participation consisted of joining three peer assessments out of the 5 sessions during which these took place. Peer rating conditions such as the number of peer raters or tasks are examined using mean expected standard errors. Also, the relationship between instructor's ratings and estimated ability is examined using the IRT model to look at variations in the number of peer raters and tasks. The results provide evidence that a set of guidelines could better organise peer assessment of the evaluation of programmina skills in actual course settings,
Date of Conference: 06-08 November 2024
Date Added to IEEE Xplore: 16 January 2025
ISBN Information:

ISSN Information:

Conference Location: Paris, France
References is not available for this document.

I. Introduction

The assessment of programming skills is not simple, as programming techniques and concepts of code writing need to be evaluated correctly. As a solution to this, peer assessment has been introduced into programming practice classes [1]–[3], and reports of its effectiveness regarding the stimulation of proactive learning have been positive [4], [5]. Therefore, peer assessment can be applied to a large scale classes [6].

Select All
1.
K. Georgouli and C. Sgouropoulou, “Collaborative peer-evaluation learning results in higher education programming-based courses,” in Proc. ICBL2013 Int. Conf. on Interactive Computer aided Blended Learning, 2013, pp. 309–314.
2.
X.-M. Wang, G.-J. Hwang, Z.-Y. Liang, and H.-Y. Wang, “Enhancing students' computer programming performances, critical thinking awareness and attitude towards programming: An online peer-assessment attempt,” Educational Technology & Society, vol. 20, pp. 58–68, 2017.
3.
A.-K. Amal, K. and D. Marie, “Evaluating a peer assessment approach in introductory programming course,” in Proc. UKICER '20, 2020, pp. 51–58.
4.
G. Thomas, D. Martin, and K. Pleasants, “Using self- and peer-assessment to enhance students' future-learning in higher education,” Journal of University Teaching & Learning Practice, vol. 8, no. 1, pp. 1–17, 2011.
5.
M. A. Jackson, A. Tran, M. P. Wenderoth, and J. H. Doherty, “Peer vs. self-grading of practice exams: Which is better? ” CBE Life Science Education, vol. 17, no. ar44, pp. 1–8, 2018.
6.
R. Alcarria, B. Bordel, D. M. deAndres, and T. Robles, “Enhanced peer assessment in MOOC evaluation through assignment and review analysis,” International Journal of Emerging Technologies in Learning, vol. 13, no. 1, pp. 206–219, 2018.
7.
M. De Marsico, A. Sterbini, and M. Temperini, “Towards a quantitative evaluation of the relationship between the domain knowledge and the ability to assess peer work,” in Proceedings of iTHET 2015, Lisbon, Portugal, 2015, pp. 1–6.
8.
M. De Marsico, A. Sterbini, and M. Temperini, “Modeling peer assessment as a personalized predictor of teacher's grades: the case of OpenAnswer,” in Proceedings of iTHET 2016, Istanbul, Turkey, 2016, pp. 1–5.
9.
R. L. Linn, Educational measurement, 3 rd edition. New York, USA: Collier Macmillan Publishers, 1989.
10.
M. Uto, “Rater-effect IRT model integrating supervised LDA for ac-curate measurement of essay writing ability,” in Proc. International Conference on Artificial Intelligence in Education, 2019, pp. 494–506.
11.
M. Uto and M. Ueno, “Empirical comparison of item response theory models with rater's parameters,” Heliyon, vol. 4, pp. 1–32, 2018.
12.
M. Nakayama, M. Uto, M. Temperini, and F. Sciarrone, “Estimating ability of programming skills using IRT based peer assessments,” in Proceedings of 19th International Conference on Information Technol-ogy Based Higher Education and Training (ITHET), 2021, pp. 1–6.
13.
M. Nakayama, F. Sciarrone, M. Temperini, and M. Uto, “A peer assessment characteristic for evaluating programming skills using IRT estimation,” in Proceedings of 20th International Conference on In-formation Technology Based Higher Education and Training (ITHET), 2022, pp. 1–8.
14.
M. Uto, D.-T. Nguyen, and M. Ueno, “Group optimization to maxi-mize peer assessment accuracy using item response theory and integer programming,” IEEE Transactions on Learning Technologies, vol. 13, no. 1, pp. 91–106, 2020.
15.
M. Nakayama, F. Sciarrone, M. Uto, and M. Temperini, “Estimating student's performance based on item response theory in a MOOC environment with peer assessment,” in Proceedings of 10th International Conference in Methodologies and Intelligent Systems for Technology Enhanced Learning, L’ Aquila, Italy, 2020, pp. 1–10.
16.
M. Nakayama, F. Sciarrone, M. Temperini, and M. Uto, “An item response theory approach to enhance peer assessment effectiveness in massive open online course,” International Journal of Distance Education Technologies, vol. 20, no. 1, pp. 1–19, 2022.
17.
M. Nakayama, F. Sciarrone, M. Uto, and M. Temperini, “Impact of the number of peers on a mutual assessment as learner's performance in a simulated MOOC environment using the IRT model,” in Proceedings of 24th International Conference Information Visualisation (IV), 2020, pp. 483–487.
18.
M. Uto and M. Ueno, “A generalized many-facet Rasch model and its Bayesian estimation using Hamiltonian Monte Carlo,” Behaviormetrika, vol. 47, pp. 469–496, 2020.
19.
M. Uto and M. Ueno, “Item response theory without restriction of equal interval scale for rater's score,” in Proc. International Conference on Artificial Intelligence in Education, 2018, pp. 363–368.

Contact IEEE to Subscribe

References

References is not available for this document.