Skip to main content

Item Response Model with Lower Order Parameters for Peer Assessment

  • Conference paper
  • First Online:
Artificial Intelligence in Education (AIED 2015)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9112))

Included in the following conference series:

  • 6081 Accesses

Abstract

Peer assessment has become popular in recent years. However, in peer assessment, a problem remains that reliability depends on the rater characteristics. For this reason, some item response models that incorporate rater parameters have been proposed. However, in previous models, the parameter estimation accuracy decreases as the number of raters increases because the number of rater parameters increases drastically. To solve that problem, this article presents a proposal of a new item response model for peer assessment that incorporates rater parameters to maintain as few rater parameters as possible.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Patz, R.J., Junker, B.W.: Applications and extensions of mcmc in irt: Multiple item types, missing data, and rated responses. Journal of Educational and Behavioral Statistics 24, 342–366 (1999)

    Article  Google Scholar 

  2. Patz, R.J., Junker, B.W., Johnson, M.S.: The hierarchical rater model for rated test items and its application to large-scale educational assessment data. Journal of Educational and Behavioral Statistics 27(4), 341–366 (1999)

    Article  Google Scholar 

  3. Piech, C., Huang, J., Chen, Z., Do, C., Ng, A., Koller, D.: Tuned models of peer assessment in MOOCs. In: Proceedings of Sixth International Conference of MIT’s Learning International Networks Consortium (2013)

    Google Scholar 

  4. Ueno, M., Okamoto, T.: Item response theory for peer assessment. In: Eighth IEEE International Conference on Advanced Learning Technologies. ICALT 2008, pp. 554–558 (2008)

    Google Scholar 

  5. Usami, S.: A polytomous item response model that simultaneously considers bias factors of raters and examinees: Estimation through a markov chain monte carlo algorithm. The Japanese Journal of Educational Psychology 58(2), 163–175 (2010)

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Masaki Uto .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Uto, M., Ueno, M. (2015). Item Response Model with Lower Order Parameters for Peer Assessment. In: Conati, C., Heffernan, N., Mitrovic, A., Verdejo, M. (eds) Artificial Intelligence in Education. AIED 2015. Lecture Notes in Computer Science(), vol 9112. Springer, Cham. https://doi.org/10.1007/978-3-319-19773-9_119

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-19773-9_119

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-19772-2

  • Online ISBN: 978-3-319-19773-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics