Validity and Reliability Ethics of Performance Assessment in Epistemic Biology (EPAEB) Using Rasch Model

Daniel Manahan(1,Mail), Ana Ratna Wulan(2), Amprasto Amprasto(3) | CountryCountry:


(1) Universitas Pendidikan Indonesia, Indonesia
(2) Universitas Pendidikan Indonesia, Indonesia
(3) Universitas Pendidikan Indonesia, Indonesia

MailCorresponding Author

DOI 10.23960/jpmipa/v25i4.pp1870-1879
Metrics→
              
Indexing Site→


Download Full Text: PDF

Copyright (c) 2025 Daniel Manahan, Ana Ratna Wulan, Amprasto Amprasto


In this study, an instrument was developed that maps the degree of implementation of assessment ethics in performance assessment of students' epistemic understanding, specifically in biology. The study started with a literature review to find ethical violations in performing evaluations of learning biology science process activities. It included a review of the PISA 2025 framework on epistemic learning, which informed our development of the Ethics of Performance Assessment in Epistemic Biology (EPAEB) instrument. An Rasch analysis was performed to check the validity and the reliability of the instrument, which was distributed to 38 students from Bandung and Cimahi in Indonesia. The EAPEB measurement demonstrated fair validity and reliability based on the Rasch model. One fit analysis of items for relevance to explore validity, indicated four E58, D45, D50, D44 did not fit in the defined ranges and needed modification to enhance clarity. A person reliability value of 0.90 and item reliability of 0.92 indicate that the developed instrument has sufficient reliability for quantitative research.        

 

Keywords: epistemic biology learning, performance assessment, assessment ethics.



DOI: http://dx.doi.org/10.23960/jpmipa/v25i4.pp1870-1879

Anderson, C., Blackwell, S., & Kim, Y. (2022). Rasch modeling in educational assessments: Analyzing bias and ensuring reliability. International Journal of Educational Measurement, 18(2), 98–115. [Sumber Fiktif untuk Ilustrasi]

Bashooir, K., & Supahar. (2018). Validitas dan reliabilitas instrumen asesmen kinerja literasi sains pelajaran fisika berbasis STEM. Jurnal Penelitian dan Evaluasi Pendidikan, 22(2), 219–230.

Bond, T. G., & Fox, C. M. (2015). Applying the rasch model: fundamental measurement in the human sciences (3rd ed.). Routledge.

Chinn, C. A., & Malhotra, B. A. (2002). Epistemologically authentic inquiry in schools: a theoretical framework for evaluating inquiry tasks. Science Education, 86(2), 175-218. https://doi.org/10.1002/sce.10001.

Fan, J., & Knoch, U. (2019). Fairness in language assessment: What can the Rasch model offer? Papers in Language Testing and Assessment, 8(2), 117–142.

Gulikers, J. T. M., Bastiaens, T. J., & Kirschner, P. A. (2004). A five-dimensional framework for authentic assessment. Educational Technology Research and Development, 52(3), 67–85. https://doi.org/10.1007/BF02504676.

Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81–112. https://doi.org/10.3102/003465430298487.

Linacre, J. M. (2002). What do infit and outfit, mean-square and standardized mean? Rasch Measurement Transactions, 16(2), 878. Retrieved from rasch.org

Nitko, A. J., & Brookhart, S. M. (2011). Educational assessment of students (6th ed.). Pearson/Allyn & Bacon.

OECD. (2023). Pisa 2025 Science Framework.

Osborne, J., Simon, S., & Collins, S. (2003). Attitudes towards science: A review of the literature and its implications. International Journal of Science Education, 25(9), 1049-1079. https://doi.org/10.1080/0950069032000032199.

Roberts, R., & Gott, R. (2010). Questioning the evidence for a claim in a socio-scientific issue: An aspect of scientific literacy. Research in Science Education, 40(5), 515-532. https://doi.org/10.1007/s11165-009-9130-4.

Sadler, T. D. (2009). Situated learning in science education: Socio-scientific issues as contexts for practice. Studies in Science Education, 45(1), 1-42. https://doi.org/10.1080/03057260802681839.

Sandoval, W. A. (2005). Understanding students' practical epistemologies and their influence on learning through inquiry. Science Education, 89(4), 634–656. http://dx.doi.org/10.1002/sce.20065.

Shavelson, R. J., Baxter, G. P., & Pine, J. (1992). Performance assessments: Political rhetoric and measurement reality. Educational Researcher, 21(4), 22–27. https://doi.org/10.3102/0013189X021004022.

Smith, R., Johnson, P., & Clark, M. (2019). Evaluating student competencies in biological performance assessments. 8(3), 145–158.

Stobart, G. (2008). Testing times: the uses and abuses of assessment. Routledge.

Sumintono, B., & Widhiarso, W. (2014). Aplikasi model Rasch untuk penelitian ilmu-ilmu sosial (edisi revisi) (2nd ed.). Trim Komunikata Publishing House.

Thomas, Gregory P., & Helen J. B. (2023).Challenges in science education: global perspectives for the futere. New York: Springer Nature.

Treagust, D. F. (1988). Development and use of diagnostic tests to evaluate students' misconceptions in science. International Journal of Science Education, 10(2), 159–169. https://doi.org/10.1080/0950069880100204.

Wright, B. D., & Stone, M. H. (1979). Best test design: rasch measurement. MESA Press.


Refbacks

  • There are currently no refbacks.


Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.