Identifikasi Muatan Differential Item Functioning Pada Data Ujian Nasional Matematika

Main Article Content

Samritin Samritin

Abstract

One of the test criteria used in ability testing is a test that is free from Differential Item Functioning (DIF). Tests containing DIF can harm certain groups and benefit certain groups as well. However, sometimes the standard tests used are unknowingly containing DIF. The National Examination must meet the DIF-free criteria to avoid erroneous decision-making regarding the results of the exam. The purpose of this study is to detect Differential Item Functioning (DIF) through a unidimensional-based item response theory approach using explorative quantitative research. The results of the national mathematics exam as data for detecting DIF. The analysis technique uses the Mantel-Haenszal and Logistic Regression. In addition, DIF detection focused on gender, namely groups of men and women. Estimation of item parameters was carried out with the help of BILOG-MG, and the identification of DIF content using the R program tool. The results of the analysis showed that of the 21 items analyzed using Mantel-Haenszal and Logistic Regression, one item (Number 27) was suspected to contain DIF. This item is likely or suspected to benefit the male group, because in completing it the students first illustrate it in the form of an image, making it easier to calculate the height of the building, so that a long enough analysis is needed to be able to answer it.

Article Details

How to Cite
Samritin, S. (2022). Identifikasi Muatan Differential Item Functioning Pada Data Ujian Nasional Matematika. Journal on Education, 4(4), 1675-1684. https://doi.org/10.31004/joe.v4i4.2508
Section
Articles

References

Angoff & William H. (1982). Use of Difficulty and Discrimination Indices for Detecting Item Bias in Handbook of Methods for Detecting Test Bias, ed. Ronard K. Beck. Baltimore: The Johns Hopkins University Press.
Augemberg, K. E. & Morgan, D. L. (2008). Differential Performance of Test Items by Geographical Regions. Paper dipresentasikan pada annual meeting ofthe National Council on Measurement in Education, New York.
Budiono. (2005). Perbandingan Metode Mantel-Haenszel, Sibtest, Regresi Logistik, dan Perbedaan Peluang Dalam Mendeteksi Keberbedaan Fungsi Butir. Jurnal Studi Dan Evaluasi Pendidikan, Vol. VII. No. 2. pp. 170-195.
Curley, W. E., & Schmitt, A. P. (1993). Revising SAT-Verbal Items to Eliminate Differential Item Functioning College Board Report. No. 93-2, ETS RR No. 93-61. New York: The College Board
Else-Quest, N. M., Hyde, J. S., & Linn, M. C. (2010). Cross-National Patterns of Gender Differences in Mathematics: A Meta-Analysis. Psychological Bulletin, 136(1), 103–127.
Gallagher, A. M., Levin, J., & Cahalan, C. (2002). Cognitive Patterns of Gender Differences on Mathematics Admissions Tests. GRE Board Report No. 96- 17P. Princeton, NJ: Educational Testing Services.
Hambleton, Ronald K., H. Swaminathan, dan H. Jane Rogers. (1991). Fundamentals of Item Response Theory. London: SAGE Publication, Inc.
Halpern, Diana F. (1997). Sex Differences in Intelligence Implication for Education. American Psychologist, Vol. 52.
Harris, A. M., & Carlton, S. T. (1993). Patterns of gender differences on mathematics items on the Scholastic Aptitude Test. Applied measurement in Education, 6(2), 137-151.
Kartowagiran, B. (2005). Perbandingan berbagai metode untuk mendeteksi bias butir (Doctoral dissertation, Universitas Gadjah Mada).
Keeves, J. (1992). The IEA Study of Science III: Changes in science edu cation and achievement: 1970 to 1984. New York: Pergamon.
Nandakumar, R. (1994). Assessing Dimensionality of a Set of Item Responses:Comparison of Different Approaches. Journal of Educational Measurement, 31(1), 17-35.
Nankervis, B. (2011). Gender Inequities in University Admission due to the Differential Validity of the SAT. Journal of College Admission, 213, 24-30.
Park, G.-P. (2008). Differential Item Functioning on an English Listening Test across Gender. TESOL Quarterly, 42(1), 115-123. doi: 10.2307/40264430
Penfield, R. D. & Lam, T. C. M. (2000). Assessing differential item functioning in performance assessment: Review and recommendations. Educational Measurement: Issues and Practice, 19, 5-15.
Retnawati, H. (2003). Keberfungsian Butir Diferensial pada Perangkat Tes Seleksi Masuk SLTP Mata Pelajaran Matematika. Jurnal Penelitian dan Evaluasi Pendidikan, 5(6).
Retnawati, H. (2013). Pendeteksian keberfungsian butir pembeda dengan indeks volume sederhana berdasarkan teori respons butir multidimensi. Jurnal penelitian dan evaluasi pendidikan, 17(2), 275-286.
Scheuneman, J. D., & Gerritz, K. (1990). Using differential item functioning procedures to explore sources of item difficulty and group performance characteristics. Journal of Educational Measurement, 27(2), 109-131.
Stark, S., Chernyshenko, O. S., & Drasgow, F. (2004). Examining the effects of differential item (functioning and differential) test functioning on selection decisions: When are statistically significant effects practically important?. Journal of Applied Psychology, 89(3), 497.