Psychometric Evaluation on Mathematics Beliefs Instrument Using Rasch Model ()
1. Introduction
One of important aspects to be studied in mathematics education is belief. Mathematics beliefs plays role in shaping teachers’ practices and creates their attitude towards teaching ( Swan, 2002 ). Being successful teachers would derive beliefs system effectively through teaching process ( Wilkins, 2008 ). Although there are increasing numbers of researches on beliefs, the focus is mainly on its relationship with other variables. The beliefs system has a dynamic structure which can be related to achievement, teaching practices and knowledge ( Muijs & Reynolds, 2002 ).
However, mathematics beliefs is limited to empirical measurement only since the focus is more on its relationship with other variables ( Swan, 2002 ) such as the effect on instructional practices ( Beswick, 2005 , 2007). As a result, by testing the chosen the instrument of mathematics beliefs would lead to a reliable validated measurement of beliefs. The justification of testing the psychometric aspect of the mathematics beliefs is to provide empirical evidence in providing a credible measurement. A meaningful measurement of any variables produces standardized instrument that can be replicated to a different study ( Saidfudin et al., 2010 ). The importance of measuring mathematics beliefs in terms of psychometric aspects is to get the knowledge of the validity, reliability, scores, format and scale of the instrument ( Ali et al., 2014 ). Some selected recent studies like Head (2012) , Ali et al. (2014) , Çakiroğlu & Işıksal (2009) and Pampaka & Williams (2010) have tested mathematics beliefs instrument using Rasch model. The Rasch measurement model refers to a technique to measure any latent traits ( Abdul Aziz et al., 2013 ). It provides information on validity of instrument which reflects to the quality of the items in the questionnaire. Despite the differences in terms of samples and mathematics beliefs constructs, the findings are found to be common in terms of instrument validity and reliability. The research gap of this study is based on samples as well as the mathematics beliefs instrument. Therefore, by providing the psychometric evaluation on testing and validation of the said instrument would fulfil the research objective.
2. Methodology
Using a quantitative approach, 254 of secondary mathematics school teachers were selected randomly to be the sample of this study. There were 51 and 203 male and female teachers respectively. A set of Mathematics Belief Instrument (MBI) which adapted from Evans (2003) used 5-point Likert scale ranging from “1” as “strongly disagree” to “5” as’ strongly agree’. Initially, there were three constructs that include mathematics beliefs towards nature (11 items), beliefs about mathematics teaching (12 items) and beliefs about mathematics learning (13 items).
However, after exploratory factor analysis, some of the items have to be deleted due to low loading factor value. It was revealed that the reliability value of 0.81indicates that the internal consistency for MBI is acceptable (Hair et al., 2010). In addition, each construct of MBI has reliability value which ranging 0.71 to 0.81. The collected data were analysed using the Racsh model in order to convert the data in logit value.
Using Winsteps software, the data was analysed in order to determine the validity and reliability of the MBI. Table 1 shows the item reliability and construct validity of 254 measured persons. The person reliability which is equivalent to KR20 has acceptable value of 0.81 (Sekaran, 2003). This indicates that consistency level of person arrangement in answering the same constructs exists despite using different set of items (Wright & Masters, 1982). Based on the separation index value of 2.07, it shows that there are 3 levels of respondent ability in terms of their mathematics beliefs.
The item reliability analysis is shown in Table 2 which indicates a reliability value of 0.98. It is an acceptable value for which the instrument to be consistent if given to another sample that has the same and near features ( Arasinah, Bakar, Ramlah, Soaib, & Zaliza, 2015 ); while the separation index shows a value of 6.27 which indicates the scale of MBI can be divided into 6 levels of difficulty. The six levels are categorized according to the respondents’ ability in answering the items of the questionnaire.
Both values of the reliability person and item respectively have fulfilled the cut-off point as suggested by Bond & Fox (2007) . In addition, the separation index for both item and person exceeds 2.0 which are considered good
Table 1. Person reliability: mathematics beliefs.
Summary of 254 measured person.
Table 2. Item reliability: mathematics beliefs.
Summary of 36 measured item.
( Fisher Jr., 2007 ). Person and item separation indices are recommended to be good between 2 to 3 and good at 5 ( Fisher Jr., 2007 ). The more levels of difficulty the better the measurement reflecting different kinds of person’s ability.
3. Result and Discussion
Figure 1 shows the Person-Item Map Distribution (PIDM) which indicates the person ability on the left side and the item difficulty on the right side. The top left is allocated for person with high ability in terms of agreeable level. While the low ability person who has less agreeable level is placed at the lower part of the scale. Based on the item difficulty, the most difficult item is K6 which is followed by K11and K26; and the least difficult items are K24 and K34. The statement for item K6 is given by “In quadratic equation, the enhancement should be given to the skill of using formula in getting correct answer”. While the easiest items in the instrument include item K24 which states “Mathematics learning is an active process” and item K34 states “In learning mathematics, understanding mathematics concept, principal and strategy is important”. The overlapping items at different level of difficulty are considered identical in terms of measuring the same construct (Bond & Fox, 2001). Overlapping items include K10 and K22; K18, K20, K23 and K32, K13; K17, K19, K25, K30 and K7; K16, K21, K28 and K3; K12, K33, K8 and K9; K2, K35, K36, K4 and K5; K14, K15 and K27; and K24 and K34.
The item difficulty is not equally distributed among respondents of the study. Most items are considered to be at easy level which is not match with the respondents’ ability. The distribution of the person ability is towards high ability. Three respondents are found to be at the top scale of person ability and all of the respondents exceed the mean of the person ability.
Next, the polarity of items is shown by Figure 2 which is used to validate the construct validity of the instrument. All items have positive values of Point Measure Correlation (PTMEA CORR) that are more than 0.30 except item K6, K26, K1, K5 and K27. However, positive values of PTMEA designate that all items measure the required construct of mathematics beliefs. It is an indicator whether the items are in line with the measured construct.
Figure 2 is a screenshot of Rasch model output showing the Infit MNSQ and Outfit MNSQ values of the MBI. Item K6 has to be removed since the value Infit and Outfit value respectively is more than the suggested value of 1.44 ( Bond & Fox, 2007 ). The recommended range of these values is 0.60 to 1.40 if the instrument uses Likert scale type ( Bond & Fox, 2007 ). An item with low fit index value shows that they are overlapping with other items and high value indicates nonhomogeneous with the other. The most difficult item of K6 has the highest Infit/Outfit value which contradicts to the easiest items of K24 and K34 which has the lowest index mentioned.
4. Conclusion
The importance of this study is to provide psychometrical characteristics of mathematics beliefs instrument. The findings of this analysis provide credibility of the MBI using Rasch Model. Although there are different levels
Figure 1. The person-item map distribution conclusion.
Figure 2. Screen shots of Rasch model output showing item distribution according to difficulty level in mathematics beliefs.
of difficulty and ability, high consistency should be justified based on the reliability of person and item respectively. The findings have become the statistical evidence in validating the instrument which should be consistent with theoretical expectations ( Ariffin & Abdul Hamid, 2009 ). Misleading items can be either replaced or omitted from being included in the instrument. For future work, the study can be discussed further based on gender and years of teaching experience. The discussion of the mathematics beliefs instrument can be extended to misfit behaviours, differential item functioning, and other related item analysis using Rasch Model. The Rasch output has created a paradigm shift in measuring respondent’s perception using instrument. A meaningful data has been obtained through Rasch model output. Individual item checking can be done which leads to a quality instrument.