Abstract
The assessment of students' understanding of the difficulty levels in physics, provided by Rasch analysis, is carried out here using two different tests as measurement tools. We used a poly-tomous version of the standard FCI test by selecting items that maintain the core of this conceptual inventory. To do this, we randomly divided the 380 responses collected during a recent FCI analysis into five groups and then restructured six items into six separate evaluation scale items. At this stage, we carefully selected items to high-light the dominance of one common-sense error in mechanics. Next, we created our own test with items designed to cover six levels of difficulty, alternating simpler calculation steps within phys-ics-based problems. For both cases, the threshold parameter was estimated using the polytomous Rasch model. The results were analysed and discussed.
Keywords: Concept Inventory, Physics, Knowledge, Index theory, The Rasch Model, Socio-metric Measuring and Instruments.
Background
The assessment of conceptual knowledge in physics often relies on the Force Concept Inventory (FCI), a tool designed to measure the depth of understanding in mechanics beyond simple procedural calculations. While the standard FCI uses binary (correct/incorrect) scoring, there is a growing interest in using polytomous (multi-level) models to better capture students' reasoning and common-sense errors. The Rasch model, specifically its polytomous application, provides a threshold parameter that measures the difficulty of transitioning between different levels of knowledge. This study aims to evaluate undergraduate students' understanding of physics by applying Rasch analysis to restructured conceptual tests, identifying how difficulty is distributed across various levels of proficiency.
Methods
The study utilized two measurement tools: a restructured, polytomous version of the standard FCI and an ad-hoc test designed to cover six levels of difficulty. For the FCI version, the researchers analyzed 380 existing student responses, grouping them into a 6-point evaluation scale based on a taxonomy of common-sense errors in mechanics. Initially, an ad-hoc test was validated by 11 physics educators, but it yielded a low Content Validity Ratio (CVR) of 0.4 due to a lack of consensus on scoring. Consequently, the authors developed "cloned items" that maintained the philosophy of the FCI. Data analysis was conducted using the polytomous Rasch model to estimate threshold parameters (τ), which represent the difficulty of obtaining a score h relative to h−1.
Results
The analysis revealed that students' measured ability typically falls into 5–6 distinct levels. A key finding was the non-uniform distribution of difficulty: the first and last threshold parameters showed significantly higher magnitudes than the intermediate levels. For example, in a sample of 280 students, the first threshold difference was as high as 3.7471, while intermediate steps dropped to 0.4317. This indicates that the "pedagogical effort" required to bring a student from a basic level to an admissible one, or from a good level to an excellent one, is much greater than the effort needed to progress through the middle range of knowledge. The threshold levels were also found to be nearly symmetrical around the zero point, reflecting the consistency of the standardized FCI-cloned test.
Conclusions
The authors conclude that combining dichotomous and polytomous scoring algorithms through Rasch analysis provides a more nuanced understanding of conceptual knowledge structures. The study highlights that difficulty thresholds are not uniformly distributed, suggesting that pedagogical investment is most "cost-efficient" when focused on students at intermediate proficiency levels. Conversely, elevating students with very low conceptual knowledge or pushing high-achievers toward total mastery requires disproportionately more resources. The study recommends that educators and researchers adopt Item Response Theory (IRT)-based techniques to inform evidence-based instructional practices and design better conceptual inventories that move beyond simple binary assessments.
CONFLICT OF INTEREST
The authors declare no conflict of interest.
REFERENCES
Citing Literature
How to cite this article:
Kushta, E., Kadena, F., Prenga,D. DOI: 10.63871…. UniVlora Scientific Journal 2025, no.I, volume II
