Abstract
Solving word problems involving 'Time' is an important skill but poor mastery of the skill among elementary students has often been reported in the literature. In addition, the available diagnostic tools in the literature might be less efficient for identifying the various errors made by many students in solving word problems. Thus, an online problem solving Cognitive Diagnostic Assessment (CDA) with Ordered Multiple-Choice (OMC) items was developed as a web application with automated scoring features to increase the efficiency in assessing Grade Five students’ mastery level of word problem-solving attributes involving the topic of Time. The online problem-solving CDA with OMC items was developed based on the principled assessment design which comprised four building blocks: (i) construct map, (ii) item design, (iii) outcome space, and (iv) measurement model. Newman's Error Analysis was adapted to specify the construct map required for the development of OMC items. In this article, we documented the empirical evidence regarding item quality, the validity of the cognitive models, the reliability of the instrument, and the diagnostic analysis of students' responses. This study involved 128 Grade Five students in three elementary schools. The results indicated that the instrument showed appropriate difficulty level and discrimination power, satisfactory model-data fit, and high reliability, but most of the students were not at the highest mastery level of the word problem-solving attributes. The findings of the study suggested that the web application could be a valid and reliable diagnostic tool for pinpointing students' errors made in solving word problems. Besides, the findings also highlight the need for intervention to enhance the students’ mastery of word problem-solving attributes.
Similar content being viewed by others
References
Alonzo, A. C., & Steedle, J. T. (2009). Developing and assessing a force and motion learning progression. Science Education, 93(3), 389–421.
Alves, C. B. (2012). Making diagnostic inferences about student performance on the Alberta education diagnostic mathematics project: An application of the Attribute Hierarchy Method. (Publication No. 919011661) [Doctoral Thesis, University of Alberta, Ann Arbor, Canada]. ProQuest Dissertations and Theses database.
Authors. (2019). International Journal of Science and Mathematics Education.
Authors. (2021). Journal of Computers in Education.
Bardhoshi, G., & Erford, B. T. (2017). Processes and procedures for estimating score reliability and precision. Measurement and Evaluation in Counseling and Development, 50(4), 256–263.
Boonen, A. J., de Koning, B. B., Jolles, J., & van der Schoot, M. (2016). Word problem solving in contemporary math education: A plea for reading comprehension skills training. Frontiers in Psychology, 7(191), 1–10.
Boora, S., Pasiphol, S., & Tangdhanakanond, K. (2015). Development of cognitive diagnostic testing on basic arithmetic operation. Procedia-Social and Behavioral Sciences, 191, 769–772.
Bradshaw, L. (2017). Diagnostic classification models. In A. A. Rupp & J. P. Leighton (Eds.), the handbook of cognition and assessment: Frameworks, methodologies, and applications (1st ed., pp. 297–327). Wiley Blackwell.
Briggs, D. C., & Alonzo, A. (2012). The psychometric modeling of ordered multiple-choice item responses for diagnostic assessment with a learning progression. In A. C. Alonzo & A. W. Gotwals (Eds.), Learning progressions in science: Current challenges and future directions (pp. 293–316). Sense Publishers.
Briggs, D. C., Alonzo, A., Schwab, C., & Wilson, M. (2006). Diagnostic assessment with ordered multiple-choice items. Educational Assessment, 11(1), 33–63.
Broaddus, A. E. (2011). An investigation into foundational concepts related to slope: An application of the Attribute Hierarchy Method (UMI No: 3487353) [Doctoral Thesis, University of Kansas, Kansas, United State]. ProQuest Dissertations and Theses database.
Cai, J., Morris, A., Hohensee, C., Hwang, S., Robison, V., & Hiebert, J. (2018). Using data to understand and improve students' learning: Empowering teachers and researchers through building and using a knowledge base. Journal for Research in Mathematics Education, 49(4), 362–372.
Clarkson, P. C. (1991). Language comprehension errors: A further investigation. Mathematics Education Research Journal, 3(2), 24–33.
Cui, Y., & Leighton, J. P. (2009). The Hierarchy Consistency Index: Evaluating person fit for cognitive diagnostic assessment. Journal of Educational Measurement, 46(4), 429–449.
De Champlain, A. F. (2010). A primer on classical test theory and item response theory for assessments in medical education. Medical Education, 44(1), 109–117.
Downing, S. M. (2004). Reliability: on the reproducibility of assessment data. Medical education, 38(9), 1006–1012.
Ebel, R. L. & Frisbie, D.A. (1991). Essentials of educational measurement (5th ed.). Prentice-Hall.
Fuchs, L. S., Fuchs, D., Seethaler, P. M., & Craddock, C. (2020a). Improving language comprehension to enhance word-problem solving. Reading & Writing Quarterly, 36(2), 142–156.
Fuchs, L., Fuchs, D., Seethaler, P. M., & Barnes, M. A. (2020b). Addressing the role of working memory in mathematical word-problem solving when designing intervention for struggling learners. ZDM, 52(1), 87–96.
Fulmer, G. W. (2015). Validating proposed learning progressions on force and motion using the force concept inventory: Findings from Singapore secondary schools. International Journal of Science and Mathematics Education, 13(6), 1235–1254.
Gay, L. R., Mills, G. E., & Airasian, P. W. (2012). Educational research: Competencies for analysis and applications (10th ed.). Merrill.
George, D., & Mallery, P. (2019). IBM SPSS statistics 26 step by step: A simple guide and reference. Routledge.
Gierl, M. J., & Zhou, J. (2008). Computer adaptive-attribute testing: A new approach to cognitive diagnostic assessment. Zeitschrift für Psychologie/Journal of Psychology, 216(1), 29–39.
Gierl, M. J., Alves, C., & Taylor-Majeau, R. (2010). Using the Attribute Hierarchy Method to make diagnostic inferences about examinees’ knowledge and skills in mathematics: An operational implementation of cognitive diagnostic assessment. International Journal of Testing, 10(4), 318–341.
Gierl, M. J., Cui, Y., & Zhou, J. (2009a). Reliability and attribute-based scoring in cognitive diagnostic assessment. Journal of Educational Measurement, 46(3), 293–313.
Gierl, M. J., Roberts, M. P. R., Alves, C., & Gotzmann, A. (2009b). Using judgments from content specialists to develop cognitive models for diagnostic assessments. Paper presented at the Annual Meeting of National Council on Measurement in Education, San Diego, CA.
Hadenfeldt, J. C., Bernholt, S., Liu, X., Neumann, K., & Parchmann, I. (2013). Using ordered multiple-choice items to assess students’ understanding of the structure and composition of matter. Journal of Chemical Education, 90(12), 1602–1608.
Hadenfeldt, J. C., Neumann, K., Bernholt, S., Liu, X., & Parchmann, I. (2016). Students’ progression in understanding the matter concept. Journal of Research in Science Teaching, 53(5), 683–708.
Haladyna, T. M. (1992). Context-dependent item sets. Educational Measurement: Issues and Practice, 11(1), 21–25.
Haladyna, T. M. (2004). Developing and validating multiple-choice test items (3rd ed.). Routledge.
Hassan, F. N. A., Mohamad, Q. A., & Azuddin, S. K. Y. (2020). The implementation of higher order thinking skills (HOTS) in Malaysia secondary school: Post PISA 2009. International Journal of Psychosocial Rehabilitation, 24(5), 5510–5517.
Ihudiebube-Splendor, C. N., & Chikeme, P. C. (2020). A descriptive cross-sectional study: Practical and feasible design in investigating health care–seeking behaviors of undergraduates.In SAGE Research Methods Cases. https://doi.org/10.4135/9781529742862
Keehner, M., Gorin, J. S., Feng, G., & Katz, I. R. (2017). Developing and validating cognitive models in assessment. In A. A. Rupp & J. P. Leighton (Eds.), The handbook of cognition and assessment: Frameworks, methodologies, and applications (1st ed., pp. 75-101). John Wiley & Sons.
LaDuca, A., Downing, S. M., & Henzel, T. R. (1995). Systematic item writing and test construction. In J. C. Impara (Ed.), Licensure testing: Purposes, procedures, and practices (pp. 117-149). Buros.
Lai, H., Gierl, M. J., & Babenko, O. (2015). Application of conditional means for diagnostic scoring. International Journal of Learning, Teaching and Educational Research, 12(3), 61–79.
Leighton, J. P., & Gierl, M. J. (2011). The learning sciences in educational assessment: The role of cognitive models. Cambridge University Press.
Leighton, J. P., Gierl, M. J., & Hunka, S. M. (2004). The Attribute Hierarchy Method for cognitive assessment: A variation on Tatsuoka's Rule-Space Approach. Journal of Educational Measurement, 41(3), 205–237.
Leong, Y. H., Tay, E. G., Toh, T. L., Quek, K. S., Toh, P. C., & Dindyal, J. (2016). Infusing mathematical problem solving in the mathematics curriculum: Replacement Units. In P. Felmer, E. Pehkonen, & J. Kilpatrick (Eds.), Posing and solving mathematical problems: Advances and new perspective (pp. 309–325). Springer.
Liljedahl, P., Santos-Trigo, M., Malaspina, U., & Bruder, R. (2016). Problem solving in mathematics education. Springer.
Lithner, J. (2008). A research framework for creative and imitative reasoning. Educational Studies in Mathematics, 67(3), 255–276.
Ministry of Education Malaysia [MOE] (2016). Kurikulum standard sekolah rendah: Standard kurikulum dan pentaksiran Matematik Tahun 4 [Primary school standard curriculum: Year 4 Mathematics curriculum and assessment standard document]. Ministry of Education Malaysia.
Newman, M. A. (1977). An analysis of sixth-grade pupil's error on written mathematical tasks. Victorian Institute for Educational Research Bulletin, 39, 31–43.
Nichols, P. D., Kobrin, J. L., Lai, E., & Koepfler, J. D. (2017). The role of theories of learning and cognition in assessment design and development. In A. A. Rupp & J. P. Leighton (Eds.), The handbook of cognition and assessment: Frameworks, methodologies, and applications (1st ed., pp. 41–74). Wiley Blackwell.
O’Connor, B. R., & Norton, S. (2020). Supporting indigenous primary students’ success in problem-solving: Learning from Newman interviews. Mathematics Education Research Journal, 1-24.
Polit, D. F., & Beck, C. T. (2006). The content validity index: Are you sure you know what's being reported? Critique and recommendations. Research in Nursing & Health, 29(5), 489–497.
Quaigrain, K., & Arhin, A. K. (2017). Using reliability and item analysis to evaluate a teacher-developed test in educational measurement and evaluation. Cogent Education, 4(1), 1–11.
Roberts, M. R., Alves, C. B., Chu, M. W., Thompson, M., Bahry, L. M., & Gotzmann, A. (2014). Testing expert based versus student based cognitive models for a Grade 3 diagnostic mathematics assessment. Applied Measurement in Education, 27(3), 173–195.
Rohmah, M., & Sutiarso, S. (2018). Analysis problem solving in mathematical using theory Newman. Eurasia Journal of Mathematics, Science and Technology Education, 14(2), 671–681.
Salkind, N. (2010) Convenience sampling. In N. Salkind (Ed.), Encyclopedia of research design (p. 254). Sage.
Scheibling-Seve, C., Pasquinelli, E., & Sander, E. (2020). Assessing conceptual knowledge through solving arithmetic word problems. Educational Studies in Mathematics, 1-19.
Schultz, M., Lawrie, G. A., Bailey, C. H., Bedford, S. B., Dargaville, T. R., O'Brien, G., .Tasker, R., Thompson, C. D., Williams, M. & Wright, A. H. (2017). Evaluation of diagnostic tools that tertiary teachers can apply to profile their students’ conceptions. International Journal of Science Education, 39(5), 565–586.
Scully, D. (2017). Constructing multiple-choice items to measure higher-order thinking. Practical Assessment, Research, and Evaluation, 22(1), 1–4.
Sia, C. J. L. (2017). Development and validation of Cognitive Diagnostic Assessment (CDA) for primary mathematics learning of time [Unpublished master's thesis]. Universiti Sains Malaysia, Malaysia.
Sia, C. J. L., & Lim, C. S. (2018). Cognitive diagnostic assessment: An alternative mode of assessment for learning. In D. R. Thompson, M. Burton, A. Cusi, & D. Wright (Eds.), Classroom assessment in mathematics (pp. 123–137). Springer.
Sibanda, L. (2017). Grade 4 learners’ linguistic difficulties in solving mathematical assessments. African Journal of Research in Mathematics, Science and Technology Education, 21(1), 86–96.
Tan, P. L., Lim, C. S., & Kor, L. K. (2017). Diagnosing primary pupils' learning of the concept of" after" in the topic "Time" through knowledge states by using cognitive diagnostic assessment. Malaysian Journal of Learning and Instruction, 14(2), 145–175.
Tavakol, M., & Dennick, R. (2011). Post-examination analysis of objective tests. Medical Teacher, 33(6), 447–458.
Thompson, N. A. (2010). KR-20. In N. Salkind (Ed.), Encyclopedia of research design (pp. 667–668). Sage.
Verschaffel, L., Van Dooren, W., & De Smedt, B. (2012). Mathematical Learning. In N. M. Seel (Eds.), Encyclopedia of the sciences of learning (pp. 2107–2110). Springer.
White, A. L. (2005). Active mathematics in classrooms: Finding out why children make mistakes-and then doing something to help them. Square One, 15(4), 15–19.
White, A. L. (2009). A revaluation of Newman's error analysis. In D. Martin, T. Fitzpatrick, R. Hunting, D. Itter, C. Lenard, T. Mills, L. Milne (Eds), Mathematics of prime importance, Proceedings of 46th Annual Conference of the Mathematics Association of Victoria (pp. 249-257). M.A.V.
White, A. L. (2018). Australian Mathematics Literacy Perspectives: Assessment. Southeast Asian Mathematics Education Journal, 8(1), 69–81.
Wijaya, A., van den Heuvel-Panhuizen, M., Doorman, M., & Robitzsch, A. (2014). Difficulties in solving context-based PISA mathematics tasks: An analysis of students' errors. The Mathematics Enthusiast, 11(3), 555–584.
Wilson, M. (2009). Measuring progressions: Assessment structures underlying a learning progression. Journal of Research in Science Teaching, 46(6), 716–730.
Acknowledgements
This study is made possible with funding from the Research University Grant (RUI) Scheme 1001/PGURU/8011027 of Universiti Sains Malaysia. The authors would like to thank the computer science experts, the web developer, and the technician from Centre for Knowledge, Communication and Technology, USM who involved in the Online CDA web application development process. The authors would also like to thank all the teachers and students who voluntarily participated in this study.
Availability of data and material
Not applicable.
Code availability
Not applicable.
Funding
This study is made possible with funding from the Research University Grant (RUI) Scheme 1001/PGURU/8011027 of Universiti Sains Malaysia, Penang, Malaysia.
Author information
Authors and Affiliations
Contributions
All authors contributed to the study conception and design. Material preparation, data collection and analysis were performed by Chin Huan and Chew Cheng Meng. The first draft of the manuscript was written by Chin Huan and Chew Cheng Meng commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Conflicts of interest/Competing interests
The authors have no conflicts of interest to declare that are relevant to the content of this article.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Word Problem Solving Construct Map
Level | Descriptor |
---|---|
0 | Does not master any attribute. |
1 | Master attribute: RU: Read and understand the word problem |
2 | Master attributes: RU: Read and understand the word problem, and T: Transform the word problem into correct mathematical sentences |
3 | Master attributes: RU: Read and understand the word problem, and T: Transform the word problem into correct mathematical sentences, and P: Perform mathematical operations necessary to solve the word problem after the transformation process |
4 | Master attributes: RU: Read and understand the word problem, and T: Transform the word problem into correct mathematical sentences, and P: Perform mathematical operations necessary to solve the word problem after the transformation process, and E: Encode the answer in an acceptable form |
Rights and permissions
About this article
Cite this article
Chin, H., Chew, C.M. Online Cognitive Diagnostic Assessment with Ordered Multiple-Choice Items for Word Problems involving ‘Time’. Educ Inf Technol 27, 7721–7748 (2022). https://doi.org/10.1007/s10639-022-10956-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10639-022-10956-2