Research Article

Student Reaction to a Modified Force Concept Inventory: The Impact of Free-Response Questions and Feedback

Mark A. J. Parker 1 * , Holly Hedgeland 2 , Nicholas Braithwaite 1 , Sally Jordan 1
More Detail
1 The Open University, UK2 University of Cambridge, UK* Corresponding Author
European Journal of Science and Mathematics Education, 10(3), July 2022, 310-323, https://doi.org/10.30935/scimath/11882
Published: 08 March 2022
OPEN ACCESS   1672 Views   930 Downloads
Download Full Text (PDF)

ABSTRACT

The study investigated student reaction to the alternative mechanics survey (AMS), a modified force concept inventory, which used automatically marked free-response questions and offered limited feedback to students after their answers had been submitted. Eight participants were observed in completing the AMS, and they were interviewed to gain insight into what had been observed; the resultant data set was analyzed by thematic analysis. This revealed six key themes: “use of free-response questions supported deep learning”, “interpretation of the AMS instructions affected answer length”, “the idea of being marked by a computer did not affect answer structure”, “participant reaction to the usability of the AMS was mostly positive”, “reactions to the AMS depended upon what participants thought it was for”, and “limited feedback was a useful addition to the AMS”. Participants gave answers of differing length, being guided by the question wording as well as by their own preferences. It was found that participants valued being given feedback on their performance. Participants reacted positively to the free-response questions and could see potential for the use of this question type, opening up possibilities for the use of automatically marked free-response questions in concept inventories in the future.

CITATION (APA)

Parker, M. A. J., Hedgeland, H., Braithwaite, N., & Jordan, S. (2022). Student Reaction to a Modified Force Concept Inventory: The Impact of Free-Response Questions and Feedback. European Journal of Science and Mathematics Education, 10(3), 310-323. https://doi.org/10.30935/scimath/11882

REFERENCES

  1. Bailey, J. M., Johnson, B., Prather, E. E., & Slater, T. F. (2012). Development and validation of the star properties concept inventory. International Journal of Science Education, 34(14), 2257-2286. https://doi.org/10.1080/09500693.2011.589869
  2. Barnum, C. B. (2010). Usability testing essentials: Ready, set, test. Morgan Kaufmann Publishers.
  3. Boud, D., & Soler, R. (2016). Sustainable assessment revisited. Assessment & Evaluation in Higher Education, 41(3), 400-413. https://doi.org/10.1080/02602938.2015.1018133
  4. Braun, V., & Clarke, V. (2006). Using thematic analysis in psychology. Qualitative Research in Psychology, 3(2), 77-101. https://doi.org/10.1191/1478088706qp063oa
  5. Braun, V., Clarke, V., & Terry, G. (2014). Thematic analysis. Qualitative Research in Clinical Health Psychology, 24, 95-114. https://doi.org/10.1007/978-1-137-29105-9_7
  6. Brown, E., & Glover, C. (2006). Evaluating written feedback. In C. Bryan, & K. Clegg (Eds.), Innovative assessment in higher education (pp. 81-91). Routledge.
  7. Bull, J., & McKenna, C. (2004). Blueprint for computer-aided assessment. Routledge. https://doi.org/10.4324/9780203464687
  8. Carless, D. (2006). Differing perceptions in the feedback process. Studies in Higher Education, 31(2), 219-233. https://doi.org/10.1080/03075070600572132
  9. Carless, D. (2020). From teacher transmission of information to student feedback literacy: Activating the learner role in feedback processes. Active Learning in Higher Education. https://doi.org/10.1177/1469787420945845
  10. Castro, A., & Andrews, G. (2018). Nursing lives in the blogosphere: A thematic analysis of anonymous online nursing narratives. Journal of Advanced Nursing, 74(2), 329-338. https://doi.org/10.1111/jan.13411
  11. Chen, J. C., Kadlowec, J., & Whittinghill, D. (2004). Work in progress: Combining concept inventories with rapid feedback to enhance learning. In Proceedings of the 34th Annual Frontiers in Education. https://doi.org/10.1109/FIE.2004.1408580
  12. Clow, D. (2013). An overview of learning analytics. Teaching in Higher Education, 18(6), 683-695. https://doi.org/10.1080/13562517.2013.827653
  13. Crisp, G. (2007). The e-assessment handbook. Continuum.
  14. Dawson, P., Henderson, M., Mahoney, P., Phillips, M., Ryan, T., Boud, D., & Molloy, E. (2018). What makes for effective feedback: Staff and student perspectives. Assessment & Evaluation in Higher Education, 44(1), 25-36. https://doi.org/10.1080/02602938.2018.1467877
  15. Ding, L., Chaby, R., Sherwood, B., & Beichner, R. (2006). Evaluating an electricity and magnetism assessment tool: Brief electricity and magnetism assessment. Physical Review Special Topics-Physics Education Research, 2, 010105. https://doi.org/10.1103/PhysRevSTPER.2.010105
  16. Docktor, J. L., & Mestre, J. P. (2014). Synthesis of discipline-based education research in physics. Physical Review Special Topics-Physics Education Research, 10, 020119. https://doi.org/10.1103/PhysRevSTPER.10.020119
  17. Draper, S. (2009). What are learners actually regulating when given feedback. British Journal of Educational Technology, 40(2), 306-315. https://doi.org/10.1111/j.1467-8535.2008.00930.x
  18. Eaton, P. (2021). Evidence of measurement invariance across gender for the force concept inventory. Physical Review Physics Education Research, 17, 010130. https://doi.org/10.1103/PhysRevPhysEducRes.17.010130
  19. Filia, K. M., Jackson, H. J., Cotton, S. M., Gardner, A., Killackey, E. J., & Cook, J. A. (2018). What is social inclusion? A thematic analysis of professional opinion. Psychiatric Rehabilitation Journal, 41(3), 183-195. https://doi.org/10.1037/prj0000304
  20. Garcia, M., & Hamilton-Giachritsis, C. (2017). Getting involved: A thematic analysis of caregivers’ perspectives in Chilean residential children’s homes. Journal of Social and Personal Relationships, 34(3), 356-375. https://doi.org/10.1177/0265407516637838
  21. Garvin-Doxas, K., Klymkowsky, M., & Elrod, S. (2007). Building, using, and maximizing the impact of concept inventories in the biological sciences: Report on a National Science Foundation-sponsored conference on the construction of concept inventories in the biological sciences. CBE Life Sciences Education, 6(4), 277-282. https://doi.org/10.1187/cbe.07-05-0031
  22. Grogan, S., & Jayne, M. (2017). Body image after mastectomy: A thematic analysis of younger women’s written accounts. Journal of Health Psychology, 22(11), 1480-1490. https://doi.org/10.1177/1359105316630137
  23. Halloun, I., & Hestenes, D. (1985). The initial knowledge state of college students. American Journal of Physics, 53, 1043-1056. https://doi.org/10.1119/1.14030
  24. Henderson, M., Phillips, M., Ryan, T., Boud, D., Dawson, P., Molloy, E., & Mahoney, P. (2019). Conditions that enable effective feedback. Higher Education Research & Development, 38(7), 1401-1416. https://doi.org/10.1080/07294360.2019.1657807
  25. Hestenes, D., Wells, M., & Swackhamer, G., (1992). Force concept inventory. The Physics Teacher, 30, 141-158. https://doi.org/10.1119/1.2343497
  26. Hufnagel, B. (2002). Development of the astronomy diagnostic test. Astronomy Education Review, 1(1), 47-51. https://doi.org/10.3847/AER2001004
  27. Hunt, T. (2012). Computer-marked assessment in Moodle: Past, present and future. Digital Education Research Network. https://dern.acer.org/dern/ict-research/page/computer-marked-assessment-in-moodle-past-present-and-future
  28. Jensen, L. X., Bearman, M., & Boud, D. (2021). Understanding feedback in online learning–A critical review and metaphor analysis. Computers & Education, 173, 104271.https://doi.org/10.1016/j.compedu.2021.104271
  29. Kluger, A. N., & DeNisi, A. (1996). The effects of feedback interventions on performance: A historical view, a meta-analysis and a preliminary feedback intervention theory. Psychological Bulletin, 119(2), 254-284. https://doi.org/10.1037/0033-2909.119.2.254
  30. Lasry, N., Rosenfield, S., Dedic, H., Dahan, A., & Reshef, O. (2011). The puzzling reliability of the force concept inventory. American Journal of Physics, 79(9), 909-914. https://doi.org/10.1119/1.3602073
  31. Lawrie, G., Wright, A., Schultz, M., Dargaville, T., O’Brien, G., Bedford, S., Williams, M., Tasker, R., Dickson, H., & Thompson, C. (2013). Using formative feedback to identify and support first-year chemistry students with missing or misconceptions. A practice report. International Journal of the First Year in Higher Education, 4(2), 111-116. https://doi.org/10.5204/intjfyhe.v4i2.179
  32. Lee, N. W., Shamsuddin, W. N. F. W, Wei, L. C., Anuardi, M. N. A. M., Heng, C. S., & Abdullah, A. N. (2021). Using online multiple choice questions with multiple attempts: A case for self-directed learning among tertiary students. International Journal of Evaluation and Research in Education, 10(2), 553-568. https://doi.org/10.11591/ijere.v10i2.21008
  33. Millar, J. (2005). Engaging students with assessment feedback: What works? An FDTL5 project literature review. Oxford Brookes University Press.
  34. Mitchell, T., Aldridge, N., Williamson, W., & Broomhead, P. (2003). Computer based testing of medical knowledge. In Proceedings of the 7th International CAA Conference (pp. 249-267).
  35. Nardi, A., & Ranieri, M. (2019). Comparing paper-based and electronic multiple-choice examinations with personal devices: Impact on students’ performance, self-efficacy and satisfaction. British Journal of Educational Technology, 50(3), 1495-1506. https://doi.org/10.1111/bjet.12644
  36. Nickerson, J., Corter, J., Esche, S., & Chassapis, C. (2007). A model for evaluation the effectiveness of remote engineering laboratories and simulations in education. Computers and Education, 49(3), 708-725. https://doi.org/10.1016/j.compedu.2005.11.019
  37. Nicol, D. (2007). E‐assessment by design: Using multiple‐choice tests to good effect. Journal of Further and higher Education, 31(1), 53-64. https://doi.org/10.1080/03098770601167922
  38. Nicol, D. (2021). The power of internal feedback: Exploiting natural comparison processes. Assessment & Evaluation in Higher Education, 46(5), 756-778. https://doi.org/10.1080/02602938.2020.1823314
  39. Nyquist, J. B. (2003). The benefits of reconstructing feedback as a larger system of formative assessment: A meta-analysis. Vanderbilt University Press.
  40. Porter, L., Taylor, C., & Webb, K. (2014). Leveraging open source principles for flexible concept inventory development. In Proceedings of the 2014 Conference on Innovation & Technology in Computer Science Education (pp. 243-248). https://doi.org/10.1145/2591708.2591722
  41. Rebello, N., & Zollman, D. (2004). The effect of distractors on student performance on the force concept inventory. American Journal of Physics, 72, 116. https://doi.org/10.1119/1.1629091
  42. Reeves, B., & Nass, C. (1996). The media equation. Stanford University Press.
  43. Robertson, A. E., Stanfield, A. C., Watt, J., Barry, F., Day, M., Cormack, M., & Melville, C. (2018). The experience and impact of anxiety in autistic adults: A thematic analysis. Research in Autism Spectrum Disorders, 46, 8-18. https://doi.org/10.1016/j.rasd.2017.11.006
  44. Sangwin, C. J. (2013). Computer aided assessment of mathematics. Oxford University Press. https://doi.org/10.1093/acprof:oso/9780199660353.001.0001
  45. Scanlon, E., Colwell, C., Cooper, M., & Di Paolo, T. (2004). Remote experiments, re-versioning and re-thinking science learning. Computers and Education, 43(1-2), 153-163. https://doi.org/10.1016/j.compedu.2003.12.010
  46. Scott, S. V. (2014). Practising what we preach: Towards a student-centred definition of feedback. Teaching in Higher Education, 19(1), 49-57. https://doi.org/10.1080/13562517.2013.827639
  47. Sedrakyan, G., Malmberg, J., Verbert, K., Jarvela, S., & Kirschner, P. A. (2020). Linking learning behavior analytics and learning science concepts: Designing a learning analytics dashboard for feedback to support learning regulation. Computers in Human Behavior, 107, 105512. https://doi.org/10.1016/j.chb.2018.05.004
  48. Shute, V. J. (2008). Focus on formative feedback. Review of Educational Research, 78(1), 153-189. https://doi.org/10.3102/0034654307313795
  49. Simon, & Snowdon, S. (2014). Multiple-choice vs free-text code-explaining examination questions. In Proceedings of the 14th Koli Calling International Conference on Computing Education Research (pp. 91-97). https://doi.org/10.1145/2674683.2674701
  50. Smedley, R. M., & Coulson, N. S. (2017). A thematic analysis of messages posted by moderators within health-related asynchronous online support forums. Patient Education and Counseling, 100(9), 1688-1693. https://doi.org/10.1016/j.pec.2017.04.008
  51. Smith, J. I., & Tanner, K. (2010). The problem of revealing how students think: Concept inventories and beyond. CBE Life Sciences Education, 9(1), 1-5. https://doi.org/10.1187/cbe.09-12-0094
  52. Thornton, R., & Sokoloff, D. (1998). Assessing student learning of Newton’s laws: The force and motion conceptual evaluation and the evaluation of active learning laboratory and lecture curricula. American Journal of Physics, 66, 338. https://doi.org/10.1119/1.18863
  53. University of Auckland. (2017). Thematic analysis. https://www.psych.auckland.ac.nz/en/about/our-research/research-groups/thematic-analysis/about-thematic-analysis.html
  54. Varagona, L. M., & Hold, J. L. (2019). Nursing students’ perceptions of faculty trustworthiness: Thematic analysis of a longitudinal study. Nurse Education Today, 72, 27-31. https://doi.org/10.1016/j.nedt.2018.10.008
  55. Walker, D. J., Topping, K., & Rodrigues, S. (2008). Student reflections on formative e-assessment: Expectations and perceptions. Learning, Media and Technology, 33(3), 221-234. https://doi.org/10.1080/17439880802324178
  56. Weaver, M. R. (2006). Do students value feedback? Student perceptions of tutors’ written responses. Assessment and Evaluation in Higher Education, 31(3), 379-394. https://doi.org/10.1080/02602930500353061
  57. Woodford, K., & Bancroft, P. (2005). Multiple choice questions not considered harmful. In Proceedings of the Seventh Australasian Computing Education Conference (pp. 109-115).
  58. Yasuda, J., Mae, N., Hull, M. M., & Taniguchi, M. (2021). Optimizing the length of computerized adaptive testing for the force concept inventory. Physical Review Physics Education Research, 17, 010115. https://doi.org/10.1103/PhysRevPhysEducRes.17.010115
  59. Zeilik, M. (2003). Birth of the astronomy diagnostic test: Prototest evolution. Astronomy Education Review, 1(2), 46-52. https://doi.org/10.3847/AER2002005
  60. Zhang, L., & VanLehn, K. (2021). Evaluation of auto-generated distractors in multiple choice questions from a semantic network. Interactive Learning Environments, 29(6), 1019-1036. https://doi.org/10.1080/10494820.2019.1619586
  61. Zilvinskis, J., Willis, J. III., & Bordon, V. M. H. (2017). An overview of learning analytics. New Directions for Higher Education, 179, 9-17. https://doi.org/10.1002/he.20239