To measure the effectiveness of the ITS, the tutor was deployed on TutorShop, and the pre- and the post-tests were embedded in the problem set. The problems were different in the two tests, but they were carefully selected in the way that the difficulty level was similar and the solutions were analogous.
The problem set was assigned to 78 students as homework. Students were given deidentified accounts and a detailed instruction file on how to log in and find the assignment. After 5 days, 12 students didn’t finish it due to technical issues. An optional follow-up online survey was sent to the students, and 15 students gave feedback about the experience of interacting with the ITS.
This study has been reviewed andgranted approval by the Carnegie Mellon University Institutional Review Board (IRB) under Exempt Review on 5/26/2022, in accordance with 45 CFR 46.104.d.1. The IRB ID of thiss tudy is STUDY2021_00000509.
The transaction data also showed that with the prompt in the instruction, the students were able to use the unknown setting function. The average score of the pre-test is 2.75 out of 4, indicating the students had relatively high prior knowledge. The average score of the post-test is 2.25 out of 4, which is lower than the average pre-test score. The p-value of the t-test is 0.0166, indicating the post-test results are significantly worse than the pre-test.
With detailed examination, it is shown by the grouped average scores that the tutor had a more positive effect on low-performing students. However, students reported positive attitudes towards the experience. There are several possible reasons for the negative learning gain as listed below.
First, the way of the questions asked in the test problems is not as straightforward as it is in the practice problems, and there are more detailed descriptions in the pre- and post-test than in the practice problems. The practice part of the ITS didn’t model the errors related to reading ability.
Second, students with higher prior knowledge (who got higher scores in the pre-test) may tend to be less careful in the post-test. As suggested by previous works, affect plays a crucial role in learning processes, but this aspect is not considered in this study.
Due to the limitation of the remote setting and timing, the students' learning behavior wasn't closely observed. Future work needs to be done on fixing the technical problems, understanding student difficulties, and redesigning the tutor to enhance learning effectiveness.
Aleven, V. (2010). Rule-based cognitive modeling for intelligent tutoring systems. In Advances in intelligent tutoring systems (pp. 33-62). Springer, Berlin, Heidelberg.
Aleven, V., Mclaren, B. M., Sewall, J., & Koedinger, K. R. (2009). A new paradigm for intelligent tutoring systems: Example-tracing tutors. International Journal of Artificial Intelligence in Education, 19(2), 105-154.
Aleven, V., McLaughlin, E. A., Glenn, R. A., & Koedinger, K. R. (2016). Instruction based on adaptive learning technologies. Handbook of research on learning and instruction, 522-560.
Aleven, V., & Sewall, J. (2010, June). Hands-on introduction to creating intelligent tutoring systems without programming using the cognitive tutor authoring tools (CTAT). In Proceedings of the 9th International Conference of the Learning Sciences-Volume 2 (pp. 511-512).
Cao, J., Yang, T., Lai, I. K. W., & Wu, J. (2021). Student acceptance of intelligent tutoring systems during COVID-19: The effect of political influence. The International Journal of Electrical Engineering & Education, 00207209211003270.
Clark, R. E., Feldon, D. F., van Merriënboer, J. J., Yates, K. A., & Early, S. (2008). Cognitive task analysis. In Handbook of research on educational communications and technology (pp. 577-593).
Routledge.Clark, R. C., & Mayer, R. E. (2016). E-learning and the science of instruction: Proven guidelines for consumers and designers of multimedia learning. john Wiley & sons.Ceiling effect (statistics). (n.d.). Wikipedia. Retrieved July 14, 2022, from https://en.wikipedia.org/wiki/Ceiling_effect_(statistics)
D’Mello, S. K., & Graesser, A. C. (2014). 31 Feeling, Thinking, and Computing with Affect-Aware Learning Technologies. The Oxford handbook of affective computing, 419.
Han, J., Zhao, W., Jiang, Q., Oubibi, M., & Hu, X. (2019, October). Intelligent tutoring system trends 2006-2018: A literature review. In 2019 Eighth International Conference on Educational Innovation through Technology (EITT) (pp. 153-159). IEEE.
Koedinger, K., Aleven, V., Roll, I., & Baker, R. (2009). In vivo experiments on whether supporting metacognition in intelligent tutoring systems yields robust learning. In Handbook of metacognition in education (pp. 395-424). Routledge.
Koedinger, K. R., & Anderson, J. R. (1993). Effective use of intelligent software in high school math classrooms.
Koedinger, K. R., Corbett, A. T., & Perfetti, C. (2012). The Knowledge‐Learning‐Instruction framework: Bridging the science‐practice chasm to enhance robust student learning. Cognitive science, 36(5), 757-798.
Koedinger, K., Cunningham, K., Skogsholm, A., & Leber, B. (2008, June). An open repository and analysis tools for fine-grained, longitudinal learner data. In Educational data mining 2008.
Malekzadeh, M., Mustafa, M. B., & Lahsasna, A. (2015). A review of emotion regulation in intelligent tutoring systems. Journal of Educational Technology & Society, 18(4), 435-445.
Nesbit, J. C., Adesope, O. O., Liu, Q., & Ma, W. (2014, July). How effective are intelligent tutoring systems in computer science education?. In 2014 IEEE 14th international conference on advanced learning technologies (pp. 99-103). IEEE.
Zhang, B., & Jia, J. (2017, June). Evaluating an intelligent tutoring system for personalized math teaching. In 2017 international symposium on educational technology (ISET) (pp. 126-130). IEEE.
杨晓晓. (2021). 让高考不再 “高” 不可攀# br#——突破 “化学反应原理” 试题的解题瓶颈# br. 化学教与学, 558(6x), 65.