| 研究生: |
鄭宇辰 YU-CHEN CHENG |
|---|---|
| 論文名稱: |
融合思考鏈機制的主動型生成式智慧教學助手之設計與其對科學學習成效之分析 Design and Analysis of a Chain-of-Thought- Enhanced Active Generative Intelligent Tutors for Imporving Science Learning Performance |
| 指導教授: | 劉晨鐘 |
| 口試委員: | |
| 學位類別: |
碩士 Master |
| 系所名稱: |
資訊電機學院 - 資訊工程學系 Department of Computer Science & Information Engineering |
| 論文出版年: | 2025 |
| 畢業學年度: | 113 |
| 語文別: | 中文 |
| 論文頁數: | 155 |
| 中文關鍵詞: | 智慧教學助手 、科學模擬 、大型語言模型 、思考鏈 |
| 外文關鍵詞: | Intelligent Tutoring System, Science Simulation, Large Language Model, Chain-of-Thought |
| 相關次數: | 點閱:17 下載:0 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
智慧教學系統是一種模擬人類教師功能的數位學習支援工具,能根據學生的學習行為提供即時回饋與個別化引導,並減少教師於課堂中的負擔。隨著大型語言模型技術發展,生成式智慧教學助手開始展現出強大的語意理解與對話能力,其中,思考鏈提示詞架構進一步引導模型先進行推理後再生成回答,使其可以產生更具邏輯性與可解釋性的回應。本研究在CoSci線上科學模擬平台上開發了一套基於大型語言模型GPT-4o的智慧教學助手,並設計了兩種引導模式,其一為僅被動根據學生提問回應的被動型助手,其二為引入思考鏈機制,會根據學生的學習狀況主動引導學生的主動型助手。
本研究招募台灣北部一所高中的22位一年級學生,所有學生依序使用兩種智慧教學助手並搭配不同主題的科學模擬進行學習。在每次活動前後皆會進行對應的物理概念測驗,並記錄學生在活動中的操作紀錄及與智慧教學助手的互動記錄,以進一步分析學生在不同引導模式下的學習成效及行為差異。
研究結果顯示,學生在兩組活動中的概念測驗後測成績皆有顯著提升,且兩組活動中的進步幅度並無顯著差異,顯示兩種引導方式均有助於學生學習。但在學習歷程與互動方式上呈現明顯差異,在被動型情境中,整體互動過程由學生主導,互動模式呈現一問一答的形式,而在主動型情境中,整體互動則由智慧助手主導,智慧助手會依據學生的情況主動進行引導、提問及給予回饋等行為。相較於被動型助手,主動型助手更能促進學生的知識轉化,並將學生的注意力從模擬轉移至與智慧助手的互動中。進一步分析發現,主動型助手的思考鏈推理品質與學生學習成效為顯著正相關,顯示若智慧助手能正確推斷學生情境並選擇適當策略,將有助於學習成效提升。此外,本研究亦從學生與智慧助手的案例中進行質性分析,以了解智慧助手具體如何引導學生及思考鏈推理過程如何影響其推理方式,並在研究的最後提出對未來改進的建議。
Intelligent Tutoring Systems are digital learning support tools designed to simulate the role of human teachers. They provide students with real-time feedback and personalized guidance based on their behaviors, helping to reduce the workload of teachers in class. With recent advances of large language models, generative AI-based tutoring assistants have shown impressive capabilities in language understanding and engaging in meaningful dialogue. The Chain-of-Thought prompting framework further enhances these systems by guiding the model to reason before generating responses, resulting in replies that are more logical and explainable.
In this study, we developed an intelligent tutoring assistant based on GPT-4o large language model on the CoSci online science simulation platform. Two types of guidance modes were designed: a passive assistant that only responds to students’ questions, and an active assistant enhanced with Chain-of-Thought framework that actively guides students based on their learning status.
A total of 22 first-year high school students from northern Taiwan participated in the study. Each student sequentially used both types of assistants, paired with different science simulations. Before and after each activity, students completed corresponding physics concept tests. Their simulation logs and interactions with assistants were recorded for further analysis of learning outcomes and behavioral differences under different guidance modes.
The results showed that students’ post-test scores significantly improved in both activities, with no significant difference in the magnitude of improvement between two modes, indicating that both guidance strategies effectively supported learning. However, their learning processes and interaction patterns differed notably. In passive mode, students led the interaction, typically in a question-answer format, while in the active mode, the interaction was led by the assistant, which actively asking questions, giving feedback, and guiding learning. The active assistant better supported knowledge transformation and shifted students' focus from simulation to conversations with assistant. Further analysis revealed a significant positive correlation between the quality of the assistant’s Chain-of-Thought reasoning and students’ learning performance. This suggests that when the system can accurately assess a student’s situation and apply suitable strategies, it can significantly enhance learning outcomes.
Additionally, qualitative case studies were conducted to reveal how the assistant guided students and how Chain-of-Thought reasoning influenced the assistant’s responses. Based on these findings, this study concludes with suggestions for future improvements.
Abdelghani, R., Murayama, K., Kidd, C., Sauzéon, H., & Oudeyer, P. Y. (2025). Investigating Middle School Students Question-Asking and Answer-Evaluation Skills When Using ChatGPT for Science Investigation. arXiv preprint arXiv:2505.01106.
Alrakhawi, H. A., Jamiat, N., & Abu-Naser, S. S. (2023). Intelligent tutoring systems in education: a systematic review of usage, tools, effects and evaluation. Journal of Theoretical and Applied Information Technology, 101(4), 1205-1226.
Batsaikhan, B. Z., & Correia, A. P. (2024). The effects of Generative Artificial Intelligence on Intelligent Tutoring Systems in higher education: A systematic review. Studies in Technology Enhanced Learning, 4(1).
Bubeck, S., Chandrasekaran, V., Eldan, R., Gehrke, J., Horvitz, E., Kamar, E., Lee, P., Lee, Y.T., Li, Y., Lundberg, S., Nori, H., Palangi, H., Ribeiro, M.T., Zhang, Y. (2023). Sparks of Artificial General Intelligence: Early experiments with GPT-4. arXiv preprint arXiv:2303.12712.
Chang, C. J., Liu, C. C., Wu, Y. T., Chang, M. H., Chiang, S. F., Chiu, B. C., Wen, C. T., Hwang, F. K., Chao, P. Y., Lai, C. H., Wu, S. W., & Chang, C. K. (2016). Students' perceptions on problem solving with collaborative computer simulation. In 24th International Conference on Computers in Education, ICCE 2016 (pp. 166-168). Asia-Pacific Society for Computers in Education.
Chia, Y.K., Chen, G., Tuan, L.A., Poria, S., Bing, L. (2023). Contrastive Chain-of-Thought Prompting. arXiv preprint arXiv:2311.09277.
Chiu, J. L., & Linn, M. C. (2011). Knowledge integration and wise engineering. Journal of Pre-College Engineering Education Research (J-PEER), 1(1), 2.
Collins, A., Brown, J. S., & Holum, A. (1991). Cognitive apprenticeship: Making thinking visible. American educator, 15(3), 6-11
Constantinou, C. P., Tsivitanidou, O. E., & Rybska, E. (2018). What is inquiry-based science teaching and learning? In O. Tsivitanidou, P. Gray, E. Rybska, L. Louca, & C. Constantinou (Eds.), Professional development for inquiry-based science teaching and learning (Vol. 5, pp. 1–23). Springer , Cham.
Cooper, G. (2023). Examining science education in ChatGPT: An exploratory study of generative artificial intelligence. Journal of science education and technology, 32(3), 444-452.
Cress, U., & Kimmerle, J. (2023). Co-constructing knowledge with generative AI tools: Reflections from a CSCL perspective. International Journal of Computer-Supported Collaborative Learning, 18(4), 607-614.
de Jong, T., & van Joolingen, W. R. (1998). Scientific Discovery Learning with Computer Simulations of Conceptual Domains. Review of Educational Research, 68(2), 179–201. https://doi.org/10.2307/1170753
Donnelly, D. F., Linn, M. C., & Ludvigsen, S. (2014). Impacts and characteristics of computer-based science inquiry learning environments for precollege students. Review of Educational Research, 84(4), 572-608.
Dunbar, K. (1993). Concept discovery in a scientific domain. Cognitive science, 17(3), 397-434.
Feng, G., Zhang, B., Gu, Y., Ye, H., He, D., Wang, L. (2023). Towards Revealing the Mystery behind Chain of Thought: A Theoretical Perspective. arXiv preprint arXiv:2305.15408.
Furtak, E. M., Seidel, T., Iverson, H., & Briggs, D. C. (2012). Experimental and quasi-experimental studies of inquiry-based science teaching: A meta-analysis. Review of educational research, 82(3), 300-329.
Graesser, A. C., Chipman, P., Haynes, B. C., & Olney, A. (2005). AutoTutor: An intelligent tutoring system with mixed-initiative dialogue. IEEE Transactions on Education, 48(4), 612-618.
Hsieh, C., Chen, S., Li, C., Fujii, Y., Ratner, A., Lee, C., Krishna, R., Pfister, T. (2023). Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models. arXiv preprint arXiv:2308.00675.
Kirschner, P. A., Sweller, J., & Clark, R. E. (2006). Why minimal guidance during instruction does not work: An analysis of the failure of constructivist, discovery, problem-based, experiential, and inquiry-based teaching. Educational psychologist, 41(2), 75-86.
Kojima, T., Gu, S.S., Reid, M., Matsuo, Y., Iwasawa, Y. (2022). Large Language Models are Zero-Shot Reasoners. arXiv preprint arXiv:2205.11916.
Kong, A., Zhao, S., Chen, H., Li, Q., Qin, Y., Sun, R., Zhou, X., Wang, E., Dong, X. (2023). Better Zero-Shot Reasoning with Role-Play Prompting. arXiv preprint arXiv:2308.07702.
Kulik, J. A., & Fletcher, J. D. (2016). Effectiveness of intelligent tutoring systems: a meta-analytic review. Review of educational research, 86(1), 42-78.
Lehtinen, A., & Viiri, J. (2017). Guidance provided by teacher and simulation for inquiry-based learning: A case study. Journal of science education and technology, 26(2), 193-206.
Liu, J., Liu, A., Lu, X., Welleck, S., West, P., Bras, R.L., Choi, Y., Hajishirzi, H. (2021). Generated Knowledge Prompting for Commonsense Reasoning. arXiv preprint arXiv:2110.08387.
Liu, S., Guo, X., Hu, X., & Zhao, X. (2024). Advancing Generative Intelligent Tutoring Systems with GPT-4: Design, Evaluation, and a Modular Framework for Future Learning Platforms. Electronics, 13(24), 4876.
Lu, P., Mishra, S., Xia, T., Qiu, L., Chang, K., Zhu, S., Tafjord, O., Clark, P., Kalyan, A. (2022). Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering. arXiv preprint arXiv:2209.09513.
Marshall, J. C., & Smart, J. B. (2013). Teachers’ transformation to inquiry-based instructional practice. Creative Education, 4(02), 132.
Meulenbroeks, R., van Rijn, R., & Reijerkerk, M. (2024). Fostering secondary school science students’ intrinsic motivation by inquiry-based learning. Research in Science Education, 54(3), 339-358.
Min, S., Lyu, X., Holtzman, A., Artetxe, M., Lewis, M., Hajishirzi, H., Zettlemoyer, L. (2022). Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?. arXiv preprint arXiv:2202.12837.
National Research Council. (2000). Inquiry and the National Science Education Standards: A Guide for Teaching and Learning. National Academies Press.
Ng, D. T. K., Tan, C. W., & Leung, J. K. L. (2024). Empowering student self‐regulated learning and science education through ChatGPT: A pioneering pilot study. British Journal of Educational Technology, 55(4), 1328-1353.
Perkins, K., Adams, W., Dubson, M., Finkelstein, N., Reid, S., Wieman, C., & LeMaster, R. (2006). PhET: Interactive simulations for teaching and learning physics. The physics teacher, 44(1), 18-23.
Podolefsky, N. S., Moore, E. B., & Perkins, K. K. (2013). Implicit scaffolding in interactive simulations: Design strategies to support multiple educational goals. arXiv preprint arXiv:1306.6544.
Prathigadapa, S., & Daud, S. M. (2025). A Review of Virtual Tutoring Systems and Student Performance Analysis Using GPT-3. Journal of Learning for Development, 12(1), 167-180.
Rutten, N., Van Joolingen, W. R., & Van Der Veen, J. T. (2012). The learning effects of computer simulations in science education. Computers & education, 58(1), 136-153.
Smetana, L. K., & Bell, R. L. (2012). Computer simulations to support science instruction and learning: A critical review of the literature. International Journal of Science Education, 34(9), 1337-1370.
Stamper, J., Xiao, R., & Hou, X. (2024, July). Enhancing llm-based feedback: Insights from intelligent tutoring systems and the learning sciences. In International Conference on Artificial Intelligence in Education (pp. 32-43). Cham: Springer Nature Switzerland.
Sui, C. J., Chen, H. C., Cheng, P. H., & Chang, C. Y. (2023). The Go-Lab platform, an inquiry-learning space: Investigation into students’ technology acceptance, knowledge integration, and learning outcomes. Journal of Science Education and Technology, 32(1), 61-77.
Sun, Y., Yan, Z., & Wu, B. (2022). How differently designed guidance influences simulation‐based inquiry learning in science education: A systematic review. Journal of Computer Assisted Learning, 38(4), 960-976.
Sun, Z., Wang, X., Tay, Y., Yang, Y., Zhou, D. (2022). Recitation-Augmented Language Models. arXiv preprint arXiv:2210.01296.
van der Graaf, J., Segers, E., & de Jong, T. (2020). Fostering integration of informational texts and virtual labs during inquiry-based learning. Contemporary educational psychology, 62, 101890.
VanLehn, K. (2011). The relative effectiveness of human tutoring, intelligent tutoring systems, and other tutoring systems. Educational psychologist, 46(4), 197-221.
Virvou, M., & Tsihrintzis, G. A. (2023, July). Is ChatGPT beneficial to education? A holistic evaluation framework based on intelligent tutoring systems. In 2023 14th International Conference on Information, Intelligence, Systems & Applications (IISA) (pp. 1-8). IEEE.
Wang, H., Wang, R., Mi, F., Deng, Y., Wang, Z., Liang, B., Xu, R., Wong, K. (2023). Cue-CoT: Chain-of-thought Prompting for Responding to In-depth Dialogue Questions with LLMs. arXiv preprint arXiv:2305.11792.
Wei, J., Wang, X., Schuurmans, D., Bosma, M., Ichter, B., Xia, F., Chi, E., Le, Q., Zhou, D. (2022). Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. arXiv preprint arXiv:2201.11903.
Yan, L., Sha, L., Zhao, L., Li, Y., Martinez‐Maldonado, R., Chen, G., ... & Gašević, D. (2024). Practical and ethical challenges of large language models in education: A systematic scoping review. British Journal of Educational Technology, 55(1), 90-112.
Zhang, Z., Zhang, A., Li, M., Zhao, H., Karypis, G., Smola, A. (2023). Multimodal Chain-of-Thought Reasoning in Language Models. arXiv preprint arXiv:2302.00923.
Zhou, D., Schärli, N., Hou, L., Wei, J., Scales, N., Wang, X., Schuurmans, D., Cui, C., Bousquet, O., Le, Q., Chi, E. (2022). Least-to-Most Prompting Enables Complex Reasoning in Large Language Models. arXiv preprint arXiv:2205.10625.
朱翊瑄 (2024),生成式人工智慧對於線上科學探究之影響,國立中央大學資訊 工程研究所,碩士論文。
許志仲 (2024),大型語言模型驅動之智慧學習助教對網路科學探究之成效評估,國立中央大學資訊工程研究所,碩士論文。
陳冠傑 (2025),具思維鏈推理與動機誘因設計之智慧助教環境建置與成效評估,國立中央大學資訊工程研究所,碩士論文。