Keywords

NLP, machine learning, cognitive science, self-regulated learning, LLM, CoALA

Abstract

As large language models (LLMs) demonstrate increasingly sophisticated language understanding capabilities, they offer transformative potential for educational technology, particularly in developing more responsive intelligent tutoring systems (ITS). To develop these systems, a synthesis that balances theoretical frameworks with practical, cutting-edge, and deployable technology would be useful. This thesis aims to advance the field of ITSs theoretically by viewing the modern advancements in natural language processing within the context of the foundational frameworks important to the fields of self-regulated learning, ITSs, and cognitive agents. Then, building on prior research on the ITS MetaTutor \citep{Azevedo2022}, this research focuses on how functions of the pedagogical agent "Sam the Strategizer" could be deployed and accessible through local on-device models. % Specifically, I fine-tuned the encoder models ModernBERT-base and ModernBERT-large, and decoder-only models such as Llama 3.2, Phi-4, and Qwen 2.5, and evaluated their performance on a custom benchmark dataset for fact-checking claims and detecting student compliance. The following research questions guided this investigation: % (1) How can theoretical frameworks from cognitive architectures and intelligent tutoring systems be integrated with emergent abilities of LLMs to design modern pedagogical agents that scaffold self-regulated learning, and what cognitive architectural modules could be used to construct an SRL scaffolding agent (Sam) based on these principles? % (2) How do small, consumer-hardware-compatible encoder-only models compare to small decoder-only models in accuracy and efficiency when fact-checking claims, and how does domain-specific fine-tuning affect these results? % (3) How do these models perform on student compliance detection, and does fine-tuning a domain-specific expert improve these results? This research demonstrates the potential for integrating theoretical cognitive architecture frameworks with modern LLM abilities in the context of ITSs. % The experimental findings suggest that small fine-tuned decoder models outperform encoder models in fact checking domain-specific claims, while small encoder models are more effective at determining student compliance, which suggests that different model architectures may be optimal for different reasoning tasks that an agent performs. % Additionally, the results indicate that although larger models tend to work better for complex fine-grained tasks, small locally-deployable models can be effective with fine-tuning, potentially increasing privacy and accessibility for students. Since the parameter count appeared to significantly increase performance, future research could more thoroughly explore the trade-off between efficiency and accuracy by deploying larger models (7-8B parameters) that still fit on some consumer hardware. % Additionally, expanding beyond the domain of the circulatory system and exploring more the performance of models on different reasoning tasks could increase the agent's completeness.

Thesis Completion Year

2025

Thesis Completion Semester

Spring

Thesis Chair

Azevedo, Roger

College

College of Engineering and Computer Science

Department

Computer Science

Thesis Discipline

Computer Science

Language

English

Access Status

Open Access

Length of Campus Access

None

Campus Location

Orlando (Main) Campus

Included in

Engineering Commons

Share

COinS
 

Rights Statement

In Copyright