Search Results for author: Conrad Borchers

Found 19 papers, 10 papers with code

LLM-Generated Feedback Supports Learning If Learners Choose to Use It

no code implementations20 Jun 2025 Danielle R. Thomas, Conrad Borchers, Shambhavi Bhushan, Erin Gatz, Shivang Gupta, Kenneth R. Koedinger

After adjusting for this effect, two out of seven lessons showed statistically significant learning benefits from LLM feedback with standardized effect sizes of 0. 28 and 0. 33.

Selection bias

Leveraging LLMs to Assess Tutor Moves in Real-Life Dialogues: A Feasibility Study

no code implementations20 Jun 2025 Danielle R. Thomas, Conrad Borchers, Jionghao Lin, Sanjit Kakarla, Shambhavi Bhushan, Erin Gatz, Shivang Gupta, Ralph Abboud, Kenneth R. Koedinger

Tutoring improves student achievement, but identifying and studying what tutoring actions are most associated with student learning at scale based on audio transcriptions is an open research problem.

Math

An Integrated Platform for Studying Learning with Intelligent Tutoring Systems: CTAT+TutorShop

no code implementations17 Jan 2025 Vincent Aleven, Conrad Borchers, Yun Huang, Tomohiro Nagashima, Bruce McLaren, Paulo Carvalho, Octav Popescu, Jonathan Sewall, Kenneth Koedinger

This platform has been used to develop and conduct an estimated 147 research studies which have run in a wide variety of laboratory and real-world educational settings, including K-12 and higher education, and have addressed a wide range of research questions.

Augmenting Human-Annotated Training Data with Large Language Model Generation and Distillation in Open-Response Assessment

no code implementations15 Jan 2025 Conrad Borchers, Danielle R. Thomas, Jionghao Lin, Ralph Abboud, Kenneth R. Koedinger

We conclude that integrating human and LLM-generated data to improve text classification models in assessment offers a scalable solution that leverages both the accuracy of human coding and the variety of LLM outputs.

Language Modeling Language Modelling +3

Toward Sufficient Statistical Power in Algorithmic Bias Assessment: A Test for ABROCA

no code implementations8 Jan 2025 Conrad Borchers

Specifically, we address (1) whether ABROCA follows any known distribution, (2) how to reliably test for algorithmic bias using ABROCA, and (3) the statistical power achievable with ABROCA-based bias assessments under typical EDM sample specifications.

Fairness

Combining Large Language Models with Tutoring System Intelligence: A Case Study in Caregiver Homework Support

1 code implementation16 Dec 2024 Devika Venugopalan, Ziwen Yan, Conrad Borchers, Jionghao Lin, Vincent Aleven

We contribute insights into how tutoring systems can best be merged with LLMs to support hybrid tutoring settings through conversational assistance, facilitating effective caregiver involvement in tutoring systems.

Large Language Model Math +1

Do Tutors Learn from Equity Training and Can Generative AI Assess It?

1 code implementation15 Dec 2024 Danielle R. Thomas, Conrad Borchers, Sanjit Kakarla, Jionghao Lin, Shambhavi Bhushan, Boyuan Guo, Erin Gatz, Kenneth R. Koedinger

Advances in generative AI via large language models (LLMs) are being used in a wide range of applications, with this present work assessing its use in the equity domain.

Few-Shot Learning

Does Multiple Choice Have a Future in the Age of Generative AI? A Posttest-only RCT

1 code implementation13 Dec 2024 Danielle R. Thomas, Conrad Borchers, Sanjit Kakarla, Jionghao Lin, Shambhavi Bhushan, Boyuan Guo, Erin Gatz, Kenneth R. Koedinger

Using a posttest-only randomized control design, we compare the performance of 234 tutors (790 lesson completions) across three conditions: MCQ only, open response only, and a combination of both.

Multiple-choice

Evaluating the Impact of Data Augmentation on Predictive Model Performance

no code implementations3 Dec 2024 Valdemar Švábenský, Conrad Borchers, Elizabeth B. Cloude, Atsushi Shimada

This paper systematically compares data augmentation techniques and their impact on prediction performance in a typical LA task: prediction of academic outcomes.

Data Augmentation

ABROCA Distributions For Algorithmic Bias Assessment: Considerations Around Interpretation

1 code implementation28 Nov 2024 Conrad Borchers, Ryan S. Baker

When assessing whether a classifier is biased, this skewness inflates ABROCA values by chance, even when data is drawn (by simulation) from populations with equivalent ROC curves.

Fairness

Data Augmentation for Sparse Multidimensional Learning Performance Data Using Generative AI

1 code implementation24 Sep 2024 Liang Zhang, Jionghao Lin, John Sabatini, Conrad Borchers, Daniel Weitekamp, Meng Cao, John Hollander, Xiangen Hu, Arthur C. Graesser

Second, a tensor factorization method is used to impute missing values in sparse tensors of collected learner data, thereby grounding the imputation on knowledge tracing tasks that predict missing performance values based on real observations.

ARC Data Augmentation +4

Predicting Learning Performance with Large Language Models: A Study in Adult Literacy

no code implementations4 Mar 2024 Liang Zhang, Jionghao Lin, Conrad Borchers, John Sabatini, John Hollander, Meng Cao, Xiangen Hu

This research is motivated by the potential of LLMs to predict learning performance based on its inherent reasoning and computational capabilities.

Knowledge Tracing Lifelong learning +1

Improving Assessment of Tutoring Practices using Retrieval-Augmented Generation

no code implementations4 Feb 2024 Zifei, Han, Jionghao Lin, Ashish Gurung, Danielle R. Thomas, Eason Chen, Conrad Borchers, Shivang Gupta, Kenneth R. Koedinger

The results indicate that the RAG prompt demonstrated more accurate performance (assessed by the level of hallucination and correctness in the generated assessment texts) and lower financial costs than the other strategies evaluated.

Hallucination Math +3

3DG: A Framework for Using Generative AI for Handling Sparse Learner Performance Data From Intelligent Tutoring Systems

1 code implementation29 Jan 2024 Liang Zhang, Jionghao Lin, Conrad Borchers, Meng Cao, Xiangen Hu

Learning performance data (e. g., quiz scores and attempts) is significant for understanding learner engagement and knowledge mastery level.

Generative Adversarial Network Imputation

Revealing Networks: Understanding Effective Teacher Practices in AI-Supported Classrooms using Transmodal Ordered Network Analysis

1 code implementation17 Dec 2023 Conrad Borchers, Yeyu Wang, Shamya Karumbaiah, Muhammad Ashiq, David Williamson Shaffer, Vincent Aleven

Taken together, offering early conceptual support to students with low learning rates could make classroom practice with AI tutors more effective.

Using Think-Aloud Data to Understand Relations between Self-Regulation Cycle Characteristics and Student Performance in Intelligent Tutoring Systems

1 code implementation9 Dec 2023 Conrad Borchers, Jiayi Zhang, Ryan S. Baker, Vincent Aleven

We discuss system re-design opportunities to add SRL support during stages of processing and paths forward for using machine learning to speed research depending on the assessment of SRL based on transcription of think-aloud data.

Insights into undergraduate pathways using course load analytics

1 code implementation20 Dec 2022 Conrad Borchers, Zachary A. Pardos

Course load analytics (CLA) inferred from LMS and enrollment features can offer a more accurate representation of course workload to students than credit hours and potentially aid in their course selection decisions.

Cannot find the paper you are looking for? You can Submit a new open access paper.