ASEE PEER - Evaluation of LLMs and Other Machine Learning Methods in the Analysis of Qualitative Survey Responses for Accessible Engineering Education Research
Asee peer logo

Evaluation of LLMs and Other Machine Learning Methods in the Analysis of Qualitative Survey Responses for Accessible Engineering Education Research

Download Paper |

Conference

2024 ASEE Annual Conference & Exposition

Location

Portland, Oregon

Publication Date

June 23, 2024

Start Date

June 23, 2024

End Date

July 12, 2024

Conference Session

Educational Research and Methods Division (ERM) Technical Session 7

Tagged Division

Educational Research and Methods Division (ERM)

Tagged Topic

Diversity

Permanent URL

https://peer.asee.org/47360

Request a correction

Paper Authors

author page

Xiuhao Ding University of Illinois at Urbana - Champaign

author page

Meghana Gopannagari University of Illinois at Urbana - Champaign

author page

Kang Sun University of Illinois at Urbana - Champaign

author page

Alan Tao University of Illinois at Urbana - Champaign

author page

Delu Louis Zhao

biography

Sujit Varadhan University of Illinois at Urbana - Champaign

visit author page

Sujit Varadhan is a Junior at the University of Illinois at Urbana-Champaign majoring in Computer Science. He is an undergraduate research assistant as well as a frontend developer on ClassTranscribe.

visit author page

author page

Bobbi Lee Battleson Hardy University of Illinois at Urbana - Champaign

author page

David Dalpiaz University of Illinois at Urbana - Champaign

biography

Chrysafis Vogiatzis University of Illinois at Urbana - Champaign Orcid 16x16 orcid.org/0000-0003-0787-9380

visit author page

Dr. Chrysafis Vogiatzis is a teaching associate professor for the Department of Industrial and Enterprise Systems Engineering at the University of Illinois Urbana-Champaign. Prior to that, Dr. Vogiatzis was an assistant professor at North Carolina Agricultural and Technical State University. His current research interests lie in network optimization and combinatorial optimization, along with their vast applications in modern socio-technical and biological systems. He is serving as the faculty advisor of the Institute of Industrial and Systems Engineers, and was awarded the 2019 and 2023 Faculty Advisor award for the North-Central region of IISE. Dr. Vogiatzis was awarded ASEE IL/IN Teacher of the Year in 2023.

visit author page

biography

Lawrence Angrave University of Illinois at Urbana - Champaign Orcid 16x16 orcid.org/0000-0001-9762-7181

visit author page

Dr. Lawrence Angrave is an award-winning computer science Teaching Professor at the University of Illinois Urbana-Champaign. He creates and researches new opportunities for accessible and inclusive equitable education.

visit author page

biography

Hongye Liu University of Illinois at Urbana - Champaign

visit author page

Hongye Liu is a Teaching Assistant Professor in the Dept. of Computer Science in UIUC. She is interested in education research to help students with disability and broaden participation in computer science.

visit author page

Download Paper |

Abstract

This research paper provides insights and guidance for selecting appropriate analytical tools in engineering educational research. Currently, educators and researchers face difficulties in gaining insights effectively from free-response survey data. We evaluate the effectiveness and accuracy of Large Language Models (LLMs), in addition to the existing methods that employ topic modeling, document clustering coupled with Support Vector Machine (SVM) and Random Forest (RF) approaches, and the unsupervised Latent Dirichlet Allocation (LDA) method. Free responses to open-ended questions from student surveys in multiple courses at University of Illinois Urbana-Champaign were previously collected by engineering education accessibility researchers. The data (N=129 with seven free response questions per student) were previously analyzed to assess the effectiveness, satisfaction, and quality of adding accessible digital notes to multiple engineering courses and the students’ perceived belongingness, and self-efficacy. Manual codings for the seven open-ended questions were generated for qualitative tasks of sentiment analysis, topic modeling, and summarization and were used in this study as a gold standard to evaluate automated text analytic approaches. Raw text from open-ended questions was converted into numerical vectors using text vectorization and word embeddings and an unsupervised analysis using document clustering and topic modeling was performed using LDA and BERT methods. In addition to conventional machine learning models, multiple pre-trained open-sourced local LLMs were evaluated (BART and LLaMA) for summarization. The remote online ChatGPT closed-model services by OpenAI (ChatGPT-3.5 and ChatGPT-4) were excluded due to subject data privacy concerns. By comparing the accuracy, recall, and depth of thematic insights derived, we evaluated how effectively the method based on each model categorized and summarized students’ responses across educational research interests of effectiveness, satisfaction, and quality of education materials. The paper will present these results and discuss the implications of our findings and conclusions.

Ding, X., & Gopannagari, M., & Sun, K., & Tao, A., & Zhao, D. L., & Varadhan, S., & Hardy, B. L. B., & Dalpiaz, D., & Vogiatzis, C., & Angrave, L., & Liu, H. (2024, June), Evaluation of LLMs and Other Machine Learning Methods in the Analysis of Qualitative Survey Responses for Accessible Engineering Education Research Paper presented at 2024 ASEE Annual Conference & Exposition, Portland, Oregon. https://peer.asee.org/47360

ASEE holds the copyright on this document. It may be read by the public free of charge. Authors may archive their work on personal websites or in institutional repositories with the following citation: © 2024 American Society for Engineering Education. Other scholars may excerpt or quote from these materials with the same citation. When excerpting or quoting from Conference Proceedings, authors should, in addition to noting the ASEE copyright, list all the original authors and their institutions and name the host city of the conference. - Last updated April 1, 2015