Performance Evaluation of Large Language Model Chatbots for Radiation Therapy Education

This study aimed to develop a large language model (LLM) chatbot for radiation therapy education and compare the performance of portable document format (PDF)- and webpage-based question-and-answer (Q&A) chatbots. An LLM chatbot was created using the EmbedChain framework, OpenAI GPT-3.5-Turbo AP...

Full description

Saved in:
Bibliographic Details
Main Authors: Jae-Hong Jung, Daegun Kim, Kyung-Bae Lee, Youngjin Lee
Format: Article
Language:English
Published: MDPI AG 2025-06-01
Series:Information
Subjects:
Online Access:https://www.mdpi.com/2078-2489/16/7/521
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1839615879120158720
author Jae-Hong Jung
Daegun Kim
Kyung-Bae Lee
Youngjin Lee
author_facet Jae-Hong Jung
Daegun Kim
Kyung-Bae Lee
Youngjin Lee
author_sort Jae-Hong Jung
collection DOAJ
description This study aimed to develop a large language model (LLM) chatbot for radiation therapy education and compare the performance of portable document format (PDF)- and webpage-based question-and-answer (Q&A) chatbots. An LLM chatbot was created using the EmbedChain framework, OpenAI GPT-3.5-Turbo API, and Gradio UI. The performance of both chatbots was evaluated based on 10 questions and their corresponding answers, using the parameters of accuracy, semantic similarity, consistency, and response time. The accuracy scores were 0.672 and 0.675 for the PDF- and webpage-based Q&A chatbots, respectively. The semantic similarity between the two chatbots was 0.928 (92.8%). The consistency score was one for both chatbots. The average response time was 3.3 s and 2.38 s for the PDF- and webpage-based chatbots, respectively. The LLM chatbot developed in this study demonstrates the potential to provide reliable responses for radiation therapy education. However, its reliability and efficiency must be further optimized to be effectively utilized as an educational tool.
format Article
id doaj-art-0a0c8c2f5933468e9cd16bf6c2060d65
institution Matheson Library
issn 2078-2489
language English
publishDate 2025-06-01
publisher MDPI AG
record_format Article
series Information
spelling doaj-art-0a0c8c2f5933468e9cd16bf6c2060d652025-07-25T13:25:00ZengMDPI AGInformation2078-24892025-06-0116752110.3390/info16070521Performance Evaluation of Large Language Model Chatbots for Radiation Therapy EducationJae-Hong Jung0Daegun Kim1Kyung-Bae Lee2Youngjin Lee3Department of Radiation Oncology, College of Medicine, Soonchunhyang University Hospital Bucheon, 170, Jomaru-ro, Wonmi-gu, Bucheon-si 14584, Republic of KoreaDepartment of Radiation Oncology, College of Medicine, Soonchunhyang University Hospital Bucheon, 170, Jomaru-ro, Wonmi-gu, Bucheon-si 14584, Republic of KoreaDepartment of Radiotechnology, Wonkwang Health Science University, 514, Iksan-daero, Iksan-si 54538, Republic of KoreaDepartment of Radiological Science, Gachon University, 191, Hambakmoe-ro, Yeonsu-gu, Incheon 21936, Republic of KoreaThis study aimed to develop a large language model (LLM) chatbot for radiation therapy education and compare the performance of portable document format (PDF)- and webpage-based question-and-answer (Q&A) chatbots. An LLM chatbot was created using the EmbedChain framework, OpenAI GPT-3.5-Turbo API, and Gradio UI. The performance of both chatbots was evaluated based on 10 questions and their corresponding answers, using the parameters of accuracy, semantic similarity, consistency, and response time. The accuracy scores were 0.672 and 0.675 for the PDF- and webpage-based Q&A chatbots, respectively. The semantic similarity between the two chatbots was 0.928 (92.8%). The consistency score was one for both chatbots. The average response time was 3.3 s and 2.38 s for the PDF- and webpage-based chatbots, respectively. The LLM chatbot developed in this study demonstrates the potential to provide reliable responses for radiation therapy education. However, its reliability and efficiency must be further optimized to be effectively utilized as an educational tool.https://www.mdpi.com/2078-2489/16/7/521radiation therapy educationprofessional educationlarge language modelchatbot
spellingShingle Jae-Hong Jung
Daegun Kim
Kyung-Bae Lee
Youngjin Lee
Performance Evaluation of Large Language Model Chatbots for Radiation Therapy Education
Information
radiation therapy education
professional education
large language model
chatbot
title Performance Evaluation of Large Language Model Chatbots for Radiation Therapy Education
title_full Performance Evaluation of Large Language Model Chatbots for Radiation Therapy Education
title_fullStr Performance Evaluation of Large Language Model Chatbots for Radiation Therapy Education
title_full_unstemmed Performance Evaluation of Large Language Model Chatbots for Radiation Therapy Education
title_short Performance Evaluation of Large Language Model Chatbots for Radiation Therapy Education
title_sort performance evaluation of large language model chatbots for radiation therapy education
topic radiation therapy education
professional education
large language model
chatbot
url https://www.mdpi.com/2078-2489/16/7/521
work_keys_str_mv AT jaehongjung performanceevaluationoflargelanguagemodelchatbotsforradiationtherapyeducation
AT daegunkim performanceevaluationoflargelanguagemodelchatbotsforradiationtherapyeducation
AT kyungbaelee performanceevaluationoflargelanguagemodelchatbotsforradiationtherapyeducation
AT youngjinlee performanceevaluationoflargelanguagemodelchatbotsforradiationtherapyeducation