Eposter Presentation
 
Accept format: PDF. The file size should not be more than 5MB
 
Accept format: PNG/JPG/WEBP. The file size should not be more than 2MB
 
Draft
Abstract
Comparative Evaluation of AI Models for Generating Urology Admission Summaries Using QNOTE Scores
Podium Abstract
Clinical Research
AI in Urology
Author's Information
4
No more than 10 authors can be listed (as per the Good Publication Practice (GPP) Guidelines).
Please ensure the authors are listed in the right order.
Taiwan
Liang-Chen Huang sam831009@gmail.com En Chu Kong Hospital Urology New Taipei City Taiwan *
Jung-Yang Yu ericyu29218218@gmail.com National Taiwan University Hospital Urology Taipei City Taiwan
Chung-Cheng Wang ericwcc@ms27.hinet.net En Chu Kong Hospital Urology New Taipei City Taiwan
Juan-Hua Hong cliffordhong622@gmail.com National Taiwan University Hospital Urology Taipei City Taiwan
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Abstract Content
The urology department routinely handles a high volume of inpatient elective procedures, contributing to a demanding clinical environment. The application of Artificial Intelligence (AI) has the potential to streamline repetitive tasks, particularly in clinical documentation. This study aims to evaluate the performance and accuracy of generative AI and large language models (GAI/LLMs) in producing admission summaries based on outpatient clinic notes.
Patients undergoing inpatient elective procedures, arranged through outpatient clinic visits between January and April 2024, were included in this study. AI models, GPT-4 (Model 1) and GPT-4o (Model 2), were prompted to generate admission summaries based on a single outpatient clinic note per patient. The quality of the generated summaries was assessed using the QNOTE scoring system, a non-disease-specific, 12-category, 44-individual-item rubric that evaluates the quality of clinical documentation across various domains.
A total of 14 patients were included in the evaluation. Both AI models produced high-quality admission summaries, with Model 1 (GPT-4) achieving an average QNOTE score of 87.92, and Model 2 (GPT-4o) scoring 91.48 (out of 100). Both models achieved perfect scores in several categories. However, Model 2 consistently outperformed Model 1 in both subjective assessments and across multiple QNOTE domains. The distribution of QNOTE scores for the outpatient clinic notes, comparing Model 1 and Model 2, is shown in Figure 1.
GAI/LLMs demonstrate the capability to generate high-quality admission summaries for inpatient elective urology procedures based on a single outpatient clinic note. GPT-4o outperformed GPT-4 in both objective and subjective evaluations. While these AI models show strong potential, it remains essential for clinicians to review the generated summaries for accuracy and consistency. Further research with larger sample sizes and continued development of AI models are necessary to validate these findings and refine their clinical application.
large language models, repetitive documentation, efficiency
https://storage.unitedwebnetwork.com/files/1237/1f48bf101b62211fc5ed576d7e6fe86b.png
Qualitative assessment of the clinical note. The bars represent the percentage of different components of the 12 elements of QNOTE and an overall note score which is at the bottom of the chart.
 
 
 
 
 
 
 
 
2043
 
Presentation Details