Data

The performance of ChatGPT-4.0o in medical imaging evaluation: a preliminary investigation

Charles Sturt University
Arruzza, Elio Stefan ; Evangelista, Carla Marie ; Chau, Minh
Viewed: [[ro.stat.viewed]] Cited: [[ro.stat.cited]] Accessed: [[ro.stat.accessed]]
ctx_ver=Z39.88-2004&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Adc&rfr_id=info%3Asid%2FANDS&rft_id=info:doi10.7910/dvn/qq8pko&rft.title=The performance of ChatGPT-4.0o in medical imaging evaluation: a preliminary investigation&rft.identifier=10.7910/dvn/qq8pko&rft.publisher=Harvard Dataverse&rft.description=This study investigated the performance of ChatGPT-4.0o in evaluating the quality of positioning in radiographic images. Thirty radiographs depicting a variety of knee, elbow, ankle, hand, pelvis, and shoulder projections were produced using anthropomorphic phantoms and uploaded to ChatGPT-4.0o. The model was prompted to provide a solution to identify any positioning errors with justification and offer improvements. A panel of radiographers assessed the solutions for radiographic quality based on established positioning criteria, with a grading scale of 1–5. In only 20% of projections, ChatGPT-4.0o correctly recognized all errors with justifications and offered correct suggestions for improvement. The most commonly occurring score was 3 (9 cases, 30%), wherein the model recognized at least 1 specific error and provided a correct improvement. The mean score was 2.9. Overall, low accuracy was demonstrated, with most projections receiving only partially correct solutions. The findings reinforce the importance of robust radiography education and clinical experience.&rft.creator=Arruzza, Elio Stefan &rft.creator=Evangelista, Carla Marie &rft.creator=Chau, Minh &rft.date=2024&rft.relation=http://researchoutput.csu.edu.au/en/publications/30ecd36d-5083-403d-a91b-3e57589bb7ae&rft_subject=Medicine, Health and Life Sciences&rft.type=dataset&rft.language=English Access the data

Access:

Open

Full description

This study investigated the performance of ChatGPT-4.0o in evaluating the quality of positioning in radiographic images. Thirty radiographs depicting a variety of knee, elbow, ankle, hand, pelvis, and shoulder projections were produced using anthropomorphic phantoms and uploaded to ChatGPT-4.0o. The model was prompted to provide a solution to identify any positioning errors with justification and offer improvements. A panel of radiographers assessed the solutions for radiographic quality based on established positioning criteria, with a grading scale of 1–5. In only 20% of projections, ChatGPT-4.0o correctly recognized all errors with justifications and offered correct suggestions for improvement. The most commonly occurring score was 3 (9 cases, 30%), wherein the model recognized at least 1 specific error and provided a correct improvement. The mean score was 2.9. Overall, low accuracy was demonstrated, with most projections receiving only partially correct solutions. The findings reinforce the importance of robust radiography education and clinical experience.

Notes

External Organisations
University of South Australia; Jones Radiology
Associated Persons
Elio Stefan Arruzza (Contributor); Carla Marie Evangelista (Creator)

Created: 2024-10-30 to 2024-10-30

Issued: 2024-10-30

Subjects

User Contributed Tags    

Login to tag this record with meaningful keywords to make it easier to discover

Identifiers