Introduction: Generative artificial intelligence (AI) technologies like GPT-4 can instantaneously provide health information to patients; however, the readability of these outputs compared to ophthalmologist-written responses is unknown. This study aimed to evaluate the readability of GPT-4-generated and ophthalmologist-written responses to patient queries about ophthalmic surgery. Methods: This retrospective cross-sectional study used 200 randomly selected patient questions about ophthalmic surgery extracted from the American Academy of Ophthalmology’s EyeSmart platform. The questions were inputted into GPT-4, and the generated responses were recorded. Ophthalmologist-written replies to the same questions were compiled for comparison. Readability of GPT-4 and ophthalmologist responses was assessed using six validated metrics: Flesch Kincaid Reading Ease (FK-RE), Flesch Kincaid Grade Level (FK-GL), Gunning Fog Score (GFS), SMOG Index (SI), Coleman Liau Index (CLI), and Automated Readability Index (ARI). Descriptive statistics, one-way ANOVA, Shapiro-Wilk, and Levene’s tests (α = 0.05) were used to compare readability between the two groups. Results: GPT-4 used a higher percentage of complex words (24.42%) compared to ophthalmologists (17.76%), although mean (standard deviation) word count per sentence was similar (18.43 [2.95] and 18.01 [6.09]). Across all metrics (FK-RE; FK-GL; GFS; SI; CLI; and ARI), GPT-4 responses were at a higher grade level (34.39 [8.51]; 13.19 [2.63]; 16.37 [2.04]; 12.18 [1.43]; 15.72 [1.40]; 12.99 [1.86]) than ophthalmologists’ responses (50.61 [15.53]; 10.71 [2.99]; 14.13 [3.55]; 10.07 [2.46]; 12.64 [2.93]; 10.40 [3.61]), with both sources necessitating a 12th-grade education for comprehension. ANOVA tests showed significance (p < 0.05) for all comparisons except word count (p = 0.438). Conclusion: The National Institutes of Health advises health information to be written at a 6th- to 7th-grade level. Both GPT-4- and ophthalmologist-written answers exceeded this recommendation, with GPT-4 showing a greater gap. Information accessibility is vital when designing patient resources, particularly with the rise of AI as an educational tool.

As technology continues to evolve, it is important to understand how it can help but also harm certain groups of people. The presence of artificial intelligence (AI) in medicine is rapidly growing. Specifically within ophthalmology, recent studies have shown that AI like ChatGPT-4 is capable of answering patient questions in similar quality levels to human ophthalmologists. How readable this information output by AI is to the average patient seeking these data is important to understanding its accessibility. Therefore, in this study we aimed to compare the readability level of responses from AI and human ophthalmologists to inform future discussions on information accessibility. Our results showed that a higher education level is needed to understand responses from ChatGPT-4 (14th grade) compared to ophthalmologists (12th grade). This is much higher than the average reading level of adults in the USA (8th). Additionally, both required an education level significantly higher than the readability recommendation for medical information from national health organizations (6–7th grade). Our findings emphasize the importance of ensuring information accessibility in the development and usage of technology in the medical space. Failing to prioritize these factors could worsen healthcare disparities for patient populations already at a disadvantage in society.

1.
Magnitude and Projections
.
The International Agency for the Prevention of Blindness
. Available from: https://www.iapb.org/learn/vision-atlas/magnitude-and-projections/ (accessed March 10, 2024).
2.
Global blindness was slowing prior to pandemic study reveals
.
Orbis
;
2024
. Available from: https://www.orbis.org/en/news/2021/new-global-blindness-data (accessed March 10, 2024).
3.
Iskander
M
,
Hu
G
,
Coulon
S
,
Seixas
AA
,
McGowan
R
,
Al-Aswad
LA
.
Health literacy and ophthalmology: a scoping review
.
Surv Ophthalmol
.
2023
;
68
(
1
):
78
103
.
4.
Gualtieri
LN
.
The doctor as the second opinion and the internet as the first
. In:
CHI ’09 extended abstracts on human factors in computing systems. CHI EA ’09
.
Association for Computing Machinery
;
2009
. p.
2489
98
.
5.
Tonsaker
T
,
Bartlett
G
,
Trpkov
C
.
Health information on the internet
.
Can Fam Physician
.
2014
;
60
(
5
):
407
8
.
6.
Jia
X
,
Pang
Y
,
Liu
LS
.
Online health information seeking behavior: a systematic review
.
Healthcare
.
2021
;
9
(
12
):
1740
.
7.
Iverson
SA
,
Howard
KB
,
Penney
BK
.
Impact of internet use on health-related behaviors and the patient-physician relationship: a survey-based study and review
.
J Am Osteopath Assoc
.
2008
;
108
(
12
):
699
711
.
8.
Tan
SS-L
,
Goonawardene
N
.
Internet health information seeking and the patient-physician relationship: a systematic review
.
J Med Internet Res
.
2017
;
19
(
1
):
e9
.
9.
Hutchinson
N
,
Baird
GL
,
Garg
M
.
Examining the reading level of internet medical information for common internal medicine diagnoses
.
Am J Med
.
2016
;
129
(
6
):
637
9
.
10.
Zahlan
A
,
Ranjan
RP
,
Hayes
D
.
Artificial intelligence innovation in healthcare: literature review, exploratory analysis, and future research
.
Technol Soc
.
2023
;
74
:
102321
.
11.
Nyblade
L
,
Stockton
MA
,
Giger
K
,
Bond
V
,
Ekstrand
ML
,
Lean
RM
, et al
.
Stigma in health facilities: why it matters and how we can change it
.
BMC Med
.
2019
;
17
(
1
):
25
.
12.
Reading level of internet medical information for common diagnoses
.
The American Journal of Medicine Blog
;
2016
. Available from: https://amjmed.org/reading-level-of-internet-medical-information-for-common-diagnoses/ (accessed November 23, 2023).
13.
Eltorai
AEM
,
Ghanian
S
,
Adams
CA
,
Born
CT
,
Daniels
AH
.
Readability of patient education materials on the American association for surgery of trauma website
.
Arch Trauma Res
.
2014
;
3
(
2
):
e18161
.
14.
Safeer
RS
,
Keenan
J
.
Health literacy: the gap between physicians and patients
.
Am Fam Physician
.
2005
;
72
(
3
):
463
8
.
15.
Rooney
MK
,
Santiago
G
,
Perni
S
,
Horowitz
DP
,
McCall
AR
,
Einstein
AJ
, et al
.
Readability of patient education materials from high-impact medical journals: a 20-year analysis
.
J Patient Exp
.
2021
;
8
:
2374373521998847
.
16.
Rodriguez
F
,
Ngo
S
,
Baird
G
,
Balla
S
,
Miles
R
,
Garg
M
.
Readability of online patient educational materials for coronary artery calcium scans and implications for health disparities
.
J Am Heart Assoc
.
2020
;
9
(
18
):
e017372
.
17.
Weiss
B
.
Health literacy: a manual for clinicians
.
American Medical Association Foundation
. Available from: http://lib.ncfh.org/pdfs/6617.pdf (accessed November 23, 2023).
18.
Baker
DW
,
Parker
RM
,
Williams
MV
,
Clark
WS
,
Nurss
J
.
The relationship of patient reading ability to self-reported health and use of health services
.
Am J Public Health
.
1997
;
87
(
6
):
1027
30
.
19.
Shedlosky-Shoemaker
R
,
Sturm
AC
,
Saleem
M
,
Kelly
KM
.
Tools for assessing readability and quality of health-related Web sites
.
J Genet Couns
.
2009
;
18
(
1
):
49
59
.
21.
About health literacy
. Available from: https://www.ahrq.gov/health-literacy/about/index.html (accessed November 23, 2023).
22.
Williams
AM
,
Muir
KW
,
Rosdahl
JA
.
Readability of patient education materials in ophthalmology: a single-institution study and systematic review
.
BMC Ophthalmol
.
2016
;
16
(
1
):
133
.
23.
Patel
PA
,
Gopali
R
,
Reddy
A
,
Patel
KK
.
The readability of ophthalmological patient education materials provided by major academic hospitals
.
Semin Ophthalmol
.
2022
;
37
(
1
):
71
6
.
24.
Cohen
SA
,
Pershing
S
.
Readability and accountability of online patient education materials for common retinal diseases
.
Ophthalmol Retina
.
2022
;
6
(
7
):
641
3
.
25.
Lin
MX
,
Li
G
,
Cui
D
,
Mathews
PM
,
Akpek
EK
.
Usability of patient education–oriented cataract surgery websites
.
Ophthalmology
.
2023
;
131
(
4
):
499
506
.
26.
Dave
T
,
Athaluri
SA
,
Singh
S
.
ChatGPT in medicine: an overview of its applications, advantages, limitations, future prospects, and ethical considerations
.
Front Artif Intell
.
2023
;
6
:
1169595
.
27.
Mondal
H
,
Mondal
S
,
Podder
I
.
Using ChatGPT for writing articles for patients’ education for dermatological diseases: a pilot study
.
Indian Dermatol Online J
.
2023
;
14
(
4
):
482
6
.
28.
Zalzal
HG
,
Abraham
A
,
Cheng
J
,
Shah
RK
.
Can ChatGPT help patients answer their otolaryngology questions
.
Laryngoscope Investig Otolaryngol
.
2024
;
9
(
1
):
e1193
.
29.
Bernstein
IA
,
Zhang
YV
,
Govil
D
,
Majid
I
,
Chang
RT
,
Sun
Y
, et al
.
Comparison of ophthalmologist and large language model chatbot responses to online patient eye care questions
.
JAMA Netw Open
.
2023
;
6
(
8
):
e2330320
.
30.
Readability test
.
WebFX
. Available from: https://www.webfx.com/tools/read-able/ (accessed November 23, 2023).
31.
Dowdle
TS
,
Nguyen
JM
,
Steadman
JN
,
Layher
H
,
Sturgeon
ALE
,
Akin
R
.
Online readability analysis: Mohs micrographic surgery postsurgical wound care
.
Adv Skin Wound Care
.
2022
;
35
(
4
):
213
8
.
32.
Ganjavi
C
,
Eppler
MB
,
Ramacciotti
LS
,
Cacciamani
GE
.
Clinical patient summaries not fit for purpose: a study in urology
.
Eur Urol Focus
.
2023
;
9
(
6
):
1068
71
.
33.
Flesch
R
.
Guide to academic writing article: how to write plain English
. Chapter 2: let’s start with the formula.
2016
. Available from: https://web.archive.org/web/20160712094308/http://www.mang.canterbury.ac.nz/writing_guide/writing/flesch.shtml (accessed November 23, 2023).
34.
Doak
CC
,
Doak
LG
,
Friedell
GH
,
Meade
CD
.
Improving comprehension for cancer patients with low literacy skills: strategies for clinicians
.
CA Cancer J Clin
.
1998
;
48
(
3
):
151
62
.
35.
Ebrahimzadeh
H
,
Davalos
R
,
Lee
PP
.
Literacy levels of ophthalmic patient education materials
.
Surv Ophthalmol
.
1997
;
42
(
2
):
152
6
.
36.
Muir
KW
,
Lee
PP
.
Health literacy and ophthalmic patient education
.
Surv Ophthalmol
.
2010
;
55
(
5
):
454
9
.
37.
GPT-4
. Available from: https://openai.com/research/gpt-4 (accessed November 23, 2023).
38.
Oh
N
,
Choi
G-S
,
Lee
WY
.
ChatGPT goes to the operating room: evaluating GPT-4 performance and its potential in surgical education and training in the era of large language models
.
Ann Surg Treat Res
.
2023
;
104
(
5
):
269
73
.
39.
Scheschenja
M
,
Viniol
S
,
Bastian
MB
,
Wessendorf
J
,
König
AM
,
Mahnken
AH
.
Feasibility of GPT-3 and GPT-4 for in-depth patient education prior to interventional radiological procedures: a comparative analysis
.
Cardiovasc Interv Radiol
.
2024
;
47
(
2
):
245
50
.
40.
Williams
MV
,
Davis
T
,
Parker
RM
,
Weiss
BD
.
The role of health literacy in patient-physician communication
.
Fam Med
.
2002
;
34
(
5
):
383
9
.
41.
Wittink
H
,
Oosterhaven
J
.
Patient education and health literacy
.
Musculoskelet Sci Pract
.
2018
;
38
:
120
7
.
42.
Muir
KW
,
Santiago-Turla
C
,
Stinnett
SS
,
Herndon
LW
,
Allingham
RR
,
Challa
P
, et al
.
Health literacy and adherence to glaucoma therapy
.
Am J Ophthalmol
.
2006
;
142
(
2
):
223
6
.
43.
Muir
KW
,
Santiago-Turla
C
,
Stinnett
SS
,
Herndon
LW
,
Allingham
RR
,
Challa
P
, et al
.
Health literacy and vision-related quality of life
.
Br J Ophthalmol
.
2008
;
92
(
6
):
779
82
.
44.
Jandorf
S
,
Krogh Nielsen
M
,
Sørensen
K
,
Sørensen
TL
.
Low health literacy levels in patients with chronic retinal disease
.
BMC Ophthalmol
.
2019
;
19
(
1
):
174
.
45.
McDonald
L
,
Turnbull
P
,
Chang
L
,
Crabb
DP
.
Taking the strain? Impact of glaucoma on patients’ informal caregivers
.
Eye
.
2020
;
34
(
1
):
197
204
.
46.
Schillinger
D
,
Grumbach
K
,
Piette
J
,
Wang
F
,
Osmond
D
,
Daher
C
, et al
.
Association of health literacy with diabetes outcomes
.
JAMA
.
2002
;
288
(
4
):
475
82
.
47.
Juzych
MS
,
Randhawa
S
,
Shukairy
A
,
Kaushal
P
,
Gupta
A
,
Shalauta
N
.
Functional health literacy in patients with glaucoma in urban settings
.
Arch Ophthalmol
.
2008
;
126
(
5
):
718
24
.
48.
Scott
B
.
Readability formulas and the active role of the reader
.
ReadabilityFormulas.com
;
2023
. Available from: https://readabilityformulas.com/readability-formulas-and-the-active-role-of-the-reader/ (accessed November 23, 2023).
You do not currently have access to this content.