Keywords
retina - optical coherence tomography - artificial intelligence - retinal imaging
Introduction
The number of patients exhibiting retinal diseases is steadily increasing. Retinal
diseases are also the most common cause of visual impairment and loss of vision amongst
older people in industrialised nations [1]. This has led to rapid development in AI (artificial intelligence) algorithms towards
automatically analysing the immense amounts of data collected every day in clinical
routine. Patients with age-related macular degeneration (AMD) and diabetic retinopathy
(DR) in particular require regular monitoring to detect complications requiring treatment
at an early stage for conditions that may otherwise lead to potentially irreversible
vision loss such as neovascular AMD (nAMD) and diabetic macular oedema (DME) [2]. Apart from that, new treatment options and imaging techniques are constantly leading
to new developments and breakthroughs in ophthalmology. The introduction of intravitreal
anti-vascular endothelial growth
factor (VEGF) injections for treating nAMD and DME have contributed to a substantial
decrease in legal blindness [3], [4]. The number of patients receiving regular intravitreal injections and follow-up
visits is set to increase with the approval of new treatments in the USA for geographic
atrophy (GA), the dry late form of AMD [5], [6]. High-resolution imaging methods such as optical coherence tomography (OCT) provide
new and detailed insights into disease processes as well as mechanisms of action in
new forms of treatment [7]. Ophthalmologists cannot detect most of these subclinical biomarkers in patients
without additional image analysis. AI-based algorithms are able to evaluate large
amounts of data accurately, objectively, and extremely rapidly, thus significantly
reducing workloads and allowing more efficient and individual disease
management for ophthalmologists treating patients in clinical routine as well as in
the healthcare system in general.
Artificial Intelligence and Deep Learning
Artificial Intelligence and Deep Learning
AI is a branch of computer science aimed at training human-like cognitive abilities
in machines or computer systems [8]. AI-based models have mainly become based on machine learning, a subfield of AI.
Machine learning (ML) is a process where algorithms learn principles and structures
independently from given data rather than from preset definitions by human specialists.
However, the ML principle relies on feature engineering, that is, predefined “features”
or biomarkers such as central retinal thickness in an OCT volume. The computer can
then find solutions and self-improve without being specifically programmed to do so.
The “learned” solution approaches can then be applied to new unknown contexts for
more efficient data analysis [8]. Machine learning has since developed into artificial neural networks (ANNs), which
mimic neural networks in the human brain and therefore also human learning behaviour.
ANNs consist of
interconnected artificial neurons arranged in several layers. The information fed
into the network is processed, sorted, and refined for final analysis across several
layers. ANNs are especially well-suited in classification or diagnostic systems such
as in screening for diabetic retinopathy based on biomarkers in colour fundus photography.
ANNs have developed into deep neural networks (DNNs) finally forming a new subfield
– deep learning (DL) [9]. These DNNs consist of several intermediate layers and fewer artificial neurons,
thus increasing their efficiency. A relevant benefit from DNNs is the continuous increase
in performance with the size of the training dataset. Hardware development in computer
processing power has also led to more rapid data processing by DNNs, which has since
exceeded classical ML. DNN models are able to extract “features” from the data autonomously
without human specialists giving them predefined definitions ([Fig. 1]). DNNs with specific convolutional neural network (CNN) architectures have emerged
as the most capable of image analysis and mimic the structure of the human visual
cortex. Specifically, CNNs are able to recognise visual patterns and identify objects
in images. DL-based algorithms need large annotated datasets for training; this improves
precision, but interpretability is more challenging in these models. It is possible
to train ML-based models with smaller annotated datasets, which leads to greater transparency
and interpretability [8].
Fig. 1 Diagram showing classical machine learning and deep learning using diabetic retinopathy
screening as an example. Machine learning algorithms need features – predefined biomarkers
– specified by human specialists, whereas deep learning identifies and classifies
these features autonomously.(Source: Schmidt-Erfurth U, Sadeghipour A, Gerendas BS
et al. Artificial intelligence in retina. Prog Retin Eye Res 2018; 67: 1 – 29. doi:10.1016/j.preteyeres.2018.07.004 [rerif])
Ophthalmology is especially suited for developing and applying novel AI-based models
due to the increasing number of non-invasive high-resolution imaging techniques, especially
on the retina. AI as used in retinal health can be divided into the following categories:
Screening, disease stage classification, segmentation and quantification of lesions,
disease progression prediction, and clinical decision support systems (CDSSs).
Diabetic Retinopathy
Almost 500 million people across the world suffer from diabetes, a number expected
to rise to 600 million by 2040. One-third of those affected exhibit diabetic retinopathy,
around 10% of which are at risk of serious loss of vision. Diabetic retinopathy is
asymptomatic in its early stages, but still requires regular monitoring to detect
later stages or complications that may lead to vision loss [10]. Vision loss is often too late to restore once sustained due to conditions such
as tractional changes in the retina. The increasing number of patients requiring regular
check-ups places a heavy burden on the healthcare system in time and expense, resulting
in limited therapeutic impact in most cases. AI-supported models may relieve the burden
on ophthalmologists and the entire healthcare system while also providing efficient
support for screening and monitoring in clinical routine, especially in identifying
patients in need of treatment.
Diabetic retinopathy: Screening
Developed in 2016, the IDx-DR system for screening diabetic retinopathy from colour
fundus photographs (CFP) was one of the first medical AI applications ([Fig. 2]). The principle is based on DL and showed a sensitivity and specificity of 87% and
91%, respectively, in the validation study. The US Food and Drug Administration (FDA)
approved the system as the first screening method for the presence of diabetic retinopathy
based on these results [11]. The system does not differentiate absence of diabetic retinopathy from presence
of diabetic retinopathy in its mild form as this stage does not require immediate
treatment. Consultation with an ophthalmologist will be recommended if the device
generates a positive result, that is, moderate or severe diabetic retinopathy. EyeArt
is another FDA-approved system already in clinical use for screening diabetic retinopathy
[12]; this system has shown
similar sensitivity and specificity as the IDx-DR system. In contrast, ophthalmologists
show a sensitivity and specificity of 73% and 91%, respectively, in diagnosing diabetic
retinopathy [13]. AI-based models for screening diabetic retinopathy have already come into clinical
use with results on par with, or even surpassing, clinical specialists.
Fig. 2 Diagram showing diabetic retinopathy screening using the IDx-DR system.(Source: Schmidt-Erfurth
U, Sadeghipour A, Gerendas BS et al. Artificial intelligence in retina. Prog Retin
Eye Res 2018; 67: 1 – 29. doi:10.1016/j.preteyeres.2018.07.004 [rerif])
Diabetic retinopathy: Staging
The International Clinical Diabetic Retinopathy Disease Severity Scale classifies
diabetic retinopathy into non-proliferative diabetic retinopathy (NPDR) and proliferative
diabetic retinopathy (PDR) with NPDR further classified into mild, moderate, and severe
[14]. Potential complications include diabetic macular oedema (DME), vitreomacular traction,
epiretinal membrane formation, and retinal detachment. AI-based algorithms can also
assist in distinguishing and classifying different stages in these cases. Two exemplary
studies have investigated algorithms for automatic diabetic retinopathy staging using
large datasets. One study using two datasets with 8,788 and 1,745 CFP images reported
a sensitivity of 90% and 87% with a specificity of 98% in detecting moderate diabetic
retinopathy or worse, a sensitivity of 84% and 88% with a specificity of 98% and 99%
for severe diabetic retinopathy or worse, and a sensitivity of 91% and 90% with a
specificity
of 98% and 99% for DME alone [15]. Another study using almost half a million CFP images achieved a sensitivity/specificity
of 91%/92% and 100%/91% in detecting clinically relevant diabetic retinopathy and
vision-threatening diabetic retinopathy [16].
Diabetic retinopathy: Prediction
We have investigated for OCT criteria affecting treatment outcomes in DME based on
information from 629 eyes from the Protocol-T study and AI-based analysis and found
a prominent correlation between intraretinal cysts and outcomes. In particular, resorption
of intraretinal cysts after the first injection gave the greatest predictive value.
However, it was not possible to give a reliable prediction for central visual acuity
[17].
Age-related Macular Degeneration
Age-related Macular Degeneration
AMD is a chronic progressive disease that leads to progressive loss of retinal tissue
and sensory cells, and therefore also to loss of vision. Around 196 million people
are affected across the world with prevalence expected to increase to 288 million
by 2040 due to increasing population age [18]. The risk of developing AMD increases exponentially with age. AMD is the most common
cause of age-related blindness in the industrialised world. One in four people over
sixty develop AMD [19]. AMD is distinguished into early, intermediate, and late forms depending on the
size of the extracellular deposits or drusen. Late AMD is distinguished into two forms,
neovascular AMD (nAMD) with macular neovascularisation (MNV) and geographic atrophy
(GA) [20]. Injections using anti-VEGF drugs are used to treat nAMD, but no GA treatment has
yet been approved in Europe. In contrast, the FDA has already issued full
approval for the first two drugs as intravitreal complement inhibitors for GA treatment
[5], [6].
Age-related macular degeneration: Screening
AMD is usually asymptomatic in its early stages, but late forms of AMD lead to rapid
and sometimes irreversible vision loss if left untreated. AI-based methods based on
CFP for automatic AMD detection have shown a sensitivity/specificity of 93%/89% as
examined on 72,610 images, and an accuracy of 90% from 130,000 images compared to
human specialists [21]. However, the role of CFP in screening for AMD is still confined to its early stages,
unlike diabetic retinopathy. Late stages exhibit subclinical parameters that cannot
be assessed reliably using CFP. In particular, a reliable assessment of exudation
in nAMD or thinning in the outer retinal layers as early atrophic changes is not possible
using CFP. OCT imaging plays an important role here, as it can faithfully display
subclinical changes at high resolution. A study based on a large dataset comprising
3,265 OCT volumes used in screening for AMD achieved a sensitivity and specificity
of 93%
compared to ophthalmologists [22]. Early forms of GA are especially suitable for screening as they initially occur
outside the fovea, with many patients only showing symptoms once the fovea or its
immediate surroundings have been reached. Major multicentre projects are currently
investigating the possibility of AI-based screening at population level, while also
involving other professional groups such as pharmacists and orthoptists.
Age-related macular degeneration: Staging
Different stages of AMD have different implications in clinical routine and treatment.
In such cases, AI-based algorithms can also help in the clinical setting and in rapid
analysis of large databases for research. Results from OCT volumes from 367 individuals
showed a sensitivity and specificity of 98% and 91% for classification into no AMD,
early AMD, intermediate AMD, GA and nAMD compared to examination by ophthalmologists
[22]. A DL-based algorithm has been developed and validated for automatic AMD classification
into four stages – no AMD, intermediate AMD, nAMD, and GA – using real-world data
on 3,765 OCT volumes from 1,849 eyes in the UK Biobank. The algorithm achieved an
area under the curve (AUC) value of 0.94 [23]. This emphasises the potential of AI-based models in rapidly and reliably analysing
large databases for both clinical routine and retrospective research on disease progression
and conversion.
Age-related macular degeneration: Prediction
Predicting conversion from early or intermediate AMD to late AMD as nAMD or GA is
especially relevant in a clinical setting. Treatment for nAMD should begin as early
as possible after diagnosis to prevent potential complications such as bleeding or
fibrosis that may lead to irreversible loss of vision. The FDA has already approved
two drugs for GA treatment, which could see approval in Europe later this year [5], [6]. Treatment for GA should also begin as soon as possible to preserve central vision
for as long as possible before the onset of subfoveal atrophy. An ML-based predictive
model for automatic prediction of conversion to late AMD was tested on OCT data from
495 patients, achieving an accuracy of 68% for predicting MNV and 80% for predicting
GA [24]. Various characteristic pathomechanisms were identified using an AI-based analysis
of different biomarkers such as drusen, hyperreflective
foci, pseudodrusen, and segmentation of the retinal layers. Drusen volume played the
main role in predicting conversion to MNV, whereas hyperreflective foci and thinning
in the outer neurosensory layers were shown to be especially relevant as biomarkers
for conversion to GA [24]. This demonstrated the ability of AI-based algorithms to provide new insights into
disease progression.
Quantifying different biomarkers is one of the main applications of AI in clinical
routine and in clinical studies towards improving disease activity assessment and
treatment response. Quantifying the various fluid compartments plays an especially
important role in nAMD; these fluid compartments include intraretinal fluid (IRF),
subretinal fluid (SRF), and sub-RPE fluid (RPE: retinal pigment epithelium) for pigment
epithelial detachment (PED) [25]. Automatically quantifying these biomarkers showed IRF in particular to predict
a more adverse prognosis whereas SRF could be tolerated to some extent [26]. Ophthalmologists are not able to quantify fluid volumes at nanolitre amounts. AI-based
models could provide support in objective and personalised therapy decisions in such
cases by providing precise and reproducible values ([Fig. 3]) [27].
Fig. 3 Segmentation of different biomarkers on CFP and OCT images. Automatic segmentation
is compared to manual segmentation by human specialists and lies within the range
of interreader variability.(Source: Schmidt-Erfurth U, Sadeghipour A, Gerendas BS
et al. Artificial intelligence in retina. Prog Retin Eye Res 2018; 67: 1 – 29. doi:10.1016/j.preteyeres.2018.07.004 [rerif])
GA has been undergoing a major paradigm shift since the FDA approved the first two
drugs for treating GA, which may also see approval in Europe later this year. Results
from pivotal studies based on fundus autofluorescence (FAF) measuring GA progression
have shown significantly slower growth in treated patients compared to untreated patients
[5], [6]. Using FAF to measure GA is characterised by the absence of fluorophoric metabolites
such as lipofuscin due to missing RPE [28]. Post-hoc analysis using AI-based models for automatic segmentation of RPE and photoreceptors
measured on the ellipsoid zone (EZ) has shown reliable visualisation of treatment
effects at RPE level using OCT [29], [30]. Automatic segmentation of EZ loss also revealed the treatment effect to be more
pronounced at EZ level, and therefore at the neurosensory layer rather
than at RPE level. Identifying a pathognomonic pattern in GA progression was also
possible using automatic segmentation and precise evaluation of OCT volumes. EZ loss
was shown to exceed RPE loss and precede the progression of RPE loss using AI-assisted
analysis of OCT volumes from Phase III OAKS and DERBY study data. EC loss also affects
treatment response. Data from the OAKS/DERBY trial, the largest successful Phase III
study on GA treatment to date, also showed significantly faster growth and better
treatment response in patients with high EZ-RPE loss ratios compared to patients with
low ratios [30].
Several biomarkers have been identified as associated with rapid growth in GA progression
assessment; these include: Pseudodrusen, hyperreflective foci, specific FAF patterns
in the junctional zone, and multifocal extrafoveal lesions. However, these biomarkers
have only been studied at population level so far; assessing individual risk of rapid
growth towards predicting local lesion spread has remained a challenge. Pioneering
studies have shown AI-based models to predict GA growth and localisation in en face
images from a single OCT volume [31], [32], [33]. AI-supported analysis provides groundbreaking insights into the relevant mechanisms
of GA progression and treatment response; these cannot be reliably assessed using
FAF, nor are human specialists able to detect and quantify them at clinical examination.
AI-based models have also been investigated for their ability to predict treatment
frequency and outcomes from anti-VEGF injections for nAMD. Selecting an appropriate
treatment regime and interval remains a challenge in nAMD treatment. Compared to clinical
trials, real-world results have shown undertreatment and more adverse outcomes for
vision in treatment with anti-VEGF injections; this is likely due to late diagnosis
or treatment delays on recurring exudation requiring treatment [27]. Pioneering studies have investigated the potential of AI-based algorithms to predict
treatment needs [34], [35]. One model has been trained to distinguish a priori between low, moderate, and high
therapy needs using on OCT volumes from 317 patients in a PRN regime (PRN: pro re
nata). The algorithm achieved an AUC of 70% to 77%, a 50% performance increase compared
to retina specialists [36]. An AI-based algorithm was also able to distinguish between interval extension suitability
and unsuitability in patients on a treat-and-extend (T&E) regime [37].
Functional response to anti-VEGF injection is subject to wide interindividual variability,
making it difficult for experienced ophthalmologists to predict. Even so, predicting
treatment response based on central visual acuity may lead to higher patient compliance
and motivation for ophthalmologists. Apart from that, factoring in a poor functional
response prognosis may save important human and cost-intensive resources. A DL-based
algorithm applied to 270 previously untreated patients randomised to receive ranibizumab
in a T&E regime achieved an accuracy level of 0.87 in predicting central visual acuity.
IRF and SRF volumes after the first injection have emerged as the most relevant biomarkers
for predicting central visual acuity [38].
Other Retinal Diseases and Systemic Comorbidities
Other Retinal Diseases and Systemic Comorbidities
Other areas where AI has been applied to retinal diseases include screening and prognostics
for retinopathy of prematurity [39], glaucoma [40], vascular occlusion [41], and hereditary retinal diseases [42]. Algorithms can also be applied in general screening between healthy individuals
and individuals with retinal disease [43]. However, the potential areas of AI application go beyond ophthalmology; morphology
of the retina and retinal vascularisation can also provide indications as to systemic
risk factors and comorbidities [44]. AI-based algorithms can recognise patterns from huge amounts of data in determining
age and gender based on CFP or OCT volume [45], [46]. One recent study used CFP to predict cardiovascular risk factors such as nicotine
consumption, body mass index, and HbA1c values using DL-based models [44]. The future will also see comorbidities such as Alzheimerʼs disease or multiple
sclerosis being detected by applying AI-based analysis to CFP or OCT scans.
Approved MDR-certified Algorithms
Approved MDR-certified Algorithms
Using AI-based algorithms in clinical routine will require a stringent regulatory
approval process beforehand. The Vienna Fluid Monitor and GA Monitor (RetInSight,
Vienna, Austria) developed at the Medical University of Vienna ([Fig. 4]) are two MDR-certified algorithms for use in retinal diseases; MDR refers to the
European Medical Device Regulation [47].
Fig. 4 Fluid Monitor (left) and GA Monitor (right) reports. These reports quantify different
fluid compartments in nanolitres, RPE and photoreceptor segmentation in mm², and corresponding
colour images on OCT scans showing precise future progression.(Source: https://retinsight.com/; Stand: 26.08.2024 [rerif])
Vienna Fluid Monitor was developed for automatic segmentation and quantification of
IRF, SRF, and PED; the algorithm has been validated on both study data and data from
clinical routine. Different fluid compartments are shown in colour on the OCT scans
with volumes given in nanolitres. Fluid dynamics can be observed over time, and the
algorithm can distinguish between active and inactive fluid, such as in degenerative
cysts.
GA Monitor is designed for automatic segmentation and quantification of RPE loss as
well as EZ loss representing photoreceptor status. This allows rapid and accurate
quantification of GA growth and classification of disease activity using the EZ-RPE
loss ratio.
Both algorithms use Heyex 2.6 software in its current version for the Heidelberg Spectralis
device; the algorithms are to be approved for each of the other OCT device manufacturers.
The system takes a matter of minutes to generate a report for the ophthalmologist
showing the quantified fluid volumes in nanolitres or atrophy segmentations in mm2 including corresponding images. Tools such as Fluid Monitor (FM) and GA Monitor (GM)
generate complete reports from the cloud at the click of a mouse with en face lesion
localisation and new quantifications, that is, any increase or decrease of activity,
from a standard OCT device (Spectralis) for the ophthalmologist to make a precise
diagnosis and decide on suitable treatment there and then at patient consultation.
The FDA currently considers EC loss to be the primary relevant biomarker for assessing
photoreceptor status in GA. This biomarker provides a reliable measure of activity
and therapeutic response for GA management
in clinical routine similar to the “fluid” biomarker in nAMD and is set to play a
crucial role in recommending GA treatment for ophthalmological patients.
Other MDR-based CE-certified systems for retinal disease management using OCT comprise
RetinAI Discovery (image and data management platform, RetinAI Medical AG) [48], iPredict (screening for AMD, DR, glaucoma using OCT and CFP, Ihealthservices Inc.)
[49], and RetinaLyze (screening for DR, AMD, and glaucoma, RetinaLyze System A/S) [50]. RetInSight GA Monitor is the only MDR-certified algorithm to display photoreceptor
loss as measured by EZ in addition to RPE loss.
Continuous development and integration of AI-based models in clinical routine will
enable efficient, accurate, and objective processing for the huge amount of retinal
imaging information available. These models provide new insights into pathomechanisms
and treatment response in clinical studies while also enabling thorough and methodical
evaluation of large databases. AI systems used as clinical decision support tools
allow objective, reproducible, and most importantly, individual treatment decisions,
initiating a paradigm shift in precision healthcare in ophthalmology.