Rofo 2024; 196(11): 1115-1124
DOI: 10.1055/a-2271-0799
Review

The radiologist as a physician – artificial intelligence as a way to overcome tension between the patient, technology, and referring physicians – a narrative review

Article in several languages: English | deutsch
Christoph Alexander Stueckle
1   Clinical Radiology, University Witten Herdecke Faculty of Health, Witten, Germany
,
Patrick Haage
2   Diagnostic and Interventional Radiology, HELIOS Universitätsklinikum Wuppertal, Germany
› Author Affiliations
 

Abstract

Background Large volumes of data increasing over time lead to a shortage of radiologistsʼ time. The use of systems based on artificial intelligence (AI) offers opportunities to relieve the burden on radiologists. The AI systems are usually optimized for a radiological area. Radiologists must understand the basic features of its technical function in order to be able to assess the weaknesses and possible errors of the system and use the strengths of the system. This “explainability” creates trust in an AI system and shows its limits.

Method Based on an expanded Medline search for the key words “radiology, artificial intelligence, referring physician interaction, patient interaction, job satisfaction, communication of findings, expectations”, subjective additional relevant articles were considered for this narrative review.

Results The use of AI is well advanced, especially in radiology. The programmer should provide the radiologist with clear explanations as to how the system works. All systems on the market have strengths and weaknesses. Some of the optimizations are unintentionally specific, as they are often adapted too precisely to a certain environment that often does not exist in practice – this is known as “overfitting”. It should also be noted that there are specific weak points in the systems, so-called “adversarial examples”, which lead to fatal misdiagnoses by the AI even though these cannot be visually distinguished from an unremarkable finding by the radiologist. The user must know which diseases the system is trained for, which organ systems are recognized and taken into account by the AI, and, accordingly, which are not properly assessed. This means that the user can and must critically review the results and adjust the findings if necessary. Correctly applied AI can result in a time savings for the radiologist. If he knows how the system works, he only has to spend a short amount of time checking the results. The time saved can be used for communication with patients and referring physicians and thus contribute to higher job satisfaction.

Conclusion Radiology is a constantly evolving specialty with enormous responsibility, as radiologists often make the diagnosis to be treated. AI-supported systems should be used consistently to provide relief and support. Radiologists need to know the strengths, weaknesses, and areas of application of these AI systems in order to save time. The time gained can be used for communication with patients and referring physicians.

Key Points

  • Explainable AI systems help to improve workflow and to save time.

  • The physician must critically review AI results, under consideration of the limitations of the AI.

  • The AI system will only provide useful results if it has been adapted to the data type and data origin.

  • The communicating radiologist interested in the patient is important for the visibility of the discipline.

Citation Format

  • Stueckle CA, Haage P. The radiologist as a physician – artificial intelligence as a way to overcome tension between the patient, technology, and referring physicians – a narrative review. Fortschr Röntgenstr 2024; 196: 1115 – 1123


#

Background

Radiology is an interface discipline. The main areas of responsibility include the analysis of images and imaging-guided treatment of certain diseases.

As a technical discipline, radiology is continuously undergoing further development. As a result of these further developments, the number of available images is increasing while scan times are decreasing. Many findings are determined in compliance with defined standards. Depending on the type of disease, these are based on scans acquired in defined planes and locations. Radiological diagnosis and intervention are thus increasingly reproducible and less susceptible to error. The continuously increasing number of images and the increasing demand for interpretation mean a greater workload for radiologists.


#

Development of radiology

As a technology-based discipline, there have been many developments in radiology since the discovery of X-rays by Conrad Röntgen in 1895. In particular, the introduction of computed tomography (CT) and magnetic resonance imaging (MRI) were major milestones that changed radiology. The first CT scanners at the start of the 1970 s provided individual images with slice thicknesses of more than 4 cm. The rotation time has become shorter, slice thickness has become smaller, and scanners have become faster. At the start of the CT era, gaps were left in the scan volume in order to ensure sufficient cooling of CT scanners and to save time [1]. As a result of the introduction of spiral CT and subsequently multidetector spiral CT and volume CT, thin-slice 3D datasets have been increasingly acquired ([Fig. 1]). Instead of scan gaps, overlapping slices are acquired today. Therefore, numerous thin-slice reconstructions are the standard today in CT. They can be supplemented with specific reconstruction algorithms and thus be made available in the desired layout for viewing and interpretation. At the same time, the number of patients examined per time unit is increasing. Consequently, the number of patients to be examined per time unit and the number of images to be viewed and interpreted are continuously increasing ([Fig. 1]). This has resulted in a significant increase in the workload for radiologists. Moreover, examinations have increased not only in number but also in complexity. In addition to morphological images, functional and dynamic evaluations and diffusion maps are increasingly created. The amount of data that radiologists must process promptly, precisely, and in a targeted manner is thus further increasing. As a result of the increasing workload, greater dissatisfaction, an increase in the number of cases of burnout, and early retirement have been seen among radiologists [2]. Modern radiology is therefore currently confronted with four major challenges: large amounts of image material to be interpreted (big data), high demand for reporting and communication, a shortage of personnel, and a significant number of patients.

Zoom Image
Fig. 1 Development of slice thickness and computed images on cranial tomography over time.

#

Artificial intelligence

The greater workload has resulted in alternative approaches regarding workflow and reporting. To allow more time for communication with patients and referring physicians, AI-supported expert systems have increasingly become a topic of interest.

Due to the image-based work in radiology, it offers ideal conditions for the use of AI for evaluation [1] [2] [3] [4]. Artificial intelligence has been incorporated into radiology in stages: in the form of the first expert systems in the 1980 s, in the form of probabilistic systems in the 1990 s, and as increasingly sophisticated deep learning models since the end of the 2000 s [5]. The number of publications addressing AI-based reporting has increased accordingly [4].

The AI systems used in radiology and generally in medicine comprise two fundamental methods: AI data is generated by learning from a human being or by extracting previously unknown information [6] [7].

In radiology, artificial intelligence is primarily used in MRI (37 % of AI systems use MR datasets) followed by CT imaging (29 %), with the most common task being segmentation (39 %) [4]. Research in neuroradiology and chest radiology is currently a main topic of interest [3] [4].

Particularly in areas like oncological imaging where comparison with previous images is essential and scan results typically have to be added to a specific evaluation system, it is helpful when the preliminary work is performed by a corresponding system [8]. Therefore, AI has been implemented for many applications in the diagnosis and segmentation of pulmonary nodules and corresponding research is being conducted [4] [8] [9].


#

Black box problems

Successful use of AI has also been increasingly reported in other areas. A current review regarding the depth of myometrial invasion shows successful diagnosis of this disease using AI. The review shows that various AI systems based on different AI techniques can help to evaluate the depth of myometrial invasion. It also shows limitations with respect to the AI systems and the evaluating radiologists. It is often unclear how an AI system reaches its results [10].

For this reason, “explainable AI” is often promoted and requested. This means that the AI system and its results should be able to be explained.

The term explainable AI refers to a series of processes and methods that allow human users to understand and trust results and output generated by machine learning algorithms. Explainable AI is used to describe an AI model, its expected effects, and potential inaccuracies. It helps to characterize model accuracy, correctness, transparency, and results during the AI-supported decision-making process. Radiologists who regularly use AI applications to optimize their workflow must understand how to achieve results that will save time. Explainable AI is extremely important for creating trust among physicians and patients when AI models are used to help make medical decisions.

The more advanced the AI system, the more difficult it is for human beings to understand how the algorithm arrived at a particular result. The entire calculation process becomes a black box that can no longer be interpreted. These black box models are created directly from the data. Not even the software engineers and data scientists who developed the algorithm can understand or explain exactly what is happening or how the AI algorithm arrived at a certain result.


#

Explainable AI

There are many advantages to the user being able to at least partially understand how an AI-supported system arrived at a certain result.

Image processing AI systems often use the data enrichment technique. This means that image data are modified in many ways in preparation to be analyzed by the neural network. Classic steps for such enrichment are geometric changes, scaling of the region of interest, intentional addition of Gaussian noise, contrast enhancement, gradual potentiation of image data, insertion of Gaussian blurring, and mathematical pruning of the dataset. These mathematical processes take place prior to the actual analysis of the dataset in the neural network. These mathematical models, which are adapted in a complex manner to the relevant task, are often not understandable for the user [11].

Explainability can help developers to ensure that the system functions as expected. It can be necessary to meet regulatory standards or it can be important to allow those affected by a decision to refute or change the result [12].

Explainable AI should follow basic principles to ensure trust between AI and human beings. The US Department of Commerce created an overview:

  • Explanation: The AI system provides or contains accompanying documents or reasons for results and/or processes.

  • Meaningful: The AI system provides explanations that are understandable for the intended consumer.

  • Explanation accuracy: The explanation is adapted specifically to the displayed result. The explanation correctly reflects the reason for generating the output and/or accurately reflects the system’s process.

  • Knowledge limits: A system only operates under conditions for which it was designed and when it reaches sufficient confidence in its output [13].


#

Explainable AI: Sources of error, risks, and subsequent adaptation

A current review from the year 2022 examined which explainability methods were used in radiology studies for the application of AI. The review came to the conclusion that explainability was achieved in 49 % of studies by providing cases/examples. No explainability was offered in 28 % of studies, visualizations and saliency maps were offered as explanations in 18 %, and the results were discussed retrospectively in 5 % [4]. In this context, examples, i. e., image datasets, coded according to the disease are reviewed on the basis of coded sample datasets or test image datasets. As a critical point, the study states that some software uses image datasets from only one hospital and only very small datasets were used in some cases. Using visualization tools, the AI software can show the developer which features in the image or dataset were used to make the primary decision. The maps show corresponding foci that the software used for orientation [4].

The systems and especially primary system testing and the corresponding adaptation of the AI systems during the software training phase typically ensure that the corresponding software functions reasonably in a narrow application field, i. e., in the framework of the learned parameters [14] [15].

To achieve verifiable and highly reliable AI results, labeled datasets are often used to train AI systems. Only image data that has been checked by a human expert and provides a clear result should be used. AI is supplied with the greatest possible amount of such data. The test is then performed – also with a reviewed dataset – and the system results to be expected are positively validated with a high probability. Specific training for a particular use case can result in overfitting of the neural network and thus ultimately in an overoptimistic expectation of the model. Overfitting occurs when an AI system learns to make predictions based on image features that are specific to the training dataset and cannot be generalized to new data.

This can then result in failure of the model in the case of datasets from other hospitals or practices. One example of overoptimization is a prevalence of certain diseases on one scanner. If for organizational reasons the majority of severely ill patients are examined on one scanner, e. g., on “CT1”, but there are additional CT units that are not used for examining this special group of patients, the AI erroneously learns that there is an increased probability for a serious disease based solely on the fact that the examination is performed on this specific scanner (CT1). If the same software is used under other conditions, the factor included in the assessment (CT1) is omitted yielding completely different results [16].

One method to avoid this overoptimization is cross-validation: a sampling procedure for repeated classification of a dataset into independent cohorts for training and testing. The separation of training and test datasets ensures that performance measurements are not distorted by direct overfitting of the model to the data. During cross-validation, the dataset is divided multiple times, the model is trained and evaluated with one subgroup in each case, and the prediction error is averaged over the test runs. The use of cross-validation allows estimation of the generalization performance of an algorithm, determination of the most suitable algorithm from multiple algorithm systems, and adjustment of model hyperparameters, i. e., fine tuning of the settings in the algorithm used to configure and train the model [7] [10].

A further method for targeted and effective selection of important radiological features within an AI algorithm is the “Least Absolute Shrinkage and Selection Operator” (LASSO), which modifies the standard regression methods in that it is limited to a certain subset of all available covariates [5] [17]. Predictor variables that can contribute to overfitting are thus removed. Via manual human-controlled segmentation of lesions in a training environment, the reproducibility of feature detection is determined by various human operators. Non-reproducible features are thus rejected [5].


#

Optimization of AI and minimization of potential weaknesses

A current study on the underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations shows the need for critical examination of AI in radiology. This result has been given significant attention and shows the limitations of the technique. The authors of the study [18] were able to show that classifiers created with the latest computer vision techniques consistently and selectively underdiagnosed underserved patient groups and that the rate of underdiagnosis in these groups, e. g., Hispanic patients, was also significantly higher without the use of AI-based systems. This means that a group that is already underserved in the current medical system is also underserved by AI-supported systems, thus showing how important it is to continuously review and examine the algorithms [18].

Software systems that test AI systems for errors and correctness of explanation models to prevent such errors are now being developed [19] [20].

AI-based reporting systems undoubtedly support reporting [2] [5]. Many AI-based systems are used in chest radiography. For example, they achieve impressive accuracy between 0.935 and 0.978 in the diagnosis of pneumothoraces [21] [22] [23]. Many AI approaches with very high diagnostic accuracy were also introduced for diagnosing COVID [24] [25] [26].

In the ideal case, AI systems should be trained and validated for the entire spectrum of possible diseases in datasets of varying quality within a certain examination modality. However, this is not yet possible due to the high variability in real clinical situations [7] [27]. Therefore, AI systems are currently only designed for a specific application and are limited to this application.

As a current development, in addition to deep learning and deep learning networks, radiomics is a topic of interest as a future technique that may provide additional advantages with respect to reporting. After segmentation of the corresponding morphological correlate, e. g., “pulmonary nodule”, AI systems use an assessment cascade containing as many learned features as possible to evaluate the detected pulmonary nodule [9].

AI studies primarily from basic research with promising results often show only limited potential for generalizability and immediate clinical implementation. There is a high risk of distortion, particularly due to the lack of external validation. Moreover, a clear and understandable explanation of how the system works and which limitations need to be taken into consideration is missing in most cases [4] [15].

In unfavorable situations, even minor changes to input data that are often invisible to the human eye can result in dramatically different classifications [5]. This different and in the worst case scenario incorrect classification results from the fact that complex neural networks also overemphasize certain features. Therefore, for example, different body types outside the norm are a problem for AI. One study from the year 2023 examined the possibility of automated volumetric analysis of the abdominal wall musculature. In this study, the AI system unsuccessfully performed automated muscle volumetric analysis in one patient with well-defined abdominal muscles apparently because the system had been trained on a certain ratio of fat to muscle. The muscles that were clearly visible to the radiologist were not correctly detected by the AI system because the subcutaneous fat tissue that is otherwise typical and was apparently always present in AI training was barely visible [28].

In connection with incorrect classification, one study shows that an imperceptible, non-random perturbation of an image to be evaluated can arbitrarily change the prediction of the neural network in spite of sufficient training. The reason for the error is complex. A sufficiently trained neural network is robust with respect to minor perturbations in the entered image dataset. A minor perturbation cannot change the object category of an image. However, there are regions within the detection matrix that result in a significant deviation in results in the network. If the perturbation or deviation is present in this image region, a maximum prediction error occurs. In information technology, the term “adversarial examples” is used in this case.

These adversarial examples are relatively robust – even if the neural network was trained with various subsets of the training data. This means that the neural network is specifically susceptible to a discrete case in the data to be analyzed and that especially “deep layer” networks that learned by means of backpropagation intrinsically have blind spots. Interestingly, the specific nature of these perturbations is not a random artifact of the learning: The same perturbation can cause a different network that was trained on another subset of the dataset to classify the same input incorrectly [16] [29].

The best use of AI in medicine is as a reliable assistant requiring supervision ([Fig. 2]). The AI system ideally indicates possible errors in the relevant analysis. One possible method is for the AI system to provide reliability intervals so that the medical expert can determine when a closer look is warranted [30].

Zoom Image
Fig. 2 The figure shows how AI can be used effectively and safely. The greater the seriousness of a diagnosis, the greater the necessity for the diagnosis to be reviewed and ultimately made by a physician. If the diagnostic work to be performed is of minor immediate importance for the patient, greater trust can be placed in the AI system.

The use of AI for assisted reporting in radiology is clearly the future, particularly when radiologists successfully use the strengths of the technology and know and avoid the weaknesses. According to Curtis P. Langlot: “‚Will AI replace radiologists?‘ is the wrong question. The right answer is: Radiologists who use AI will replace radiologists who don’t.“ [2].

The use of AI systems can already begin today in patient management and appointment management, can continue to be used in reporting, and can provide significant support in the scientific evaluation of acquired data in order to save time and resources ([Fig. 3]).

Zoom Image
Fig. 3 Overview of the various application areas and uses of AI in radiology.

Thus, in the future, radiologists will ideally be able to use multiple AI-supported systems to ease their workload in various ways in the daily routine at the hospital/practice.


#

Radiology in communication

The systematic use of all AI-assisted reporting options will save radiologists a significant amount of time ([Fig. 4]). Radiologists can and should use this for patients and referring colleagues.

Zoom Image
Fig. 4 Diagram of workload in radiology. Various factors result in a continuous increase in the workload of radiologists. Targeted use of AI can save time. How the time that is gained is ultimately used can be decided on an individual basis.

There are significant differences between radiologists working in a practice/health care center and those working at a hospital. In most hospitals and clinics, the radiology department provides findings in writing, possibly combined with a clinical discussion, and the treating physician communicates with the patient. This concept is already established and absolutely desired by clinical colleagues [31]. It must be stated that interdisciplinary case discussions in hospitals improve the assessment of a patient’s clinical picture as a result of interactive communication which benefits patients, clinicians, and radiologists. This shows how important it is to invest the time gained by the use of AI in communication [32] [33] [34].

If specialist training in radiology takes place in a practice or at a health care center, findings are often still communicated to the patient directly by the physician. Patients want this communication and demand relatively little to be satisfied with the doctor-patient interaction [35]. A lack of time and an excessive workload in recent years have resulted in this direct communication of findings being less common – resulting in dissatisfaction on the part of both patients and physicians [35] [36] [37].

Radiologists complain that the ability to adequately speak with patients is insufficient and professional visibility is also often insufficient [38]. A non-representative patient questionnaire performed as part of one of our studies showed that 71 % of 386 surveyed patients reported that they did not have an opportunity to discuss the examination with the radiologist. This trend has also been confirmed by a study by the RSNA. In a large survey among members of the RSNA, 73 % stated that they do not have enough time to speak with patients due to workload and work density [39].

A well-written case history by a well-known radiologist shows that radiology can indeed be part of the clinical concept in patient diagnostics when small but significant things in the patient history that can often diagnose a complex clinical picture are taken into consideration when communicating with the patient [40].

The situation is slightly different in the case of oncology patients. It is often virtually impossible for radiologists to provide information about further treatment or in the case of disease progression about a change in treatment since oncological treatments are highly complex. It should ideally be clarified in advance with the patient and the referring oncologist that the referring oncologist will discuss findings with the patient.

In other cases, e. g., after trauma and corresponding diagnosis of exclusion, patients and also treating colleagues appreciate having a brief discussion with the radiologist about the disease and treatment to be expected. In addition, immediate communication of findings significantly shortens the time to treatment for the patient or in the case of diagnosis of exclusion the patient can immediately resume usual activities [41].

A rarely considered secondary effect is that the physician can have a positive effect on upcoming treatment as a result of expectation effects. Such side effect-free treatment effects should be increasingly considered and implemented in radiology and treating radiology [42]. With respect to communication and the communication of findings, radiologists can use expectation effects to create pretherapeutic expectations that can have a positive effect on upcoming treatment [43] [44].

An initial consultation with a radiologist can also relieve some of the burden on the health care system. For example, informing patients of the low-risk nature of their disease can prevent them from seeking care from another discipline. This requires time and background knowledge [45].

Referring physicians have clear demands regarding radiology: Diagnostic reports should be understandable and address the particular medical issue. In addition, quick communication of findings is desired. However, radiologists have stated that the requested examination method is sometimes selected incorrectly and the medical question is often not formulated precisely enough [46]. This shows that further intensive work on communication is needed on both sides. Personal contact should be established where appropriate or work shadowing should even be performed in order to optimize collaboration with respect to patients. With the systematic use of AI as a reporting tool, a time savings can be achieved resulting in better patient-oriented collaboration possibilities ([Fig. 4] and [Tab. 1]).

Table 1

Explainable AI: A practical example.

Multiparametric MRI of the prostate:

AI

User reviews the results

User makes changes as needed

Detect prostate

Prostate fully examined

Outer margins of the prostate

Divide prostate into zones

Zone categorization correct

Zone classification

Mark suspicious masses (PIRADS 3–5)

Masses detected correctly

Mark additional masses, remove incorrectly marked masses

Masses are evaluated depending on zone and size

Evaluation correct? PIRADS criteria correctly applied?

Remove incorrectly marked lesions, change PIRADS category suggested by AI

Overall evaluation according to PIRADS

PIRADS category correct?

PIRADS category adjusted

Additional findings, AI typically not designed to detect pathologies outside of the prostate

Evaluation of the structures included in the dataset

Diagnosis according to PIRADS V2.1:

In practice, AI often generates a prostate marked in color as the first step. The prostate is marked as an organ. This is the first item that the user needs to review. The AI system then marks suspicious lesions and groups them according to signal behavior in the relevant PIRADS category.

Since different evaluation criteria are used in the peripheral zone and the transition zone, it is important for the user to take a close look at the lesion and its location. Is the location really the peripheral zone and is DWI thus the main criterion or is the lesion in the transition zone and is the morphology in T2 weighting the main criterion and does the signal behavior in DWI need to be viewed on a secondary basis in this case?

The user must therefore observe and review several criteria. Software with good programming is user-friendly. With just minimal practice, it is possible to review and release or change results with a couple of careful clicks.

A practical example:

AI: Prostate is segmented.

Physician: OK.

AI: No suspicious lesions in the prostate marked.

Physician detects only nodules typical for adenoma: OK.

Additional detection of a mass measuring approximately 12 × 14 × 12 mm in the rectosigmoid junction.

AI: PIRADS: II.

Physician: PIRADS II. Additionally – and very importantly for the patient in this case: Suspicion of infiltrative tumorous process in the rectosigmoid junction.


#

Use of time gained as a result of AI

In the ideal case, the use of AI-supported radiology systems saves time. This time can be used in different ways. There is the risk in our health care system that, for economic reasons, the time gained by using AI will not be invested in communication but rather will be seen as an opportunity to further increase the number of patients examined per time unit. A modification of compensation would be one possibility to make doctor-patient communication more attractive. However, there is still the risk here that the time will only partly be used for communication so that the rest of the time can be invested in further increasing the number of examinations. Every radiologist should ultimately decide for themselves how to use the time gained as a result of the use of AI. Communication with referring physicians and patients is certainly desirable but it is not the only possibility.


#

Summary

Considering the numerous examinations, patients, reports, and referring physicians, radiology must above all be reliable, safe, and communicative. The systematic use of AI-supported systems helps radiologists to save time. AI must be implemented correctly and in a targeted manner. Radiologists must be familiar with the strengths and weaknesses of the AI system being used in order to optimally lighten the radiologist’s workload. Since current AI systems are optimized for a narrow field of activity, multiple systems need to be used for the best results.

Under optimal conditions, the use of AI systems results in a time savings for radiologists. The additional time can be used in various ways. More patients can be examined, examinations can be more comprehensive, or the time can be used for interaction.

In my opinion, the time gained as a result of the use of AI should be used for targeted communication with patients and referring colleagues. A targeted exchange results in better treatment of patients and higher satisfaction among radiologists.

Explainable artificial intelligence is the future of radiology. It will require human supervision, will save time, and will improve diagnosis.

The popular narrative of “device medicine” could change to “talking medicine”. Radiologist would become a serious clinical partner – at least a chance.


#
#

Correspondence

PD Dr. Christoph Alexander Stueckle
Radiology, Marienhospital Osnabrück
Bischhofsstraße 8
49074 Osnabrück
Germany   
Phone: +49/5 41/3 26 44 52   
Fax: +49/5 41/3 26 44 68   

Publication History

Received: 26 July 2023

Accepted after revision: 27 January 2024

Article published online:
03 April 2024

© 2024. Thieme. All rights reserved.

Georg Thieme Verlag KG
Rüdigerstraße 14, 70469 Stuttgart, Germany


Zoom Image
Fig. 1 Development of slice thickness and computed images on cranial tomography over time.
Zoom Image
Fig. 2 The figure shows how AI can be used effectively and safely. The greater the seriousness of a diagnosis, the greater the necessity for the diagnosis to be reviewed and ultimately made by a physician. If the diagnostic work to be performed is of minor immediate importance for the patient, greater trust can be placed in the AI system.
Zoom Image
Fig. 3 Overview of the various application areas and uses of AI in radiology.
Zoom Image
Fig. 4 Diagram of workload in radiology. Various factors result in a continuous increase in the workload of radiologists. Targeted use of AI can save time. How the time that is gained is ultimately used can be decided on an individual basis.
Zoom Image
Abb. 1 Dargestellt ist die Entwicklung der Schichtdicke und der errechneten Bilder in der kranialen Computertomografie über die Zeit.
Zoom Image
Abb. 2 Die Illustration zeigt, wie KI sinnvoll und sicher genutzt werden kann. Je gravierender die Auswirkung der Diagnose für den Patienten ist, umso erforderlicher ist es, dass die Diagnose ärztlich kontrolliert und gestellt wird. Ist die auszuführende diagnostische Arbeit von nur minderer unmittelbarer Bedeutung für den Patienten, kann umso mehr der KI vertraut werden.
Zoom Image
Abb. 3 Übersichtsgrafik über die verschiedenen Einsatzbereiche und Anwendungen der KI in der Radiologie.
Zoom Image
Abb. 4 Skizze der Arbeitsverdichtung in der Radiologie. Verschiedene Faktoren erhöhen beständig die Arbeitsbelastung des Radiologen. Durch gezielten Einsatz von KI kann mehr Zeit gewonnen werden. Wie die gewonnene Zeit letztlich genutzt wird, bleibt jedem selbst überlassen.