Subscribe to RSS

DOI: 10.1055/s-0044-1800723
Searching for Value Sensitive Design in Applied Health AI: A Narrative Review
Summary
Objective: Recent advances in the implementation of healthcare artificial intelligence (AI) have drawn attention toward design methods to address the impacts on workflow. Lesser known than human-centered design, Value Sensitive Design (VSD) is an established framework integrating values into conceptual, technical, and empirical investigations of technology. We sought to study the current state of the literature intersecting elements of VSD with practical applications of healthcare AI.
Methods: Using a modified VSD framework attentive to AI-specific values, we conducted a narrative review informed by PRISMA guidelines and assessed VSD elements across design and implementation case studies.
Results: Our search produced 819 articles that went through multiple rounds of review. Nine studies qualified for full-text review. Most of the studies focused on values for the individual or professional practice such as trust and autonomy. Attention to organizational (e.g., stewardship, employee well-being) and societal (e.g., equity, justice) values was lacking. Studies were primarily from the U.S. and Western Europe.
Conclusion: Future design studies might better incorporate components of VSD by considering larger domains, organizational and societal, in value identification and to bridge to design processes that are not just human-centered but value sensitive. The small number of heterogeneous studies underlines the importance of broader studies of elements of VSD to inform healthcare AI in practice.
#
Keywords
Value-sensitive design - Artificial Intelligence - Ethical value – Universal Design - Healthcare1. Introduction
Value Sensitive Design (VSD) represents a comprehensive approach to embed human values in the creation of technology. A topical and timely area of healthcare innovation, artificial intelligence (AI) describes the use of computers and technology to simulate intelligent behavior comparable to human critical thinking [[1]]. Public perception of recent advances in AI and widespread availability of AI-based tools have catalyzed interest in their clinical application [[2] [3] [4] [5] [6]]. Current applications of AI in healthcare include medical imaging and diagnostics, personalized medicine, virtual health assistants and chatbots, and predictive analytics and risk assessment [[7]]. These applications rely on computational methods ranging from machine learning (ML), natural language processing (NLP), computer vision (CV), and generative models such as large language models (LLMs).
Healthcare AI has been a mainstay of predictive and preventive clinical decision support (CDS) for decades and has been the subject of rigorous evaluation [[9]]. Scalable algorithmic tools present the promise of dramatically improving care while also risking the perils of harm at scale [[10]]. Creating and deploying AI without robust ethical frameworks and rigorous testing might perpetuate bias, discrimination, and introduce privacy, security risks [[11]]. Regardless of the nature of underlying AI, generative or not, actionable healthcare AI requires careful integration of algorithmic insights into clinical workflow, which can be achieved through effective design that engages patients and healthcare professionals, including clinicians, nurses, and allied health staff.
Numerous views of design have emerged over the last fifty years [[12]]. Design disciplines emanate from different industries and areas, such as human factors and interaction design linked to experimental psychology, and design thinking which is rooted in the science of creativity [[12]] More recently in healthcare, human-centered design (HCD) has emerged as a means of achieving effective design and lessening potential for risk [[13] [14] [15]]. In parallel, concepts like algorithmovigilance for healthcare AI have broadened the discussion around preventing adverse effects of implemented AI [[16]]. VSD intends to place human values at the heart of technological design. Batya Friedman et al., first introduced VSD in the 1990s, challenging the prevailing notion of technology as value-neutral or simply a tool for human activities. Per Friedman et al., values are defined as “what is important to people in their lives, with a focus on ethics and morality” [[13]]. Commonly recognized examples of such values include autonomy, trust, equity, and justice. The VSD framework has been applied across various fields and has become a recognized approach for addressing values in technological design [[17]]. While sharing some similarities with HCD in terms of considering users' needs and experiences, VSD might complement HCD by considering social and ethical implications of technology and impacts on various stakeholders and society at large. The framework emphasizes three main areas of investigation: conceptual, empirical, and technical. Each area builds on the prior, with conceptual investigations articulating values and any value conflicts that might result from technology use. Empirical investigations apply qualitative and quantitative methods to crystallize and document stakeholders' values, needs, and practices. Finally, technical investigations focus on the design and features of the technology itself, examining how values are or can be supported.
We emphasize the importance of applying VSD to healthcare AI because of the potentially rapid development cycle, scalable nature, lack of knowledge among healthcare managers about how it works, and its potential societal impact. The accelerated pace of AI advancement has prompted ongoing debate concerning oversight and regulation. Government entities, professional societies, scientific leadership groups, and other organizations worldwide have begun issuing principles and guidance to attempt to outline and mitigate large-scale societal risks [[18] [19] [20] [21]]. Such concerns underline the need to characterize the societal values that should be represented in the legislation and professional standards that will undoubtedly emerge. Similarly, while HCD incorporates numerous strengths in the pursuit of effective AI tools, a recent systematic review suggests key sociotechnical gaps remain: technical limitations (e.g., predicted AI output will be inaccurate), workflow misalignment (e.g., AI tools will add to workload), attitudinal barriers (e.g., resistance to using AI tools, lack of desire to use AI tools), informational barriers (e.g., AI output lacks actionable information to make the relevant clinical decision), usability issues (e.g., lengthy ‘click-throughs’ as inputs to use AI tools), and environmental barriers (e.g., requirements of AI tools such as high quality retinal imagery not available in routine practice) [[9]]. Many of these gaps might be lessened by design that centers not just on the performance of the humans interacting with the technology, but also on their values and the values of other relevant stakeholders.
To better understand how well values are captured in modern healthcare AI, we conducted a narrative review of the literature intersecting VSD, AI, and the design and implementation of AI-driven healthcare technology. We assessed recent literature, 2022 through 2023 against Friedman's tripartite framework including modifications proposed by Umbrello et al., incorporating Artificial Intelligence for Social Good (AI4SG) [[17],[22]]. We then evaluated all eligible studies for the presence of elements of VSD, critically doing so regardless of whether VSD was directly acknowledged, as it is a newer and lesser-known concept in the field than HCD. By demonstrating and dissecting VSD methodology through real-world case studies, we aim to highlight areas of strength and potential directions for improvement in the design of AI-driven technology that is both human-centered and value-sensitive.
#
2. Methods
We conducted a narrative review informed by the PRISMA guidelines for systematic reviews [[23]] with expectation that the state of the literature would not permit meta-analysis. We searched PubMed, using a search query representing AI, design, and value. The PubMed search query is shown (Box 1).
(value sensitive design[tiab] OR value-sensitive design OR value[tiab] OR social value[tiab] OR social good[tiab] OR ethical value[tiab] ) AND (“Universal Design”[Mesh] OR universal design[tiab] OR design[tiab] OR human-centered design[tiab] OR “User-Centered Design”[Mesh] OR user-centered[tiab] OR user-centred[tiab] OR user-centric[tiab] ) AND (“Artificial Intelligence”[Mesh] OR artificial intelligence[tiab] OR AI [tiab] OR intelligent design[tiab] OR intelligence design[tiab] OR machine intelligence[tiab] OR computational intelligence[tiab] OR “Machine Learning”[Mesh] OR machine learning[tiab] OR learning machine[tiab] OR “Deep Learning”[Mesh] OR deep learning[tiab] OR “Natural Language Processing”[Mesh] OR natural language processing[tiab]) AND (2022:2023[pdat])
The initial list of publications was reviewed via title and abstract review by two authors (Y.L., L.N.) to establish a full-text candidate list. The authors reviewed initial results to eliminate papers that did not address a healthcare area and/or were not English-language. Systematic reviews were also excluded though they informed study design and background as above. Retracted studies identified by initial search were excluded. The final group of studies was reviewed for elements of VSD and the presence of a specific AI application, tabulated, and summarized with respect to clinical application area, geographic diversity, methods, stakeholder engagement methods, and VSD findings.
#
3. Results
From 2022 through January 2024, we identified 819 publications via the initial search; five of the 819 had been retracted and were excluded. The initial reviews resulted in 70 studies requiring full-text consideration and nine (9) studies being selected for full-text review and extraction. A flow diagram summarizing the search is shown ([Figure 1]).


[Table 1] summarizes all studies reviewed in full-text including study objective and key findings relevant to our survey. In the final group of studies, most were original research or case studies. Clinical application area of these studies varied widely from rare disease [[24]], dementia [[25]], sexual health [[26]], mental health [[27]], to stress management [[28]]. No two studies that met full text eligibility were in the same clinical domain. In terms of AI application area, use cases primarily described predictive model and CDS or consumer health informatics applications with mobile devices (smartphones).
3.1. Stakeholder Engagement and Methods
Unlike many AI-driven CDS studies aimed at clinician-users, studies in this survey engaged a wide range of stakeholders from the general public, employees in the workplace, patients with lived experience, clinicians, managers, and biomedical scientists. Geographic diversity was limited, however, with the studies set solely in European countries and the United States. Methods used included interviews (three studies), focus groups (two studies), participatory design workshops (two studies), surveys (two studies), and literature review (one study).
#
3.2. VSD Themes Addressed
The predominant element of VSD addressed in six of nine included studies was that of value identification. Two addressed context analysis and the same two addressed prototyping.
#
3.3. Studies incorporating methods of VSD
Cagliero et al., [[29]] explored the values of stakeholders in the context of a ML tool for advanced care planning (ACP) using a values-collision approach that describes the tensions or collisions between various stakeholders' perspectives. Stakeholders included clinicians, designers, administrators, and patients. They found collisions in the values of the stakeholders that related to model characteristics, model implementation, and intervention benefits and harms. The authors propose the values-collision approach as a useful strategy for using tensions as a source of useful input for the design team to make improvements with ethical consequences in mind.
Cenci et al., [[25]] used the VSD framework in a citizen science project focused on the design of an mHealth app for dementia patients that involved location surveillance as a feature. The investigators used focus groups to elicit values of stakeholders including citizens with dementia, their family members, clinical staff and volunteers, and leaders of the local smart city initiative. They describe the use of a procedural-deliberative method to engage a wide variety of stakeholders in the co-creation of the ultimate design. Of note, the investigators referred to participants in their study as “citizens” and not patients.
#
3.4. Studies identifying trust as a core value
Götzl et al., [[27]] used focus groups, interviews, and a survey to examine the perspectives of key stakeholders of a mobile mental health app. The investigators documented a variety of values and related dimensions including acceptability of specific types of data sharing, perspectives on the role of an app in relation to traditional mental health providers, and the level of trust in various types of institutions.
Hallowell et al., [[24]] examined a specific domain, trust, in a study that documented the conditions under which stakeholders felt ML could be trusted in the diagnosis of rare disease. The authors interviewed geneticists, data scientists, bioinformaticians, industry representatives, and patient representatives. Participants characterized relational trust between patients and clinicians as a key element in determining the acceptability of a technology, and expressed both concerns and enthusiasm about the prospect of commercial involvement in the process.
Helenason et al., [[30]] engaged primary care physicians in a simulation and interviews that explored feasibility of AI-based CDS to detect cutaneous melanoma. The physicians expressed concerns about trusting the technology, and these concerns converged with issues related to the clinical context, specifically that the providers in this context are responsible for the effective use of clinical guidelines for a wide variety of clinical issues.
Helman et al., [[31]] engaged various clinical personnel in focus groups in the development of a graphical user interface (GUI) to display the AI-based risk scores for instability in the intensive care unit (ICU). Stakeholder values that emerged were trust, the impact of the tool on practice and decision-making of novice personnel, and weighing of AI outputs during decision-making.
Fischer et al., [[32]] interviewed Dutch obstetric professionals to gain their perspectives on the use of AI in obstetric care using three common AI scenarios: diagnosis, workflow optimization/ triage, and personalized treatment. Two key paradoxes emerged: the first was the desire to know the factors that drive predictive models while acknowledging that models exceed human capacity in pattern identification, and the second was the desire for the model to have access to as many parameters as possible for accuracy, while acknowledging the importance of contextual factors that the model could not capture.
#
3.5. Studies emphasizing organizational or societal values not addressed above
Kerr et al., [[28]] used a literature review and survey to examine employee concerns about digital stress management interventions (dSMI), including those with and without ML-based just-in-time adaptive interventions (JITAI). The literature review identified health and well-being as a central value, supporting use of dSMI in the workplace. Employees were concerned about privacy, autonomy, identity, and accountability in relation to the use of such tools.
Wang et al., [[26]] use Gibson's concept of affordances [[33]] to frame and facilitate a study of features that promote use of a sexual health chatbot among youth, and to explore actual use of the chatbot. An affordance is a feature or resource that assists users in appropriate use of an object. The authors' work resulted in evidence across fifteen affordances they deemed functional (including new ones that emerged in this study), however several of the affordances reflected societal values, such as inclusivity and accessibility. The complete list of affordances included accessibility, multimodality, nonlinearity, compellability, “queriosity” (encouraging user curiosity), editability, visibility, interactivity, customizability, trackability, scalability, localizability (enabling linkages between users and services in their local area), inclusivity, connectivity, and actionability. This study illustrates how values can be derived using existing frameworks for characterizing human interactions with technology.
#
#
4. Discussion
This narrative review evaluated the recent biomedical literature at the intersection of applied healthcare AI and VSD. We emphasized applied studies and those with design-focused research methods, which led to numerous Empirical investigations per the VSD framework. The literature on design of AI-driven technology largely addressed value identification across heterogeneous clinical domains. VSD concepts are not common topics in healthcare AI, so we assessed for elements of the framework regardless of whether the language of VSD was specifically included in the manuscript. Studies identified were conducted almost entirely in the United States and Western Europe, leaving significant opportunity to broaden similar efforts globally. In sum, we identified a small number of heterogeneous studies emphasizing values relevant to professional or individual practice (see [Figure 2]). While such values are critical to effective healthcare AI, VSD remains understudied in this highly topical domain of biomedical innovation. As such, we cannot yet recommend broad uptake of VSD principles into healthcare AI studies but do advocate for more comprehensive investigations of VSD in future work.


Regarding the specific values identified or explored in the papers, we found that they largely focused on values relevant to professional or individual practice as shown in the inner square of [Figure 2]. We found that studies such as Hallowell et al., [[24]], Helanason et al. [[30]], and Kerr et al., [[28]] focused on acceptability or usability of technology and were also positioned methodologically to identify such practice-based values related to technology. Organizational values were represented in Kerr et al. [[28]], where employee well-being was the primary driving force for the project. Societal values were less often elucidated though Wang et al., addressed some in their study of affordances in the context of an educational chatbot.
4.1. Implications for AI development and use in healthcare
AI-driven technology focused on individual or practice level values might implicate values for organizations or society. While every applied study cannot comprehensively elucidate and address all possible values, a wider lens on value identification might identify potential gaps that might widen during or after implementation. This approach also might lead to more holistic evaluation of AI-driven clinical technology in practice. For example, clinicians using text generative tools to speed clinical documentation might note impacts of such technology on their autonomy and need for trust. However, such tools also might impact the well-being of those working at the organization as well as broaching concerns around equity depending on how the models were trained, and whether training corpora were representative of the patient population at that organization.
Value identification also might reveal value tensions, which arise when multiple human values that co-exist in a given context come into opposition with one another [[17]]. Per Friedman et al. [[17]], the optimal solution aims to attend to all values intact rather than selecting one over another. Cagliero et al., [[29]]provided an illustration of value tensions and the resulting effort to address such tensions. Value tensions can manifest at multiple levels - within individuals attempting to balance competing values internally, between individuals who prioritize values differently, and between entire groups who have contradictory perspectives. As use of AI expands, so will the potential for these tensions and the need for effective means of resolving them.
#
4.2. Strengths and limitations
Strengths of this work include a novel focus on the intersection of AI and VSD. We reviewed recent literature, ensuring the results described are timely and account at least in part for the emergence in recent years of a focus on human centered design in informatics [[14]]. We broadened published definitions of VSD based in interpretations by Umbrello et al., [[22]]. We also reviewed studies for presence of VSD elements regardless of whether the authors applied VSD directly, as we hypothesized modern HCD would incorporate VSD concepts given the attention to issues such as trust, autonomy, etc., represented with variable verbiage as we found here.
Limitations of this study include an emphasis on applied AI-driven technology design in healthcare, which biased results toward Empirical VSD investigations. The focus on AI, HCD and VSD, might also have led to relevant studies being excluded from further review. Our search was limited to PubMed.
#
#
5. Conclusions
Elements of VSD have been included in empirical studies of AI-driven technology in healthcare with clear emphases on values to individual and professional practice, especially trust. As users gain trust in such technology, system designers might broaden their design approaches to glean values across organizational and societal dimensions. This attention might inform more effective tools, more comprehensive evaluation, and more impactful design.
#
#
No conflict of interest has been declared by the author(s).
-
References
- 1 Amisha, Malik P, Pathania M, Rathaur VK. Overview of artificial intelligence in medicine. J Family Med Prim Care. 2019 Jul;8(7):2328–31.
- 2 Brown LA, Benhamou K, May AM, Mu W, Berk R. Machine Learning Algorithms in Suicide Prevention: Clinician Interpretations as Barriers to Implementation. J Clin Psychiatry. 2020 Apr 21;81(3):10951.
- 3 Secinaro S, Calandra D, Secinaro A, Muthurangu V, Biancone P. The role of artificial intelligence in healthcare: a structured literature review. BMC Med Inform Decis Mak. 2021 Apr 10;21(1):125.
- 4 Jiang F, Jiang Y, Zhi H, Dong Y, Li H, Ma S, et al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol [Internet]. 2017 Dec 1 [cited 2024 Apr 29];2(4). Available from: https://svn.bmj.com/content/2/4/230
- 5 Kansagara D, Englander H, Salanitro A, Kagen D, Theobald C, Freeman M, et al. Risk Prediction Models for Hospital Readmission: A Systematic Review. JAMA. 2011 Oct 19;306(15):1688–98.
- 6 Goldstein BA, Navar AM, Pencina MJ, Ioannidis JPA. Opportunities and challenges in developing risk prediction models with electronic health records data: a systematic review. J Am Med Inform Assoc. 2017 Jan;24(1):198–208.
- 7 Turing AM. Computing Machinery and Intelligence. In: Epstein R, Roberts G, Beber G, editors. Parsing the Turing Test: Philosophical and Methodological Issues in the Quest for the Thinking Computer [Internet]. Dordrecht: Springer Netherlands; 2009 [cited 2024 Apr 29]. p. 23–65. Available from: https://doi.org/10.1007/978-1-4020-6710-5_3
- 8 Zhang P, Kamel Boulos MN. Generative AI in Medicine and Healthcare: Promises, Opportunities and Challenges. Future Internet. 2023 Sep;15(9):286.
- 9 Wang L, Zhang Z, Wang D, Cao W, Zhou X, Zhang P, et al. Human-centered design and evaluation of AI-empowered clinical decision support systems: a systematic review. Frontiers in Computer Science [Internet]. 2023 [cited 2024 Jan 9];5. Available from: https://www.frontiersin.org/articles/10.3389/fcomp.2023.1187299
- 10 Matheny M, Israni ST, Ahmed M, Whicher D. Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril.
- 11 Kiener M. Artificial intelligence in medicine and the disclosure of risks. AI & Soc. 2021 Sep 1;36(3):705–13.
- 12 Auernhammer J, Zallio M, Domingo L, Leifer L. Facets of Human-Centered Design: The Evolution of Designing by, with, and for People. In: Meinel C, Leifer L, editors. Design Thinking Research : Achieving Real Innovation [Internet]. Cham: Springer International Publishing; 2022 [cited 2024 Feb 14]. p. 227–45. (Understanding Innovation). Available from: https://doi.org/10.1007/978-3-031-09297-8_12
- 13 Salwei ME, Hoonakker PL, Pulia M, Wiegmann D, Patterson BW, Carayon P. Retrospective analysis of the human-centered design process used to develop a clinical decision support in the emergency department: PE Dx Study Part 2. Human Factors in Healthcare. 2023;4:100055.
- 14 Chen Y, Clayton EW, Novak LL, Anders S, Malin B. Human-Centered Design to Address Biases in Artificial Intelligence. J Med Internet Res. 2023 Mar 24;25:e43251.
- 15 Shneiderman B. Human-Centered Artificial Intelligence: Reliable, Safe & Trustworthy. International Journal of Human–Computer Interaction. 2020 Apr 2;36(6):495–504.
- 16 Embi PJ. Algorithmovigilance—Advancing Methods to Analyze and Monitor Artificial Intelligence–Driven Health Care for Effectiveness and Equity. JAMA Network Open. 2021 Apr 15;4(4):e214622.
- 17 Friedman B, Hendry DG. Value Sensitive Design: Shaping Technology with Moral Imagination. MIT Press; 2019. 258 p.
- 18 The White House [Internet]. 2023 [cited 2024 Feb 20]. OMB Releases Implementation Guidance Following President Biden's Executive Order on Artificial Intelligence | OMB. Available from: https://www.whitehouse.gov/omb/briefing-room/2023/11/01/omb-releases-implementation-guidance-following-president-bidens-executive-order-on-artificial-intelligence/
- 19 Commissioner O of the. Focus Area: Artificial Intelligence. FDA [Internet]. 2022 Sep 6 [cited 2024 Feb 20]; Available from: https://www.fda.gov/science-research/focus-areas-regulatory-science-report/focus-area-artificial-intelligence
- 20 Matheny ME, Whicher D, Thadaney Israni S. Artificial Intelligence in Health Care: A Report From the National Academy of Medicine. JAMA. 2020 Feb 11;323(6):509–10.
- 21 Commission launches AI innovation package to support Artificial Intelligence startups and SMEs | Shaping Europe's digital future [Internet]. 2024 [cited 2024 Feb 20]. Available from: https://digital-strategy.ec.europa.eu/en/news/commission-launches-ai-innovation-package-support-artificial-intelligence-startups-and-smes
- 22 Umbrello S, van de Poel I. Mapping value sensitive design onto AI for social good principles. AI Ethics. 2021 Aug 1;1(3):283–96.
- 23 Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021 Mar 29;372:n71.
- 24 Hallowell N, Badger S, Sauerbrei A, Nellåker C, Kerasidou A. “I don't think people are ready to trust these algorithms at face value”: trust and the use of machine learning algorithms in the diagnosis of rare disease. BMC Medical Ethics. 2022 Nov 16;23(1):112.
- 25 Cenci A, Ilskov SJ, Andersen NS, Chiarandini M. The participatory value-sensitive design (VSD) of a mHealth app targeting citizens with dementia in a Danish municipality. AI Ethics. 2023 Apr 13;1–27.
- 26 Wang H, Gupta S, Singhal A, Muttreja P, Singh S, Sharma P, et al. An Artificial Intelligence Chatbot for Young People's Sexual and Reproductive Health in India (SnehAI): Instrumental Case Study. J Med Internet Res. 2022 Jan 3;24(1):e29969.
- 27 Götzl C, Hiller S, Rauschenberg C, Schick A, Fechtelpeter J, Fischer Abaigar U, et al. Artificial intelligence-informed mobile mental health apps for young people: a mixed-methods approach on users' and stakeholders' perspectives. Child and Adolescent Psychiatry and Mental Health. 2022 Nov 17;16(1):86.
- 28 Kerr JI, Naegelin M, Benk M, V Wangenheim F, Meins E, Viganò E, et al. Investigating Employees' Concerns and Wishes Regarding Digital Stress Management Interventions With Value Sensitive Design: Mixed Methods Study. J Med Internet Res. 2023 Apr 13;25:e44131.
- 29 Cagliero D, Deuitch N, Shah N, Feudtner C, Char D. A framework to identify ethical concerns with ML-guided care workflows: a case study of mortality prediction to guide advance care planning. J Am Med Inform Assoc. 2023 Apr 19;30(5):819–27.
- 30 Helenason J, Ekström C, Falk M, Papachristou P. Exploring the feasibility of an artificial intelligence based clinical decision support system for cutaneous melanoma detection in primary care – a mixed method study. Scandinavian Journal of Primary Health Care. 2023;0(0):1–10.
- 31 Helman S, Terry MA, Pellathy T, Hravnak M, George E, Al-Zaiti S, et al. Engaging Multidisciplinary Clinical Users in the Design of an Artificial Intelligence-Powered Graphical User Interface for Intensive Care Unit Instability Decision Support. Appl Clin Inform. 2023 Aug;14(4):789–802.
- 32 Fischer A, Rietveld A, Teunissen P, Hoogendoorn M, Bakker P. What is the future of artificial intelligence in obstetrics? A qualitative study among healthcare professionals. BMJ Open. 2023 Oct 1;13(10):e076017.
- 33 Gibson JJ. The Theory of Affordances. In: The ecological approach to visual perception. Boulder, Colorado: Taylot & Francis; 1979. p. 119–37.
Correspondence to:
Publication History
Article published online:
08 April 2025
© 2024. The Author(s). This is an open access article published by Thieme under the terms of the Creative Commons Attribution License, permitting unrestricted use, distribution, and reproduction so long as the original work is properly cited. (https://creativecommons.org/licenses/by/4.0/)
Georg Thieme Verlag KG
Rüdigerstraße 14, 70469 Stuttgart, Germany
-
References
- 1 Amisha, Malik P, Pathania M, Rathaur VK. Overview of artificial intelligence in medicine. J Family Med Prim Care. 2019 Jul;8(7):2328–31.
- 2 Brown LA, Benhamou K, May AM, Mu W, Berk R. Machine Learning Algorithms in Suicide Prevention: Clinician Interpretations as Barriers to Implementation. J Clin Psychiatry. 2020 Apr 21;81(3):10951.
- 3 Secinaro S, Calandra D, Secinaro A, Muthurangu V, Biancone P. The role of artificial intelligence in healthcare: a structured literature review. BMC Med Inform Decis Mak. 2021 Apr 10;21(1):125.
- 4 Jiang F, Jiang Y, Zhi H, Dong Y, Li H, Ma S, et al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol [Internet]. 2017 Dec 1 [cited 2024 Apr 29];2(4). Available from: https://svn.bmj.com/content/2/4/230
- 5 Kansagara D, Englander H, Salanitro A, Kagen D, Theobald C, Freeman M, et al. Risk Prediction Models for Hospital Readmission: A Systematic Review. JAMA. 2011 Oct 19;306(15):1688–98.
- 6 Goldstein BA, Navar AM, Pencina MJ, Ioannidis JPA. Opportunities and challenges in developing risk prediction models with electronic health records data: a systematic review. J Am Med Inform Assoc. 2017 Jan;24(1):198–208.
- 7 Turing AM. Computing Machinery and Intelligence. In: Epstein R, Roberts G, Beber G, editors. Parsing the Turing Test: Philosophical and Methodological Issues in the Quest for the Thinking Computer [Internet]. Dordrecht: Springer Netherlands; 2009 [cited 2024 Apr 29]. p. 23–65. Available from: https://doi.org/10.1007/978-1-4020-6710-5_3
- 8 Zhang P, Kamel Boulos MN. Generative AI in Medicine and Healthcare: Promises, Opportunities and Challenges. Future Internet. 2023 Sep;15(9):286.
- 9 Wang L, Zhang Z, Wang D, Cao W, Zhou X, Zhang P, et al. Human-centered design and evaluation of AI-empowered clinical decision support systems: a systematic review. Frontiers in Computer Science [Internet]. 2023 [cited 2024 Jan 9];5. Available from: https://www.frontiersin.org/articles/10.3389/fcomp.2023.1187299
- 10 Matheny M, Israni ST, Ahmed M, Whicher D. Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril.
- 11 Kiener M. Artificial intelligence in medicine and the disclosure of risks. AI & Soc. 2021 Sep 1;36(3):705–13.
- 12 Auernhammer J, Zallio M, Domingo L, Leifer L. Facets of Human-Centered Design: The Evolution of Designing by, with, and for People. In: Meinel C, Leifer L, editors. Design Thinking Research : Achieving Real Innovation [Internet]. Cham: Springer International Publishing; 2022 [cited 2024 Feb 14]. p. 227–45. (Understanding Innovation). Available from: https://doi.org/10.1007/978-3-031-09297-8_12
- 13 Salwei ME, Hoonakker PL, Pulia M, Wiegmann D, Patterson BW, Carayon P. Retrospective analysis of the human-centered design process used to develop a clinical decision support in the emergency department: PE Dx Study Part 2. Human Factors in Healthcare. 2023;4:100055.
- 14 Chen Y, Clayton EW, Novak LL, Anders S, Malin B. Human-Centered Design to Address Biases in Artificial Intelligence. J Med Internet Res. 2023 Mar 24;25:e43251.
- 15 Shneiderman B. Human-Centered Artificial Intelligence: Reliable, Safe & Trustworthy. International Journal of Human–Computer Interaction. 2020 Apr 2;36(6):495–504.
- 16 Embi PJ. Algorithmovigilance—Advancing Methods to Analyze and Monitor Artificial Intelligence–Driven Health Care for Effectiveness and Equity. JAMA Network Open. 2021 Apr 15;4(4):e214622.
- 17 Friedman B, Hendry DG. Value Sensitive Design: Shaping Technology with Moral Imagination. MIT Press; 2019. 258 p.
- 18 The White House [Internet]. 2023 [cited 2024 Feb 20]. OMB Releases Implementation Guidance Following President Biden's Executive Order on Artificial Intelligence | OMB. Available from: https://www.whitehouse.gov/omb/briefing-room/2023/11/01/omb-releases-implementation-guidance-following-president-bidens-executive-order-on-artificial-intelligence/
- 19 Commissioner O of the. Focus Area: Artificial Intelligence. FDA [Internet]. 2022 Sep 6 [cited 2024 Feb 20]; Available from: https://www.fda.gov/science-research/focus-areas-regulatory-science-report/focus-area-artificial-intelligence
- 20 Matheny ME, Whicher D, Thadaney Israni S. Artificial Intelligence in Health Care: A Report From the National Academy of Medicine. JAMA. 2020 Feb 11;323(6):509–10.
- 21 Commission launches AI innovation package to support Artificial Intelligence startups and SMEs | Shaping Europe's digital future [Internet]. 2024 [cited 2024 Feb 20]. Available from: https://digital-strategy.ec.europa.eu/en/news/commission-launches-ai-innovation-package-support-artificial-intelligence-startups-and-smes
- 22 Umbrello S, van de Poel I. Mapping value sensitive design onto AI for social good principles. AI Ethics. 2021 Aug 1;1(3):283–96.
- 23 Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021 Mar 29;372:n71.
- 24 Hallowell N, Badger S, Sauerbrei A, Nellåker C, Kerasidou A. “I don't think people are ready to trust these algorithms at face value”: trust and the use of machine learning algorithms in the diagnosis of rare disease. BMC Medical Ethics. 2022 Nov 16;23(1):112.
- 25 Cenci A, Ilskov SJ, Andersen NS, Chiarandini M. The participatory value-sensitive design (VSD) of a mHealth app targeting citizens with dementia in a Danish municipality. AI Ethics. 2023 Apr 13;1–27.
- 26 Wang H, Gupta S, Singhal A, Muttreja P, Singh S, Sharma P, et al. An Artificial Intelligence Chatbot for Young People's Sexual and Reproductive Health in India (SnehAI): Instrumental Case Study. J Med Internet Res. 2022 Jan 3;24(1):e29969.
- 27 Götzl C, Hiller S, Rauschenberg C, Schick A, Fechtelpeter J, Fischer Abaigar U, et al. Artificial intelligence-informed mobile mental health apps for young people: a mixed-methods approach on users' and stakeholders' perspectives. Child and Adolescent Psychiatry and Mental Health. 2022 Nov 17;16(1):86.
- 28 Kerr JI, Naegelin M, Benk M, V Wangenheim F, Meins E, Viganò E, et al. Investigating Employees' Concerns and Wishes Regarding Digital Stress Management Interventions With Value Sensitive Design: Mixed Methods Study. J Med Internet Res. 2023 Apr 13;25:e44131.
- 29 Cagliero D, Deuitch N, Shah N, Feudtner C, Char D. A framework to identify ethical concerns with ML-guided care workflows: a case study of mortality prediction to guide advance care planning. J Am Med Inform Assoc. 2023 Apr 19;30(5):819–27.
- 30 Helenason J, Ekström C, Falk M, Papachristou P. Exploring the feasibility of an artificial intelligence based clinical decision support system for cutaneous melanoma detection in primary care – a mixed method study. Scandinavian Journal of Primary Health Care. 2023;0(0):1–10.
- 31 Helman S, Terry MA, Pellathy T, Hravnak M, George E, Al-Zaiti S, et al. Engaging Multidisciplinary Clinical Users in the Design of an Artificial Intelligence-Powered Graphical User Interface for Intensive Care Unit Instability Decision Support. Appl Clin Inform. 2023 Aug;14(4):789–802.
- 32 Fischer A, Rietveld A, Teunissen P, Hoogendoorn M, Bakker P. What is the future of artificial intelligence in obstetrics? A qualitative study among healthcare professionals. BMJ Open. 2023 Oct 1;13(10):e076017.
- 33 Gibson JJ. The Theory of Affordances. In: The ecological approach to visual perception. Boulder, Colorado: Taylot & Francis; 1979. p. 119–37.



