Abstract
Digitalization in medicine offers a significant opportunity to transform healthcare systems by providing novel digital tools and services to guide personalized prevention, prediction, diagnosis, treatment and disease management. This transformation raises a number of novel socio-ethical considerations for individuals and society as a whole, which need to be appropriately addressed to ensure that digital medical devices (DMDs) are widely adopted and benefit all patients as well as healthcare service providers. In this narrative review, based on a broad literature search in PubMed, Web of Science, Google Scholar, we outline five core socio-ethical considerations in digital medicine that intersect with the notions of equity and digital inclusion: (i) access, use and engagement with DMDs, (ii) inclusiveness in DMD clinical trials, (iii) algorithm fairness, (iv) surveillance and datafication, and (v) data privacy and trust. By integrating literature from multidisciplinary fields, including social, medical, and computer sciences, we shed light on challenges and opportunities related to the development and adoption of DMDs. We begin with an overview of the different types of DMDs, followed by in-depth discussions of five socio-ethical implications associated with their deployment. Concluding our review, we provide evidence-based multilevel recommendations aimed at fostering a more inclusive digital landscape to ensure that the development and integration of DMDs in healthcare mitigate rather than cause, maintain or exacerbate health inequities.
Keywords
Introduction
Emerging patient-centered technologies in the growing field of digital medicine and health come with the promise to improve patient health outcomes and healthcare delivery by supporting personalized healthcare. 1 Digital Medical Devices (DMDs) and related health technologies, including sensors, wearables, apps, personal health records, and management platforms, have the capacity to inform, manage, diagnose, monitor, or offer treatment throughout the healthcare journey. Their use and adoption is of particular importance due to the complex healthcare interventions required for the growing number of individuals living with chronic conditions, who might be more dependent on these technologies. 2 In the wider literature, DMDs have been described as digital tools or health technologies that are undergoing clinical validation and are incorporated into the healthcare ecosystem in order to inform prevention, diagnosis, and provide treatment and management of health conditions. 3 Thus, through a data-driven approach, DMDs can support the clinical decisions of healthcare professionals and can encourage and support positive health behaviors, and thus improve health and care outcomes for all members of society, including those within underserved communities.
DMDs can broadly be classified into three main categories
4
:
Those designed to provide more efficiently managed healthcare services—such as electronic prescribing or triaging systems; those that support people in the self-management of their health and wellness—such as communication channels between patients and healthcare professionals, health-promoting information or advice on self-management, electronic diaries, etc.; those used to diagnose, monitor, and treat health conditions, as well as to inform about disease context and care options—this for example includes digital therapeutics directly treating a distinct aspect of the disease, or digital diagnostics used to support clinical decision-making.
DMDs are designed with the primary goal of enhancing the quality of patient care by enabling remote monitoring, facilitating early diagnosis and access to healthcare services, and actively engaging and empowering both patients and healthcare professionals.5,6 However, the mode of action of DMDs raises unique societal challenges that differ from those associated with traditional pharmacological interventions.
7
As healthcare systems increasingly incorporate their use, there are growing social and ethical concerns regarding the potential risks and consequences of introducing them into society. In digital medicine, socio-ethical challenges are primarily addressed within the interdisciplinary frameworks of ELSI (Ethical, Legal, and Social Implications) and Responsible Innovation.8,9 These approaches ensure that technological advancements are developed and implemented in ways that maximize benefits while minimizing negative consequences for individuals, particularly respecting their autonomy, privacy, and trust. By emphasizing fairness, these frameworks aim to guarantee that DMDs benefit all segments of society without exacerbating health inequities—that is, unfair and avoidable differences in health across the population and between different groups.
10
While a substantial body of literature addresses various social and ethical aspects of digital medicine, these discussions often remain confined within the silos of individual disciplines. In this narrative review, we take a multidisciplinary approach and delve into comprehensive discussions on five core socio-ethical considerations associated with the digital determinants of health: (i) equity in access, use and engagement with DMDs, (ii) inclusiveness in DMD clinical trials, (iii) algorithm fairness, (iv) surveillance and datafication, and (v) data privacy and trust. Concluding this review, we present evidence-based recommendations across multiple levels of governance that encompass the individual, organizational, and national levels, that aim to drive policy action towards addressing equity and inclusion in digital medicine.
Methods
This review draws on narrative review method, that is best suited to provide a comprehensive overview of a complex and interdisciplinary topic and to provoke further reflection on a broader subject of interest11,12 Moreover, narrative reviews provide a valuable knowledge base used to inform policy and practice. 11 For this review, we carried out a literature search in leading academic databases: PubMed, Google Scholar, and Web of Science up to November 2023 by using phrases and keywords combining “digital” “ethical,” “socio-ethical,” “health,” “medicine” “technologies,” “medical devices,” “equity,” “inclusion,” “diversity,” “inequalities” “inequity,” “divide.” We searched for empirical and case studies, theoretical articles, reviews and commentaries from experts, and grey literature with full text in the English language. Following a thorough review of the articles, we synthesized five key topics of policy relevance that are identified and summarized in this review.
Equity in access, use and engagement with digital medical devices
The inequities in the adoption of DMDs documented among social groups reflect those that exist in terms of health outcomes and quality of healthcare received. In all cases, poorer outcomes are found among the most socially disadvantaged communities.10,13 Therefore, one of the most critical prerequisites before DMDs can have a positive impact on people's lives is that they are accepted and used by all members of society.1,14 However, a large body of evidence has documented inequities in the access to, use of and engagement with DMDs between social groups and even between countries and regions.14–17 Different aspects underlying the social and digital determinants of health, among which age, gender, ethnicity, socioeconomic status, cultural and linguistic background, place of residence, disease severity and digital literacy, as well as reimbursement policies, can influence the way in which DMDs are used.1,8,13,18–20 Studies in the United States (US), United Kingdom (UK) and mainland Europe have repeatedly shown that video telehealth visits and mobile health applications were more frequently downloaded and used by those who were younger and had higher incomes and educational levels.18–20 This pattern is also mirrored in the use of personal health records: those with socially advantaged backgrounds tend to have higher levels of access and engagement.21–23 In an ever-evolving digital landscape, the digital divide paradigm has shifted. Initially, its primary concern of the digital divide was access to information and communication technology (ICT) and internet penetration. The focus then moved to digital skills and effective engagement with digital technologies. 24 Indeed, in more economically developed countries in which the majority of the population owns a smartphone and has access to the internet, 25 the primary concern regarding access to DMDs often revolves around engagement and adherence. Merely having access to devices and the internet does not guarantee their meaningful use and engagement.1,23,26 Users often find that DMDs are designed for a very particular target group—those who are able-bodied, digitally and health literate, and motivated and that they often require time-consuming and complex data entry processes. 6 For example, DMD users with chronic conditions have highlighted the psychological burden of constant reminders of their poor health, mistrust, as well as a lack of time and skills needed to navigate novel digital tools, thus pointing towards the manifestation of the digital health paradox that asserts that population groups that can most benefit from DMDs are also the ones most likely to be excluded from them.17,27
While it has been well noted that early adopters of digital innovations are in general those with higher economic, educational and social capital,1,27,28 the pattern can also be seen in terms of adherence to the use of the technology, with lower engagement and higher drop off rates among those with lower income and educational levels. 20 Some of the mechanisms behind these patterns might be higher awareness, digital capabilities as well as supportive social networks, and financial affluence which enable health maintaining strategies and self-efficacy among the most advantaged.1, 29–31
When engaging with the issue of inequities in the access to, use of and engagement with DMDs, it is vital to not only consider socioeconomic, gender, or ethnic inequities in isolation, but also to explore how inequities are mutually reinforcing or intersecting between gender and socioeconomic position, 32 as well as how they may intersect with broader social and cultural norms. 33 If these challenges are not addressed, widespread adoption of DMDs may thus exacerbate health inequities, a phenomenon known as “intervention-generated inequality.” 34 It would also be a lost opportunity since socially disadvantaged individuals stand to benefit the most from these interventions.17,29,35 Therefore, careful consideration of the needs of socially disadvantaged populations is required. On one hand, the design of the DMDs needs to be guided by people-centered approaches, such as through participatory and user-centered design, to reflect the everyday experiences of a diverse range of users. 36 On the other hand, the development and personalization of health interventions addressing health inequities need to take into account the needs of socially disadvantaged members of society.1,17 Building on the capability approach, 1 if conversion factors such as socioeconomic status, digital skills or disease severity are taken into account, DMDs can have the potential to provide choices and opportunities to increase people's capabilities to function and attain a good state of health and well-being. 37 They can be leveraged to reach underserved communities by providing them with personalized and culturally appropriate information, treatment and care. This could enable improvements in health literacy and skills essential for disease management and the maintenance of good health as well as facilitate access to healthcare services especially for those living in rural areas. All of this will ultimately contribute towards creating a more inclusive and equitable healthcare environment.1,37
Diversity and inclusion in clinical trials in digital medicine research
Clinical trials are considered to provide the most robust evidence on the safety, effectiveness, and efficiency of (digital) health interventions. The results of these studies therefore need to be applicable and generalizable to an increasingly diverse population. However, multiple studies have now shown that racial and ethnic communities, individuals from lower socioeconomic backgrounds, undocumented migrants as well as older people have been underrepresented, misrepresented or made entirely invisible in clinical trials, health datasets, and research studies in general,38,39 which has recently been described as “health data poverty.” 40 This may be among the reasons that developments in research have not always translated into more effective treatment options for these groups. 41 As an example, the COVID-19 pandemic exposed critical flaws related to the lack of diversity in clinical trials, and diagnostic instruments being adapted to characteristics of predominantly white samples, such as Pulse Oximetry readings which led to a higher misclassification of patient oxygenation status in less represented social groups, and consequently higher rates of forgone COVID treatment for these minorities in clinical practice.42,43
Some of the potential mechanisms that underpin the underrepresentation of disadvantaged population groups in clinical trials and other healthcare and population health research initiatives are thought to be the lack of awareness of research initiatives by disadvantaged communities, limited opportunities for participation, logistical reasons (e.g. cost, transportation, device and internet availability, or time commitment), as well as higher levels of mistrust in science, medical professionals and the health system as a whole.33,44 Another related issue that could lead to an important selection bias is retention in digital health studies. Studies in the digital medicine field have shown higher attrition rates in clinical trials on the effectiveness of digital technologies among socially disadvantaged communities. 45 While both digital and non-digital trials face challenges with attrition among socially disadvantaged communities, digital trials may have additional unique barriers related to digital literacy, technology access, and cultural relevance. This shows that targeted recruitment of more disadvantaged individuals into these studies needs to be accompanied by support for the most vulnerable to continue study participation. Such support might include reminders, feedback based on collected data, and clear instructions and techical support on how to use DMDs. It is within the researcher's responsibility to design studies that ensure diversity and recruit participants from varied backgrounds, thereby reflecting the demographic diversity of the population affected by the disease. It is crucial to transparently report data on effectiveness of digital health intervention by the socioeconomic status, age, race, ethnicity, and gender of trial participants to identify and address gaps in study representation and the generalizability of findings. Additionally, researchers must engage in raising awareness about the importance of research participation and provide necessary logistical support to facilitate the involvement of diverse populations. However, we must also consider the mutually reinforcing effects between the different factors. Figure 1 shows a hypothetical directed acyclic graph (DAG) developed by authors based on the included literature that highlights how participation and attrition in clinical trials of DMDs can be influenced by the abovementioned factors.

A hypothetical DAG that highlights the interconnectedness of factors influencing participation and attrition rates of socially disadvantaged communities in clinical trials of DMDs.
The use of real-world evidence from routine healthcare procedures offers a great opportunity as it comes with fewer limitations compared to the challenges in the recruitment and retention of diverse populations in clinical trials.46,47 Electronic health records, claims data, or disease registers can gather valuable information from large and diverse populations, thus providing useful information on inequities in healthcare access and outcomes, quality of care or treatment effects for disadvantaged communities. 27 However, they also come with limitations in terms of data quality and by definition exclude populations who might not have access to personal health records, or fall in between the cracks of the healthcare system. 27 Thus, bridging the gap between science and health improvement requires a substantial effort to ensure the representation and engagement of diverse populations in clinical trials and research within the digital medicine field.
Fairness in artificial intelligence and machine learning
One of the biggest concerns in implementing DMDs is the socio-ethical implications of their application of artificial intelligence (AI) and machine learning (ML) methods and consequently, their potential to have a negative effect on underserved populations. 48 While advancements in AI can offer vast opportunities to support healthcare practices and to deliver personalized care at lower cost, particularly through increasing predictive accuracy to support patients care pathways, diagnostics, and treatment decisions,49,50 recent studies have highlighted that AI and ML approaches can in fact introduce and reproduce inequities in the way patients are diagnosed, given treatment and billed for healthcare or even provided with welfare benefits.51,52
Algorithmic bias might emerge from data collection through incomplete or unrepresentative data. In some countries, such as France and Germany, data on racial and ethnic minorities is not recorded. 48 Likewise, bias might also arise from outcome definitions, labeling and deciding on threshold values.48,53 For instance, scholars have argued that the oversimplification of racial categories as well as the fact that the sex and gender dimensions have been ignored in the design of algorithms, have contributed to poor results and discriminatory outcomes.54,55 As an example, some studies find that women have a higher prevalence of depression than men, however, these results might be influenced by the validity of the measuring instruments, since the expression of depression and emotional vulnerability in men and women differs, partly due to the societal gender roles related to masculinity. 56 Another example emerges from one of the goals of AI, which is to identify patients at higher risk of disease. Predictive models require accurate labeling for patients with a particular disease, as well as for those without, and the establishment of cutoff points for category boundaries across different disease stages. 53 The question is then when, by whom, and how this labelling occurs. Considering the development of algorithms that are not solely developed from mathematical concepts but rather incorporate intersectional theories and knowledge on the determinants of health is of paramount importance.
On the other hand, AI bias might arise during model development such as through inaccurate assumptions about the causal relationships between different parameters included in the model. 48 As observed differences in health outcomes based on race, gender or sex are strongly intertwined with historical, socio-cultural and institutional contexts, with healthcare settings frequently adding to racialized experiences and discrimination, 57 biases related to confounding factors are becoming increasingly important. With the increasing deployment of algorithms, biases may prevent adequate diagnosis, treatment and care for already underserved populations and increase health inequities. A pragmatic approach would involve recognizing that only a few scenarios exist wherein the health disparity between majority and minority social groups significantly diminishes. 58 Instead, it would be more prudent to systematically examine stratified analyses to assess potential benefits for both majority and minority populations, deeming these benefits sufficient to justify the use of the new device. 59
A large body of evidence has now reported gender and racial biases as a consequence of AI applications across a range of health conditions, such as cardiovascular diseases, cancer, as well as mental and maternal health.60,61 For example, several healthcare algorithms used in clinical practice include a race category to arrive at treatment or referral decisions with possibly problematic outcomes for minority groups. As an example, one algorithm calculates a heart failure risk score in order to predict the risk of death for hospitalized patients. Because of a race-based adjustment algorithm, patients identified as “Black” were categorized as having a lower mortality risk, resulting in lower rates of referral to specialized care. 60 As race is a social rather than a biological construct, controlling for race might lead to an estimation bias resulting in unintended consequences, such as the under-treatment or over-treatment of racial minorities and perpetuating the discrimination of minorities.60,62
Despite the increase in the awareness of such biases in AI and ML, more work is needed to bring transparency to AI algorithms in order to make possible the effective scrutiny of the assumptions at the heart of the models. 63 For now, AI and ML models for health applications have largely been developed post-data collection. 50 Ethics and governance need to be addressed and reevaluated also after the deployment of the algorithms.63–65 It is important to understand how personalized predictions and clinical decision support algorithms may vary for individuals of different groups, and to assess the extent to which these variations can be linked to validated research findings. Another way forward is to adapt the FAIR principles 2 of AI development and deployment in order to promote the transparency, reproducibility, and responsible use of AI technologies. 66 While AI and ML are increasingly employed in diagnostic prediction, monitoring, and healthcare decision-making, it is crucial to raise awareness among healthcare professionals and patients regarding the risks, consequences, and opportunities associated with AI in relation to equity. A recent article by Chen (2020) argues that if AI is used purposefully, it could identify and correct inequities, through the development of “fair risk prediction models” to account and adjust for the underrepresentation of socially disadvantaged groups in clinical trials. 67
Addressing datafication and the continuous surveillance of health behaviors and outcomes
The growing availability and use of digital tools, such as wearables, and monitoring systems enable the collection of a wide range of data via GPS tracking, Bluetooth, cameras and voice recording. This provides a novel opportunity to gather, store, share, and analyze vast amounts of health data across an individual's lifespan. It allows for continuously collect valuable patient information on health behaviors such as diet, heart rate, movement or sleep patterns, which could benefit different healthcare actors such as healthcare professionals, patients and healthcare commissioners. Importantly, these tools offer the potential to reduce health disparities by providing underserved populations with greater access to healthcare services and by offering personalized interventions via AI-driven risk stratification. 68
Data collected through digital tools and their metadata can provide quick and objective information about human behaviors and interactions: where the individuals are located, with whom they are interacting, and for how long. 69 This gives rise to the risk of so-called patient “datafication”—a term coined by Cukier and Mayer-Schoenberger (2013) for the transformation of people's daily activities into quantified data flows captured through continuous surveillance and self-tracking. 70 This emphasis on self-monitoring and self-management through continuous surveillance, although beneficial in understanding health conditions and behaviors, can come with important socio-ethical concerns related to the concept of equity. Indeed, while these tools can provide patients with greater autonomy in the management of their condition, the emphasis on patient empowerment, lifestyle intervention and self-care through personal responsibility fails to recognize the broader digital, political, social and economic determinants of health.71,72 Limited digital literacy or socioeconomic circumstances make it hard for individuals to adopt and comply with behavioral health interventions. 73 Moreover, many patients, particularly those from socially disadvantaged backgrounds, may struggle to understand the complex medical data generated by continuous monitoring. 74 This difficulty can exacerbate feelings of anxiety and stress and can negatively impact the ability of those with lower health literacy to make informed health choices. Datafication could also introduce a disciplining effect and even be used to restrict entitlement to healthcare resources for individuals in greater need. Indeed, scholars have highlighted the emerging trend in the USA of encouraging employees to engage in self-monitoring activities aimed at behavioral change such as exercising, keeping a healthy diet, or smoking cessation in return for financial gains or healthcare insurance discounts.75–77
In the healthcare context, the data available through the linkage of a wide range of DMDs could be used to stratify patients based on certain risk scores using data from medical or even non-medical contexts (patients’ social connectedness, socioeconomic position, etc.). Using these risk scores, healthcare providers could then decide whether to offer further treatment or prevention programs. Hogle (2016), for instance, gives a real-world example of a health insurance company that uses a “drug adherence index” to identify noncompliant patients (or those at risk of noncompliance) based on the behavior within a given prescribed medicine protocol. Since a patient's socioeconomic circumstances is determinant of non-adherence to medication, 78 this could lead to the classification of certain patients with socially disadvantaged backgrounds as “higher cost,” with potentially lasting implications for their healthcare trajectory and further exacerbation of health inequities. In a value-driven healthcare system, in which providers are rewarded based on improving patient health outcomes, there is an incentive to flag “problematic” patients. 76 This is explained by “surveillance capitalism” rooted in political economy in which data is commodified by private companies and big corporations and used to shape people's actions and behaviors, ultimately leading to profit. 79 That said, it is worth noting that all these risks depend on the way in which the information collected through continuous surveillance is used and regulated. Indeed, identifying the types of patients struggling with adherence could provide important information to healthcare professionals and serve, for example, to facilitate targeted additional health and social care services to help patients in need.
As outlined above, the datafication of health can present both opportunities and challenges. Further research is required to better comprehend how individuals from socially advantaged and disadvantaged backgrounds can equally benefit from the data collected through continuous surveillance. Data from the DMDs should be user-friendly, offering easy-to-read information to prevent information overload or misinterpretation. Risk-benefit analyses need to be complemented with research on individual attitudes and preferences for the treatment and usage of their health data from multiple sources. Additionally, using privacy-preserving techniques and well-built data infrastructures, researchers may be able to leverage health information contained within health data without the need to access personal data. 80 Data would continue to be stored in secure environments, thus ensuring their safety, while researchers are enabled to perform analyses using aggregate information. 81 Current efforts to develop so-called synthetic data, which mimics real-world health data without containing actual patient information, as well as reliance on federated data networks such as the European Health Data Space (EHDS) have a great potential to mitigate some of the risks related to datafication in the European context. 81 However, technical advancement related to health data needs to be followed by developing easy-to-read information on the risks and benefits of DMDs that are meaningful for patients and citizens and adapted to their language and health literacy proficiency levels. On an individual and societal level, it has been argued that people are often willing to accept risks if the benefits significantly outweigh the potential drawbacks. 82 These considerations must therefore be guided by clear benefits and policy decisions justified within a relevant ethical framework.
Addressing privacy and trust in digital medicine
Privacy, confidentiality and trust are fundamental values defined by the oath of Hippocrates and comprise an essential part of medical ethics and professional regulations. The issue of privacy due to the digitalization of medicine is complex and multi-faceted and entails respecting the individual's right to manage the collection, use, and sharing of their personal information. 83 It considers data stewardship, data protection, security and data transparency as well as data governance. In the context of this review, the emphasis will be placed on the way in which privacy intersects with the notion of equity. This means engaging with the issue of how trustworthy DMDs are perceived by the users in terms of data privacy. Given that these devices generate vast amounts of personal health data, trust and privacy are both critical and closely intertwined concepts when it comes to assessing their socio-ethical considerations in digital medicine.
Trust in DMDs typically refers to the level of beliefs and confidence that individuals have in the technical aspects of data usage in terms of accuracy, reliability or data security, as well as the assurance that the system will provide benefits without causing harm.83,84 While in the European context, the technical aspects of data privacy and data security are outlined in the General Data Protection Regulation (GDPR)—data encryptions, anonymisation and access control, when considering socio-ethical concerns of DMDs—trust can be established through such data privacy on a regulatory level (such as through GDPR), and it is intertwined with how data privacy is ultimately perceived by end users (trustworthiness of DMDs).17,36 Trust plays a central role in the acceptance and use of these devices, and it also hinges on the issue of transparency.
36
Regarding equity, evidence indicates that people from lower socioeconomic groups, older demographic groups - a dominant user group in digital medicine, and ethnic minorities might be more vulnerable to security breaches due to the lack of linguistically and culturally appropriate information on cybersecurity, and also may have less trust in healthcare services and the security of their personal health information.33,85 One key issue in relation to the establishment of trust in DMDs is the fact that these innovations fall under the general umbrella of medical devices, which may inadequately address the unique complexities brought forward by DMDs.1,7 Without clear perceived health benefits, the assurance of transparency and robust privacy measures, some populations might be reluctant to engage with these health technologies. On the flip side, some populations might be hesitant to provide information about technologies developed by public institutions due to prior negative experiences with public healthcare providers or a lack of trust in public institutions in general. For example, it has been shown in the context of the USA that there is a higher likelihood of distrust in the healthcare system for African Americans than whites, as well as higher rates of reported racism and concerns about privacy arising from patient electronic health records
Another closely related issue affecting trust in DMDs concerns the extent to which the users are privacy literate and know how to effectively understand and protect the privacy of the digital information they allow these technologies to collect. Although distinct, digital literacy and privacy literacy go hand in hand and as such might disproportionally affect socially disadvantaged populations.87,88 Those with higher digital privacy literacy in the context of a digital world can more successfully protect themselves from security threats and therefore be more confident when engaging with DMDs. 89 This can in turn empower and enable them to leverage their health data during health visits. Conversely, individuals with limited digital privacy literacy may be hesitant to engage with new health technologies potentially due to concerns of exposing their health information and therefore hinder the opportunity to actively engage in their health. Given that health information is a particularly sensitive issue, the development of innovative privacy-preserving data infrastructures and increasing awareness of their existence, particularly among socially disadvantaged individuals, is an essential requirement for addressing the issues of trust and privacy in digital medicine.
Multilevel recommendations to advance equity, diversity and inclusion
Depending on how DMDs are designed, deployed and integrated into the healthcare systems, they can exacerbate or potentially alleviate health inequities by reaching and serving underserved communities. To achieve the latter, we suggest a range of individual, organizational (e.g. academia, healthcare, industry) and national level recommendations, summarized in Figure 2, and further elaborated in Appendix 1.

Multilevel recommendations for the advancement of equity across different socio-ethical challenges in digital medicine.
Individual level
To ensure equitable access, use, and adherence to DMDs, efforts should be made to create an inclusive digital health ecosystem by providing internet access, and affordable equipment, complemented by financial, technical, and social support for those in need.To raise awareness about clinical trials through culturally appropriate outreach and incentives like transportation costs and childcare should be offered to research participants. Additionally, efforts should be made to identify and address algorithm bias in AI and to enhance literacy related to data privacy and security.
Organizational level
Organizations should adopt participatory design approaches in developing DMDs, focusing on the inclusion of diverse populations. They should continuously monitor disparities in DMD access and use and ensure that trials include and retain participants from diverse cultural and socioeconomic backgrounds, providing necessary instructions on the use of digital technologies and user-friendly feedback on the collected data. In addition, oversampling of underrepresented populations and transparency in data reporting should be rigorously implemented. Collaborations between AI researchers and social scientists should be fostered to ensure fairness and transparency in AI applications. Organizations must implement security infrastructure to protect health data and design privacy-preserving DMDs by principle.
National level
National policies should integrate equity considerations in health technology assessments, promote sustainable digital health business models, and ensure continuous monitoring of DMD inequities in access to, use of, and engagement with DMDs as well as fair health outcomes associated with their use. Standards for including underrepresented populations in research should be developed, along with guidelines for addressing algorithm bias in AI. A clear ethical framework for health data use should be established, with investments in federated data approaches to enhance transparency and trust in digital health systems.
Limitations
While some critics of the narrative review argue that it might lead to the selective use of evidence to support a particular point, this is balanced by the fact that the selection of literature was guided by how pertinent the evidence was to important policy questions that aid decision-makers. 11 This process was aided by the multidisciplinary expertise of the authors (coming from social medicine, social psychology, medical and computer science as well as health policy), which allowed for the careful selection and review of key topics across a wide range of disciplines relevant to equity in digital medicine. Although we recognize the importance of country-specific contexts, most of the recommendations in this narrative review are not tailored to any single country but will apply to most countries within the Global North. Instead, they aim to provide researchers and policymakers with tools to analyze specific interventions within their unique contexts. Lastly, In this review, we focused our discussion on DMDs used in healthcare, acknowledging the slow yet growing trend of their integration into the healthcare systems.
Conclusion
Equity as a lens can shed light on a range of socio-ethical challenges and opportunities linked to the development and deployment of DMDs in the healthcare system and in the wider society. Although not exhaustive, the five socio-ethical considerations discussed in this article underscore the importance of a critical examination and reflection, prompting consideration of potential socio-ethical challenges that may emerge in the current trend of developing data-driven innovations. While digital medicine offers a promising new avenue to optimize healthcare outcomes through the development of robust evidenced-based interventions—and thus improving clinical decision-making, addressing the socio-ethical implications of their use is essential to advance health equity.
Supplemental Material
sj-docx-1-dhj-10.1177_20552076241277705 - Supplemental material for Socio-ethical challenges and opportunities for advancing diversity, equity, and inclusion in digital medicine
Supplemental material, sj-docx-1-dhj-10.1177_20552076241277705 for Socio-ethical challenges and opportunities for advancing diversity, equity, and inclusion in digital medicine by Ivana Paccoud, Anja K. Leist, Isabel Schwaninger and Robin van Kessel, Jochen Klucken in DIGITAL HEALTH
Footnotes
Contributorship
I.P. and J.K. conceptualized the manuscript. I.P. drafted the manuscript. All authors (I.P., A.L., R.V.K., I.S., J.K.) contributed to the conception, critically reviewed the manuscript, and approved it.
Declaration of conflicting interests
The authors declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: J.K. declares an unpaid role in the European Taskforce for Harmonized Evaluation of Digital Medical Devices. A.L. has served on advisory boards with Roche (remunerated). None of the other authors report conflicts of interest.
Ethical approval
Not applicable.
Funding
The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: the Luxembourg National Research Fund: (dHealthPD: 14146272).
Guarantor
JK.
Informed consent
Informed consent was not required as no personal data were processed in this study.
Supplemental material
Supplemental material for this article is available online.
Notes
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
