Abstract
Artificial intelligence (AI) has been extensively researched in medicine, but its practical application remains limited. Meanwhile, there are various disparities in existing AI-enabled clinical studies, which pose a challenge to global health equity. In this study, we conducted an in-depth analysis of the geo-economic distribution of 159 AI-enabled clinical studies, as well as the gender disparities among these studies. We aim to reveal these disparities from a global literature perspective, thus highlighting the need for equitable access to medical AI technologies.
Similar content being viewed by others
In the rapidly developing field of healthcare, artificial intelligence (AI) has emerged as a pivotal force driving innovation in clinical research and improving the efficiency of clinical studies1,2,3,4,5,6. However, despite rapid technological advancements, its practical application in clinical settings remains limited. Concurrently, there are complex disparities in current AI-enabled clinical studies. These disparities, which include data and algorithms, participants and subjects, and access to cutting-edge technologies7,8,9,10, challenge the equitable implementation of AI solutions11,12.
We identified 159 clinical studies of AI-enabled applications from Embase, MEDLINE, and CINAHL through a systematic review. Among these studies, 109 were conducted in hospital settings, while 50 took place in non-hospital environments. Notably, 51.6% (82/159) of the studies utilized AI for treatment and management, and 40.9% (65/159) focused on AI-assisted diagnosis, with a significant portion related to gastroenterology (see Supplementary Table 1). Moreover, 5.0% (8/159) of the studies applied AI for prognosis, and 2.5% (4/159) studies explored its use in patient education. In this study, we primarily analyzed the geo-economic distributions as well as the gender disparities among the study subjects.
As depicted in Fig. 1a, the majority of studies were conducted in North America, Europe, and East Asia, with the United States (44 studies) and China (43 studies) leading. A significant portion (74.0%) of the clinical studies were implemented in high-income countries, 23.7% in upper-middle-income countries, 1.7% in lower-middle-income countries, and only one clinical study was conducted in low-income countries (i.e., Mozambique), as shown in Fig. 1b. Meanwhile, we analyzed the geo-economic distribution of 318 first and last authors, which revealed a similar distribution pattern. Additionally, we observed that funding status is correlated with country income level; the funding rate in high-income countries was as high as 83.8%, while it was only 68.3% in upper-middle-income countries, as shown in Fig. 1c.
We further explored the gender disparities among the subjects in these AI-enabled clinical studies. After excluding studies of gender-specific diseases, 146 studies reported gender information, of which only 3 (2.1%) reported an equal number of male and female subjects. We then classified the gender ratio of the remaining studies into three categories: “low disparity (0.7–1)”, “moderate disparity (0.3–0.7]”, and “high disparity (0–0.3]”. As illustrated in Fig. 2, 10.3% (15/146) of the clinical studies exhibited high gender disparity, while another 36.3% (53/146) demonstrated moderate gender disparity. Among the 15 clinical studies with high gender disparity, males predominated in 8 studies, 25% (2/8) of which were related to obstructive sleep apnea (OSA) due to the higher prevalence of OSA in males13. Females were the majority in 7 studies, of which 28.6% (2/7) were linked to obesity14. Additionally, one study had a higher proportion of female subjects due to the inclusion of a subset of patients undergoing gynecological surgeries. More information is detailed in Supplementary Table 2.
In summary, our study examined 159 AI-enabled clinical studies (of which only 6.3% (10/159) were international multicenter studies), and the results indicate significant disparities in both the geo-economic distribution of these clinical studies and the gender of the study subjects.
The leading countries and authors are primarily from high-income and upper-middle-income countries. The limited representation of lower-middle-income and low-income countries is concerning since these countries may have different disease profiles and are, therefore, unable to benefit from existing research15. Moreover, the presence of gender disparities in clinical studies could lead to potential health inequalities.
Considering these disparities, it is crucial to involve a diverse group of researchers and gender-balanced subjects in the design of AI-enabled clinical studies. Additionally, international multi-center clinical studies should be encouraged, particularly including sites from lower-middle-income and low-income countries. This will enable AI tools to be effectively applied and evaluated in diverse healthcare systems around the world. The pursuit of health equity will lay a solid foundation for the sustainable development and application of AI in healthcare.
Methods
Search strategy
We conducted a systematic literature search using Embase, MEDLINE, and CINAHL to identify AI-enabled clinical studies published prior to January 3, 2024. The search strategy we used included keywords and Medical Subject Headings (MeSH) terms related to “Artificial Intelligence” and “Clinical Study”. Additionally, we manually reviewed the references of the included studies to identify more relevant studies (Detailed search strategy can be found in Supplementary Material Search Strategy and PRISMA flow diagram can be found in Supplementary Fig. 1).
Inclusion and exclusion criteria
Each article was independently screened by two researchers (R.Y. and S.V.N.) to determine eligibility. Disagreements were resolved through consultation with a third researcher (Y.K.). Studies were included if they: (1) incorporated a significant AI component, defined as a nonlinear computational model (including, but not limited to, support vector machine, decision trees, neural networks, etc.)6; (2) were applied in clinical settings, influencing the patient’s health management; and (3) were published as a full-text article in a peer-reviewed, English-language journal. Additionally, we excluded studies that used linear models (such as linear regression and logistic regression), conducted secondary analyses or did not integrate AI algorithms into clinical practice. For each included study, we extracted information including gender information of the subjects, the country of clinical study implementation, the income level of the country (based on World Bank’s classifications16), the funding, and so on. All extracted information can be found in Supplementary Table 1.
Data availability
The data used in the manuscript can be found in Supplementary Material.
References
Rajpurkar, P., Chen, E., Banerjee, O. & Topol, E. J. AI in health and medicine. Nat. Med. 28, 31–38 (2022).
Hinton, G. Deep learning—a technology with the potential to transform health care. JAMA 320, 1101–1102 (2018).
Woo, M. An AI boost for clinical trials. Nature 573, S100–S102 (2019).
Yang, R. et al. Large language models in health care: development, applications, and challenges. Health Care Sci. 2, 255–263 (2023).
Ke, Y. H. et al. Enhancing diagnostic accuracy through multi-agent conversations: using large language models to mitigate cognitive bias. Preprint at arXiv:1504.14589 (2024).
Han, R. et al. Randomised controlled trials evaluating artificial intelligence in clinical practice: a scoping review. Lancet Digit. Health 6, e367–e373 (2024).
Chen, I. Y., Joshi, S. & Ghassemi, M. Treating health disparities with artificial intelligence. Nat. Med. 26, 16–17 (2020).
Nordling, L. A fairer way forward for AI in health care. Nature 573, S103–S105 (2019).
Celi, L. A. et al. Sources of bias in artificial intelligence that perpetuate healthcare disparities—a global review. PLoS Digit. Health 1, e0000022 (2022).
Liu, M. et al. A translational perspective towards clinical AI fairness. npj Digit. Med. 6, 1–6 (2023).
Serra-Burriel, M., Locher, L. & Vokinger, K. N. Development pipeline and geographic representation of trials for artificial intelligence/machine learning-enabled medical devices (2010 to 2023). NEJM AI https://doi.org/10.1056/aipc2300038 (2023).
Alberto, I. R. I. et al. A scientometric analysis of fairness in health AI literature. PLoS Glob. Public Health 4, e0002513 (2024).
O’Connor, C., Thornley, K. S. & Hanly, P. J. Gender differences in the polysomnographic features of obstructive sleep apnea. Am. J. Respir. Crit. Care Med. 161, 1465–1472 (2000).
Cooper, A. J., Gupta, S. R., Moustafa, A. F. & Chao, A. M. Sex/gender differences in obesity prevalence, comorbidities, and treatment. Curr. Obes. Rep. 10, 458–466 (2021).
Black, E. & Richmond, R. Improving early detection of breast cancer in sub-Saharan Africa: why mammography may not be the way forward. Glob. Health 15, 1–11 (2019).
Hamadeh, N. et al. New World Bank Country Classifications by Income Level: 2022–2023 (World Bank Blogs, 2022).
Acknowledgements
This work was supported by the Duke-NUS Signature Research Program funded by the Ministry of Health, Singapore. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not reflect the views of the Ministry of Health. Additionally, we appreciate Jonathan Liew’s contribution to information extraction during the revision stage.
Author information
Authors and Affiliations
Contributions
N.L. conceived the study. R.Y., S.V.N., M.L., Y.N., and N.L. designed the study. R.Y., S.V.N., Y.K., and D.D. collected data. R.Y. conducted data analyses. R.Y. and S.V.N. drafted the manuscript, with further development by N.L. N.L. supervised the study. All authors contributed to the revision of the manuscript and approval of the final version.
Corresponding author
Ethics declarations
Competing interests
N.L. is an Editorial Board Member for npj Digital Medicine. He played no role in the peer review of this manuscript. The remaining authors declare that there are no other financial or non-financial competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Yang, R., Nair, S.V., Ke, Y. et al. Disparities in clinical studies of AI enabled applications from a global perspective. npj Digit. Med. 7, 209 (2024). https://doi.org/10.1038/s41746-024-01212-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41746-024-01212-7