Wearable biometric monitoring devices (BMDs) and artificial intelligence (AI) enable the remote measurement and analysis of patient data in real time. These technologies have generated a lot of “hype,” but their real-world effectiveness will depend on patients’ uptake. Our objective was to describe patients’ perceptions of the use of BMDs and AI in healthcare. We recruited adult patients with chronic conditions in France from the “Community of Patients for Research” (ComPaRe). Participants (1) answered quantitative and open-ended questions about the potential benefits and dangers of using of these new technologies and (2) participated in a case-vignette experiment to assess their readiness for using BMDs and AI in healthcare. Vignettes covered the use of AI to screen for skin cancer, remote monitoring of chronic conditions to predict exacerbations, smart clothes to guide physical therapy, and AI chatbots to answer emergency calls. A total of 1183 patients (51% response rate) were enrolled between May and June 2018. Overall, 20% considered that the benefits of technology (e.g., improving the reactivity in care and reducing the burden of treatment) greatly outweighed the dangers. Only 3% of participants felt that negative aspects (inadequate replacement of human intelligence, risks of hacking and misuse of private patient data) greatly outweighed potential benefits. We found that 35% of patients would refuse to integrate at least one existing or soon-to-be available intervention using BMDs and AI-based tools in their care. Accounting for patients’ perspectives will help make the most of technology without impairing the human aspects of care, generating a burden or intruding on patients’ lives.
The development of wearable biometric monitoring devices (BMDs) (i.e., sensors embedded in smartphones or wearable equipment to collect physiological, biological, environmental or behavioral information) allows for remote, high-frequency and high-resolution monitoring of patients’ health outside the hospital.1,2,3 Coupled with the progress of artificial intelligence (AI), the thousands of data points collected from BMDs may help in informing diagnosis, predicting patient outcomes, and helping care professionals select the best treatment for their patients.4,5 These two technical revolutions have generated a lot of hope and “hype,” and myriad digital and AI-based tools for healthcare have been developed.6,7,8
Today, AI can outperform medical practitioners in the analysis of skin lesions, pathology slides, electrocardiograms or medical imaging data.9,10,11,12 Continuous glucose monitoring systems combined with closed-loop insulin delivery systems can improve type 2 diabetes mellitus control.13 Multiple AI algorithms using data from BMDs are being tested to detect unknown disease, predict patient outcomes and provide reactive guidance or proactive interventions.14,15,16,17,18 Despite these good preliminary results, the real-world effectiveness of such interventions that occur outside of hospitals is still uncertain and will depend on patients’ engagement, uptake and adherence to these interventions.19
The literature on patients’ views of the use of BMDs and AI in healthcare is scarce and relies mainly on context-specific studies, with limited generalizability, and on evaluations of specific interventions reported in some trials evaluating these technologies, which do not reflect whether patients would have engaged in these interventions outside of the research context.20,21 In this study, we aimed to describe chronic patients’ perceptions of the use of BMDs and AI-based tools in healthcare and assess their readiness to integrate such technologies in their care.
A total of 1183 patients with chronic conditions [861 (73%) female] participated in the study between May and June 2018 (participation rate: 48%) (Table 1). The mean age was 49.7 years (SD = 14.5). Patients’ conditions included diabetes (n = 121), asthma (n = 77), rheumatologic conditions (n = 367), neurological disorders (n = 234) and cancer (n = 107). A total of 649 (54%) participants had multimorbidity (mean number of conditions 2.5 [SD = 2.4]). In total, 590 (50%) participants reported using e-health or m-health tools for health [smartphone apps for 246 (21%), wellness wearable devices for 61 (5%), medically prescribed wearable device (e.g., continuous glucose monitoring tool) for 50 (4%), and health internet services (eg, online appointment tools) for 190 (16%)].
Patients’ perceptions of the use of BMDs and AI in healthcare
After calibrating the dataset to obtain estimates representative of the French population of patients with chronic conditions, we found that 47% of participants considered BMDs and AI as a great opportunity (rating > 7/10) (IQR of opportunity ratings: [5-9]) (Supplementary Table 1). With the open-ended questions, patients identified 47 potential benefits of the use of technology in healthcare. They believed that it could improve their follow-up and the reactivity of care (55%), reduce their burden of treatment (23%) or facilitate physicians’ work (eg, by automating repetitive tasks) (21%) (Fig. 1, Table 2, Supplementary Table 2). For example, a patient described these new technologies as “the only way [for a physician] to simultaneously take into account all multiple parameters necessary to adjust diabetes treatment: insulin sensitivity, duration of action, blood sugar levels, physical activity, continuous measurement…” (35-year-old man reporting diabetes and Hashimoto’s thyroiditis).
In contrast, 11% participants considered BMDs and AI as a great danger (rating >7/10) (IQR of danger ratings: [2–6]). With the open-ended questions, patients identified 31 potential risks for the use of technology in healthcare. They feared that it could inadequately replace human intelligence in care (28%), represent serious risks for hacking (13%), or lead to misuse of private patient data by caregivers, insurance companies, etc. (14%) (Fig. 1, Table 3, Supplementary Table 3). For example, a patient stated that: “[we will need] to be extra careful about personal data. There are risks or drawbacks if some information is disclosed to social networks, banks, insurance or work. It will be necessary for patients to be educated on that.” (60-year-old woman reporting rheumatoid arthritis, high blood pressure, and hypercholesterolemia).
Overall, 20% participants considered that the potential benefits of technology greatly outweighed its potential dangers (opportunity > 7/danger < 3), whereas only 3% felt that negative aspects outweighed potential benefits (opportunity < 3/danger > 7) (Supplementary Table 1).
Assessment of patients’ readiness to integrate interventions using BMDs or AI in their own care
Figure 2 and Supplementary Fig. 1 present patients’ readiness to accept four available interventions that use BMDs and AI in healthcare: (1) patients’ skin photographs and AI to screen for skin cancer;10,22 (2) wearable sensors for continuous and real-time monitoring and the analysis of data collected by AI to predict flares of their chronic conditions;14 (3) a smart shirt and AI to guide physical therapy;23 and (4) an AI chatbot to help patients determine how urgent their problems are.24 Approximately 20% of patients with chronic conditions were opposed to the use of BMDs and AI-based tools in their care in all presented situations. This proportion ranged from 17% for the use of patients’ skin photographs and AI to screen for skin cancer to 21% for the use of a smart shirt and AI to guide physical therapy. Accordingly, about 80% of participants were ready for the use of technology in their care. However, only a fraction of these patients were ready for the use of AI without human control (from 10% for the use of patients’ skin photographs and AI to screen for skin cancer to 36% for the use of AI chatbots to assist in determining how urgent their problems were).
We identified six clusters—or patient profiles—regarding patients’ readiness for the interventions described in the four vignettes (Fig. 3). First, 13% of participants were globally against any use of BMDs and AI (Cluster 1). Second, 22% of participants would refuse the use of BMDs and AI in one of the different situations (Clusters 2–4). Thus, only 65% of patients would agree with the integration of all interventions presented in their care. Among them, 65% (41% of the total population) would only accept BMDs and AI if their use was controlled by humans (Cluster 5), while 35% (22% of the total population) were ready for some level of automation in their care, even without human control, especially for the use of smart clothes and AI to guide physical therapy or for AI chatbots to answer emergency calls (Cluster 6). Only 3% of patients would agree with full automation of care processes for all four vignettes presented.
We found no clear association between patients’ demographic or clinical characteristics and their readiness to use BMDs and AI-based tools in their care (Supplementary Table 4).
In this study, we report two experiments to document the perception of patients with chronic conditions on the use of BMDs and AI in care. Only 50% of patients felt that the development of digital tools and AI in healthcare was an important opportunity and 11% considered it a danger. In particular, patients feared that the misuse of technology would lead to unwanted replacement of humans and threaten the humanistic aspect of health and care. In the second part of our study, we showed that 35% of patients would refuse to integrate at least one existing or soon-to-be available interventions using BMD- and AI-tools in their care. In addition, only a minority of participants were ready to integrate fully automatic AI-based tools in their care. Our results may explain the high drop-out of participants in the first large-scale implementations of digital monitoring strategies (90% incomplete follow-up for MyHeart Counts and 55% incomplete follow-up data for the Healthy Pregnancy Research Program).25,26,27
Our results highlight that patients intuitively think that AI should help clinicians “predict” outcomes, but that decisions, actions, and recommendations should remain a human task. Technology would be like as a “driver assistance” for clinicians.12 Even among patients who were the most ready for the use of technology in their care, they would only see AI as a complement—and not as replacement—for human care for situations related to sensitive topics (cancer) or which involved lasting interventions (monitoring for chronic conditions).
Our study has several strengths. As of today, it is the most comprehensive description of patients’ perceptions of the potential benefits and risks regarding the use of BMDs and AI interventions in healthcare and their readiness to integrate these technologies in their own care. Our findings were strengthened by the use of robust methods for both the quantitative and qualitative part of our study. Vignettes have been shown to accurately reflect participants’ preferences and real decisions;28 questionnaires with open-ended questions enabled the exploration of the broad range of patients’ perspectives regarding the use of BMDs and AI in their care without the preconceptions that could have arisen in a “classical” survey. Our study also benefited from a large sample size with wide diversity in patient characteristics (age, educational level, conditions, multimorbidity, use of m-health or e-health, etc.).
Our study has some limitations. First, despite statistical calibration, results from this study must be extrapolated with caution. Indeed, primary data were from a population of patients engaged in a citizen science initiative to accelerate research on their chronic conditions, so they may be more enthusiastic to use technology in healthcare. Second, we assessed whether patients would be ready to use digital tools and AI in their care by using clinical vignettes that were voluntarily simple. Especially, the vignettes did not detail the exact modalities of the use of technology (e.g., How many consultations could be prevented? How many alerts would the patient and/or his/her physician receive? Who would store and/or have access to collected data?). We chose this simple format because we aimed at getting the “feeling” of whether patients were ready or not for the use of BMDs and AI in their care rather than just assessing how they would accept specific interventions. Moreover, in practice, the decision to use digital tools or AI for the care of a given patient would require a discussion between the patient and the physician(s) on perceived advantages, barriers, and fears. As for any therapeutic intervention, the decision is not just that of the patient but rather a shared decision-making process with the patient–clinician dyad.
The literature on patients’ perceptions of the use of BMDs and AI in care is scarce. First, patient-reported data (qualitative or quantitative) were collected in some studies evaluating digital technologies and AI-based tools.13,29,30 However, these results are specific to both a given intervention and a given context and do not reflect patients’ uptake of these interventions if they had to be scaled up.20 Second, a handful of studies have explored patients’ perceptions of wearable devices and IA outside of the context of an ongoing digital-tool evaluation study.19,20,21,31,32,33,34 However, these studies were often limited in sample size and participant diversity or focused on a specific subject. Finally, to our knowledge, only one study broadly assessed patients’ perceptions of the use of AI in healthcare. This study from Syneos Health Communications involved 800 patients with atrial fibrillation, type 2 diabetes mellitus, and breast cancer and showed that 16–19% of participants were “excited” about use of AI in healthcare and 32 to 42% were “unexcited”.35 Our results provide the largest and most comprehensive view of chronic patients’ perspectives of the use of these technologies in healthcare.
Healthcare systems in high-income countries such as France strive to care for patients with chronic conditions within overburdened practices and consultations constrained to short visits.36,37 There is a mismatch between what care systems can and need to deliver.38 Therefore, many clinicians, researchers and decision makers are looking to BMDs and AI to find the “magic bullet” to transform healthcare. Although the challenges of quality and safety regarding the use of AI in care have already been noted (distributional shift, automation complacency, reward hacking, unscalable oversight, etc.),39 the perspective of patients has often been neglected or forgotten. Our results emphasize that patients are not ready for fully automated care. This perspective must be taken into account to avoid unjustified AI hype and to accurately assess the potential impact of implementing BMD- and AI-based interventions, at scale. For full potential of these interventions, device manufacturers, prescribing clinicians, care organizations and regulation authorities will need to account for patient-reported benefits and perceived risks, as we identified.
The number of studies evaluating new BMD- or AI-based tools is rapidly increasing and their costs are decreasing. The current literature focuses on the technological aspects of these tools but neglects patients’ perspectives of their use in healthcare. In this paper, we showed that most patients would agree to use these new technologies in their care if controlled by human caregivers. These findings call for a novel reflection about how technology should be integrated in care processes to avoid a negative impact on patient care, the generation of unnecessary burdens or the intrusion in their lives.
This study involved two complementary parts. First, we mixed quantitative and qualitative methods to understand the potential benefits and dangers of the use of BMDs and AI in healthcare as perceived by patients with chronic conditions. Second, we used vignettes to assess chronic patients’ readiness to integrate specific interventions involving these new technologies in their care.
Setting and participants
Participants were recruited within the “Community of Patients for Research” (ComPaRe), an ongoing citizen science project based on an e-cohort of patients with chronic conditions, in France. Participants of “ComPaRe” are adults (>18 years old) who report having at least one chronic condition (defined as a condition requiring healthcare for at least 6 months). Patients join the project to donate time to accelerate research on their conditions by answering regular patient-reported outcomes and patient-reported experience measurements online, suggesting ideas for new research, or participating in the set-up or analysis of research projects.40 All participants provide electronic consent before participating in the e-cohort. ComPaRe was approved by the Institutional Review Board of Hôtel-Dieu Hospital, Paris (IRB: 0008367).
To enhance the external validity of our results, our data were calibrated on margins for sex-specific weights for age and educational level derived from national census data for the French population reporting at least one chronic condition.41,42 The main study results are presented with the recalibrated data. Raw results are provided as supplementary tables and figures.
Description of patients’ perceptions of the use of BMDs and AI in healthcare
A sequential explanatory mixed-methods design was used to explore patients’ perceptions of the use of BMDs and AI in healthcare.43 The purpose of the sequential explanatory mixed methods design was to use qualitative methods to inform the analysis of the initial quantitative results and develop a comprehensive understanding of why patients could perceive the use of BMDs and AI in care as an opportunity or danger.
First, patients rated their perceptions of the use of these new technologies by answering two structured questions “Do you think that the increasing use of digital technologies, BMDs, and AI in healthcare is an opportunity/a danger?” Ratings were collected with numeric scales ranging from 0 (no opportunity/danger) to 10 (great opportunity/danger). Answers to these questions were described by mean (SD) scores and categorized to identify participants seeing BMDs and AI as a great opportunity/danger (rating > 7/10) or small opportunity/danger (rating < 3/10).
Then, patients supplemented their numeric evaluations with open-ended comments on the benefits and risks they perceived regarding the use of these new technologies. Open-text data were evaluated by thematic analysis without the use of a specific theoretical lens. Indeed, our study was not meant to explain but rather to inventory perceptions from participants. To summarize, our analysis involved (1) the extraction by two investigators (VTT and CR), in double, of “in vivo codes”: literal terms used by participants to explain and describe their perceptions regarding the use of BMDs and AI in healthcare; (2) the comparison of these codes in order to recognize and group those that were similar, based on the context, people and processes involved; and (3) the creation, during regular meetings between the investigators, of a consensual and stable classification for codes expressing similar domains and grouping them into larger categories.
Assessment of patients’ readiness to integrate BMDs or AI in their own care
In a second part, we presented participants with four “vignettes”—or systematically elaborated descriptions of concrete situations aimed at examining decision-making processes44—about existing or soon-to-be available interventions using BMDs and AI in healthcare. The four situations evaluated were the use of (1) patients’ skin photographs and AI to screen for skin cancer rather than consultations with a dermatologist;10,22 (2) wearable sensors for continuous and real-time monitoring and the analysis of data collected by AI to predict flares of their chronic conditions rather than usual follow-up (doctor visits, tests, etc.);14 (3) a smart shirt and AI to guide physical therapy rather than visits to a physiotherapist;23 and (4) an AI chatbot to help patients determine how urgent their problems are rather than calling an emergency telephone number.24 For each of these situations, participants were asked to evaluate their readiness to switch from current care to the new intervention with the question “If there were solid scientific evidence that the [new BMD or AI intervention] would be equivalent or better than [the current standard of care] in the given situation, would you agree to use the new intervention in your own care?” Participants’ answers ranged from −3 (would not use the intervention) to 0 (would only use the new intervention if controlled by a human caregiver), and +3 (would use the intervention and it could replace some interventions currently implemented by human caregivers).
We assessed whether homogenous groups of patients with similar readiness to use BMDs and AI could be identified from our data. For this, we considered participants’ responses to each vignette as a continuous value ranging from −3 to +3 and clustered patients according to their answers to the four vignettes with a k-means algorithm taking into account the weighted structure of the calibrated dataset. Clusters were described by both demographic and clinical characteristics (age, sex, educational level, multimorbidity, duration since the diagnosis of the first chronic condition, and previous use of e-health or m-health technology).
Analyses involved use of R v3.3 (http://www.R-project.org, the R Foundation for Statistical Computing, Vienna, Austria).
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
All data in ComPaRe are available to academic research teams provided a scientific protocol approved by the ComPaRe scientific committee.
The code used to process the dataset (R version 3.3 script) is available from the corresponding author on reasonable request.
Topol, E. J., Steinhubl, S. R. & Torkamani, A. Digital medical tools and sensors. Jama 313, 353–354 (2015).
Elenko, E., Underwood, L. & Zohar, D. Defining digital medicine. Nat. Biotechnol. 33, 456–461 (2015).
Arneric, S. P. et al. Biometric monitoring devices for assessing end points in clinical trials: developing an ecosystem. Nat. Rev. Drug Discov. 16, 736 (2017).
Fagherazzi, G. & Ravaud, P. Digital diabetes: perspectives for diabetes prevention, management and research. Diabetes Metab. doi:10.1016/j.diabet.2018.08.012 (2018).
Hinton, G. Deep learning-a technology with the potential to transform health care. Jama 320, 1101–1102 (2018).
Steinhubl, S. R., Muse, E. D. & Topol, E. J. Can mobile health technologies transform health care? Jama 310, 2395–2396 (2013).
Topol, E. J. Transforming Medicine via Digital Innovation. Sci. Transl. Med. 2, 16cm14 (2010).
Wicks, P. et al. It’s a long shot, but it just might work! Perspectives on the future of medicine. BMC Med. 14, 176 (2016).
Haenssle, H. A. et al. Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann. Oncol. 29, 1836–1842 (2018).
Esteva, A. et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature 542, 115–118 (2017).
Lakhani, P. & Sundaram, B. Deep learning at chest radiography: automated classification of pulmonary tuberculosis by using convolutional neural networks. Radiology 284, 574–582 (2017).
Topol, E. J. High-performance medicine: the convergence of human and artificial intelligence. Nat. Med. 25, 44–56 (2019).
Bally, L. et al. Closed-loop insulin delivery for glycemic control in noncritical care. New Engl. J. Med. 379, 547–556 (2018).
Burnham, J. P., Lu, C., Yaeger, L. H., Bailey, T. C. & Kollef, M. H. Using wearable technology to predict health outcomes: a literature review. J. Am. Med. Inform. Assoc.: JAMIA. https://doi.org/10.1093/jamia/ocy082 (2018).
Yates, T. et al. Association between change in daily ambulatory activity and cardiovascular events in people with impaired glucose tolerance (NAVIGATOR trial): a cohort analysis. Lancet (Lond., Engl.) 383, 1059–1066 (2014).
Pyrkov, T. V. et al. Extracting biological age from biomedical data via deep learning: too much of a good thing? Sci. Rep. 8, 5210 (2018).
Gresham, G. et al. Wearable activity monitors to assess performance status and predict clinical outcomes in advanced cancer patients. npj Digital Medicine 1, 27 (2018).
Turakhia, M. P. et al. Rationale and design of a large-scale, app-based study to identify cardiac arrhythmias using a smartwatch: The Apple Heart Study. Am. heart J. 207, 66–75 (2019).
Lennon, M. R. et al. Readiness for delivering digital health at scale: lessons from a longitudinal qualitative evaluation of a national digital health innovation program in the United Kingdom. J. Med. Internet Res. 19, e42 (2017).
O’Connor, S. et al. Understanding factors affecting patient and public engagement and recruitment to digital health interventions: a systematic review of qualitative studies. BMC Med. Inform. Decis. Mak. 16, 120 (2016).
Mosconi, P., Radrezza, S., Lettieri, E. & Santoro, E. Use of health apps and wearable devices: survey among italian associations for patient advocacy. JMIR Mhealth Uhealth 7, e10242 (2019).
Fujisawa, Y. et al. Deep learning-based, computer-aided classifier developed with a small dataset of clinical images surpasses board-certified dermatologists in skin tumor diagnosis. Br. J. Dermatol. https://doi.org/10.1111/bjd.16826 (2018).
Bobin, M., Amroun, H., Anastassova, M., Boukallel, M. & Ammi, M. in IEEE International Conference on Systems, Man, and Cybernetics (SMC2017).
Babylon. Babylon chatbot, https://www.babylonhealth.com/ (2017).
Krummel, T. M. The Rise of Wearable Technology in Health Care. JAMA Netw. open 2, e187672 (2019).
Radin, J. et al. The Healthy Pregnancy Research Program: transforming pregnancy research through a ResearchKit app. npj Digital Medicine 1, 45 (2018).
McConnell, M. V. et al. Feasibility of obtaining measures of lifestyle from a smartphone app: the MyHeart Counts Cardiovascular Health Study. JAMA Cardiol. 2, 67–76 (2017).
Hainmueller, J., Hangartner, D. & Yamamoto, T. Validating vignette and conjoint survey experiments against real-world behavior. Proc. Natl Acad. Sci. USA 112, 2395–2400 (2015).
Keel, S. et al. Feasibility and patient acceptability of a novel artificial intelligence-based screening model for diabetic retinopathy at endocrinology outpatient services: a pilot study. Sci Rep. 8, 4330 (2018).
Kim, R. H. & Patel, M. S. Barriers and opportunities for using wearable devices to increase physical activity among veterans: pilot study. JMIR Form. Res. 2, e10945 (2018).
Messer, L. H., Johnson, R., Driscoll, K. A. & Jones, J. Best friend or spy: a qualitative meta-synthesis on the impact of continuous glucose monitoring on life with Type 1 diabetes. Diabet. Med.: a J. Br. Diabet. Assoc. https://doi.org/10.1111/dme.13568 (2017).
Daus, H., Kislicyn, N., Heuer, S. & Backenstrass, M. Disease management apps and technical assistance systems for bipolar disorder: Investigating the patients point of view. J. Affect. Disord. 229, 351–357 (2018).
Fensli, R. & E., B. in International Joint Conference on Biomedical Engineering Systems and Technologies (BIOSTEC 2008). (Springer).
Bergmann, J. H., Chandaria, V. & McGregor, A. Wearable and implantable sensors: the patient’s perspective. Sens. (Basel, Switz.) 12, 16695–16709 (2012).
Syneos Health Communications. The Unheard Voice, https://syneoshealthcommunications.com/perspectives/artificial-intelligence-for-authentic-engagement.
Mechanic, D., McAlpine, D. D. & Rosenthal, M. Are patients’ office visits with physicians getting shorter? New Engl. J. Med. 344, 198–204 (2001).
Elmore, N. et al. Investigating the relationship between consultation length and patient experience: a cross-sectional study in primary care. Br. J. Gen. Pract.: J. R. Coll. Gen. Pract. 66, e896–e903 (2016).
Kvedar, J. C. & Fogel, A. L. mHealth advances clinical research, bit by bit. Nat. Biotechnol. 35, 337–339 (2017).
Challen, R. et al. Artificial intelligence, bias and clinical safety. BMJ Qual. Saf. https://doi.org/10.1136/bmjqs-2018-008370 (2019).
ComPaRe. Community of Patients for Research, http://www.compare.aphp.fr (2018).
Institut National de la Statistique et des etudes économiques. La macro SAS CALMAR, 2018).
Direction de la recherche, d. é., de l’évaluation et des statistiques. L’état de santé de la population en France - RAPPORT 2017. (Ministère des Solidarités et de la Santé - République Française, Paris, 2017).
Creswell, J. & Clark, V. Designing and conducting mixed methods research. (SAGE Publications, 2011).
Alexander, C. & Becker, H. The use of vignettes in survey research. Public Opin. Q. 42, 93–104 (1978).
The authors thank Isabelle Pane, Elise Diard, Solene Gouesbet, Razak Ellafi and all members of the ComPaRe team for their help in data collection; Laura Smales (BioMedEditing, Toronto, Canada) for manuscript editing; and Elise Diard for help in drafting figures.
The authors declare no competing interests.
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Tran, VT., Riveros, C. & Ravaud, P. Patients’ views of wearable devices and AI in healthcare: findings from the ComPaRe e-cohort. npj Digit. Med. 2, 53 (2019). https://doi.org/10.1038/s41746-019-0132-y
This article is cited by
Patient perspectives on data sharing regarding implementing and using artificial intelligence in general practice – a qualitative study
BMC Health Services Research (2023)
Artificial intelligence-based clinical decision support in pediatrics
Pediatric Research (2023)
Unsupervised IMU-based evaluation of at-home exercise programmes: a feasibility study
BMC Sports Science, Medicine and Rehabilitation (2022)
Moral exemplars for the virtuous machine: the clinician’s role in ethical artificial intelligence for healthcare
AI and Ethics (2022)
Exploring perceptions of healthcare technologies enabled by artificial intelligence: an online, scenario-based survey
BMC Medical Informatics and Decision Making (2021)