Abstract
Mobile phone applications (“apps”) are potentially an effective, low-burden method to collect patient-reported outcomes outside the clinical setting. Using such apps consistently and in a timely way is critical for complete and accurate data capture, but no studies of concurrent reporting by cancer patient–caregiver dyads have been published in the peer-reviewed literature. This study assessed app engagement, defined as adherence, timing, and attrition with two smartphone applications, one for adult cancer patients and one for their informal caregivers. This was a single-arm, pilot study in which adult cancer patients undergoing IV chemotherapy or immunotherapy used the DigiBioMarC app, and their caregivers used the TOGETHERCare app, for approximately one month to report weekly on the patients’ symptoms and wellbeing. Using app timestamp metadata, we assessed user adherence, overall and by participant characteristics. Fifty patient–caregiver dyads completed the study. Within the one-month study period, both adult cancer patients and their informal caregivers were highly adherent, with app activity completion at 86% for cancer patients and 84% for caregivers. Caregivers completed 86% of symptom reports, while cancer patients completed 89% of symptom reports. Cancer patients and their caregivers completed most activities within 48 h of availability on the app. These results suggest that the DigiBioMarC and TOGETHERCare apps can be used to collect patient- and caregiver-reported outcomes data during intensive treatment. From our research, we conclude that metadata from mobile apps can be used to inform clinical teams about study participants' engagement and wellbeing outside the clinical setting.
Similar content being viewed by others
Introduction
There is increasing awareness that remote health monitoring can be beneficial during a clinical trial1,2,3 and for routine health care4,5,6. Smartphone applications (“apps”) that help patients track symptoms and general health through electronic patient-reported outcomes (ePROs) are potentially an effective and low-burden method to collect trial and healthcare data outside the clinical setting7,8,9,10,11,12. Longitudinal data capture via a smartphone app may be a powerful approach to monitor patient progress and/or adverse events in clinical trials1,2 or usual care13,14 without the need for frequent in-person visits to a clinic.
In this relatively new area, little has been documented regarding what works best to encourage participants to consistently use mobile app interventions15,16,17. High-quality data depends on the user to regularly enter accurate information; incomplete questionnaires can compromise the quality of ePRO data18,19.
The term “engagement” has been used to discuss the extent of mobile app utilization by users20. Although engagement has not been consistently measured15, it is primarily conceptualized with three components: behavioral, cognitive, and affective21. The behavioral component—the focus of this article—is generally measured quantitatively through physical interaction of the user with the mobile health system, including metadata captured by the app, adherence with app activities, number of log-ins, time spent on each activity, and number of features accessed and screens viewed22,23. Semi-structured interviews were completed in previously published usability studies of each app; these interviews addressed some of the cognitive and affective components of engagement24,25.
This project developed and tested DigiBioMarC24 a smartphone app for individuals with cancer (“patients”), and TOGETHERCare25, a smartphone app for informal caregivers, to collect health, activity, and psychosocial data about patients and caregivers frequently and easily.
Cancer treatment may result in adverse events that require dose reductions or other interventions, making frequent and timely reporting of patient symptoms important26. There is a large body of literature about the use of mobile health apps by patients to report their health status27,28, but to date no other studies of concurrent reporting by cancer patient–caregiver dyads have been published in the peer-reviewed literature.
The aim of this article is to explore behavioral engagement with apps that collect ePROs by reporting on the adherence of both patients and caregivers, focusing on whether cancer patients and caregivers completed app activities as planned, the amount of time spent on expected activities, and associations of participants’ characteristics to explore app burden and potential obstacles to engagement.
Methods
Study design and sample
This is a single-arm, prospective study in which adult cancer patients within the Kaiser Permanente Northern California (KPNC) system and their informal caregivers were asked to use the DigiBioMarC (for cancer patients) or TOGETHERCare (for caregivers) smartphone apps for 28 days after enrollment (the study period). KPNC is an integrated healthcare delivery system serving over 4.5 million members representative of the general Northern California population29. The KPNC Institutional Review Board reviewed and approved the study protocol.
Cancer patients scheduled to receive intravenous chemotherapy or immunotherapy during the study period, were identified using the KPNC electronic health record system and were recruited by email invitations from October 2020 to March 2021. Recruitment emails were sent to 2155 potential patients. Of the 247 respondents (11% of recruitment emails), 166 (67% of respondents) were determined to be ineligible (no iPhone (n = 33), no eligible/willing caregiver (n = 24), no scheduled IV therapy during the study (n = 13), not English speaking (n = 5), physician indicated a contradiction to participation (n = 42), invalid emails (n = 41), deceased (n = 2), ineligible unspecified (n = 6)), 20 declined to participate before eligibility could be confirmed, and 7 declined to participate after learning more details about the study. No background information was tabulated for those that did not consent to participate in the study as we were restricted under our IRB approved protocol to information from the EHR only for those individuals that provided informed consent and HIPAA authorization.
Recruitment and all methods were performed in accordance with the relevant guidelines and consent were completed remotely through a video or phone call. During this call, the protocol coordinator assisted interested and eligible participants in downloading the app and instructed them on how to review and sign the informed consent document in the app. No other training was provided in how to navigate the applications. Participants were required to provide written informed consent within the application prior to proceeding with the study. Both patients and caregivers who continued responding to activities in the app through the study period and completed the two semi-structured virtual interviews received a $100 gift certificate for their time and effort. Based on the IRB recommendations, data collected through the app were not relayed to the participating patients’ or caregivers’ clinical team. The exception was that the research associate was to be automatically alerted to survey scores for mental health measures from specific surveys if they exceeded a predetermined threshold. If an alert was received, the research associate was to call the participant, ask if they were in immediate danger and if so call 911 to request the police conduct a “health and welfare check.” If not in immediate danger but they expressed distress, the research assistant was to ask if they wanted a Kaiser clinician to talk with them.
Eligibility requirements for cancer patients included being 18 years of age or older, receiving intravenous chemotherapy or immunotherapy treatments at a KPNC medical facility at the time of the study, having an iPhone 6 or higher, and the ability to read and communicate in English24. Patients were also required to have an informal caregiver to be eligible to participate in the study. The informal caregivers were identified by the patient and screened for eligibility during recruitment. Caregivers did not need to be KPNC members but did need to be 18 years of age or older, have an iPhone 6 or higher, and able to read and communicate in English. There was no requirement for the length of time the caregiver had lived with the patient. Cancer patients with severe mental illness or insufficient cognition to consent (as determined by their physician) were not eligible.
App activities and schedule
As previously described in greater detail24,25, the DigiBioMarC and TOGETHERCare apps were composed of “activities” defined as surveys as well as physical assessments. Surveys assessed the cancer patients’ symptoms and functioning, social and financial resources, wellbeing, and stress, among other measures from the perspective of the cancer patient (DigiBioMarC) and their caregiver (TOGETHERCare). Physical assessments, gait speed30 and “sit to stand”31 tasks, were to be completed by all patient participants but were not asked of caregivers.
Each time an app activity was made available to an individual participant, it counted as an “instance” of that activity. The number of instances varied for different activities. The study protocol expected, but did not mandate, participants to complete the activity within 24 h. Some activities were expected for completion only at the beginning of the study period (i.e., one instance), some at the beginning and end (i.e., two instances), and some were scheduled for completion either every 7 or 14 days (i.e., three to five instances). The intervals between instances of an activity were standardized such that after a user completed a planned activity, it could not be completed again before the minimum expected days had elapsed, depending on the planned interval. The activity remained open until it was completed. The frequency and timing of the app activities were intended to capture relevant health and wellbeing data pertaining to cancer patients. Research staff did not prompt participants by phone, email, or text to complete open activities. Instead, notifications or reminders were sent within the app to caregiver participants three times per week, and to patients every two to three days, midday on a preset schedule. These reminders were not linked to specific activities or whether the participant had completed them but were instead general reminders to complete any open activities.
While the active study duration was 28 days, for the purposes of this examination, we allowed activities to be completed within a 33-day window past enrollment to allow a few extra days for users to become familiar with the apps and study expectations.
Measures
Adherence. Using the timestamp data (app operational metadata), the study assessed app adherence as: (1) completion of all expected app activities within 33 days; (2) completion of individual app activities; (3) completion of expected app activities within 24 and 48 h; and (4) average time to complete individual app activities at the beginning and end of the study.
Based on the expected instances, we calculated denominators for the percentages of app activities completed. The numerators for these percentages are the total completions (including all study participants within each group) over the entire active study period.
We measured the app time commitment as the minutes to complete each activity based on metadata, changes in completion time over the active study period, and total time at the beginning and end of the study period, when the greatest number of activities were expected.
Potential obstacles to app adherence. To provide contextuality of the participants’ adherence and potential app burden, we collected information from the patients’ KPNC medical records about the patients’ cancer status, namely, type, stage, and presentation of cancer at study enrollment. In the apps themselves, we collected information about the participants’ social demographics (age, gender, ethnicity, race, educational attainment, employment status) and the caregivers’ perceived amount of time that they provided care at the beginning and end of the study period. We conducted semi-structured interviews with participants on approximately the seventh day of app use and again after participants completed the final app surveys, using a secure videoconferencing program. The results of these surveys have been previously reported24,25.
Analysis
The analysis is descriptive and exploratory. First, we summarized the baseline characteristics of the study participants. Second, we calculated summary statistics for the adherence measures and the average minutes spent on each activity. Third, we performed two-sided Fisher Exact tests to assess possible associations between the participant characteristics and completion of 80 + percent of the app activities within the active study period.
Results
Study participant baseline characteristics
Fifty of the 54 enrolled patient–caregiver dyads completed the study (3 patients stopped participating due to serious declines in health and one patient did not start the planned treatment making them ineligible). Seventy-six percent of the patients had advanced cancer (stage 3 or 4), and 28% had relapsed diagnosis as of their enrollment date. The caregivers in the enrolled dyads were predominantly male (62%) and spouses/partners of the cancer patients (78%), who were predominantly female (78%) (Table 1). Eighty percent of the patients were 50 years of age or older with about a quarter of them ≥ 70 years. The caregiver cohort was overall younger with 50% aged 50–69 years. Close to two-thirds of the participants (combined patients and caregivers) were white, and over 80% identified as non-Hispanic or Latino. Fifty-four percent of the caregivers and 70% of the patients had an associate degree or higher, respectively. Almost all caregivers (n = 48) and cancer patients (n = 49) had “adequate” cancer health literacy as indicated by the 6-Item Cancer Health Literacy Test (CHLT-6). Unemployment at the start of the study was higher in patients than in caregivers (68% vs 42%, respectively).
Overall completion of expected app activities
The mean number of activities completed during the active study period was 24.2 out of 28, (standard deviation [SD] = 4.4) for patients (86% completion rate) and 35.6 out of 42 (SD = 7.1) for caregivers (84% completion rate) (Supplement 1). Seventy-six percent (76%) of patients and 66% of caregivers completed 80% or more of the expected activities, respectively. More activity completion in the caregiver group was related to more activity completion in the patient group (correlation coefficient = 0.5, p ≤ 0.001). No participant survey responses met the threshold to trigger a mental health email notification to the research assistant.
Completion of individual app activities
One hundred percent of the “About You” surveys were completed by all study participants (Table 2). For the caregivers’ repeated surveys, the “Every Other Day” survey was least completed (77.1% of the expected surveys were completed) and the “Quick Check-in” (PHQ4) survey was most completed (93.0% were completed). For the cancer patients’ repeated activities, the “Gait Speed” assessment was the least completed (79.6% completion rate) and, again, the “Quick Check-in” survey was the most completed (95.0% completion rate). For both caregivers and patients, the completion rate of more frequent activities was lower compared with less frequent ones.
Ninety percent or more of the “About You” surveys were completed within 24 h from the time it became available for both patients and caregivers (Table 2). Lower percentages of the weekly activities were completed within 24 h. For example, 60% of the “Short Symptom Reports” (PRO-CTCAE) surveys were completed by caregivers, and 52% of the “Gait Speed” and “Sit and Stand” physical assessments were completed by cancer patients within 24 h. Expanding the completion window to 48 h increased the percentage completed by a range of 2.2% (caregivers’ “About You”) to 28.1% (caregivers’ “COVID-19 Patient Impact”). The high completion rate of the “Every Other Day” survey by caregivers at the onset of the study began to decline starting at the fourth of 14 expected surveys. By the ninth survey, only 76% of the caregivers completed it (Supplement 2).
Time to complete individual app activities
The average time it took participants to complete the app activities are provided in Table 3. The average time for the cancer patients to complete app activities at baseline ranged from 0.8 min (SD = 0.5) for the “Fast 4” survey to 4.9 min (SD = 1.8) for “COVID-19 Your Input” survey. Except for the caregivers’ time to complete “App Feedback,” the time to complete app activities declined slightly over time. The time of day when app activities were completed is included in Supplement 3.
Completion of app activities by background characteristics
Although none of the caregivers’ or patients’ characteristics, including age, gender, and race, were significantly associated with completing 80% or more of the app activities during the active study period (Supplement 4), we observed some trends. Adherence was overall high across all age categories among patients and reached 90% in patients ≥ 70 years of age. Notably, while the lowest rate adherence was seen among patients aged 50–69, caregivers in this age group were the most compliant (75%). Higher levels of education (associate degree or higher) were marginally (p = 0.0592) associated with lower completion of app activities during the active study period.
Caregiving time commitment also was marginally associated with study adherence at both baseline and end of study. When collapsing the reported caregiving time commitment by caregivers to < 20 h versus ≥ 20 h per week, a different trend emerged; at baseline, the rate of adherence by caregivers who committed < 20 h of caregiving was lower than for those who committed ≥ 20 h (59% vs 73%, respectively). This trend became statistically significant at the end of the study, with 92% adherence among caregivers who committed ≥ 20 h caring for their patients compared with only 56% among caregivers committing < 20 h of care (p = 0.034, data not shown).
Adherence among cancer patients with late-stage disease (3 and 4) was higher than adherence among patients diagnosed with early-stage disease (81.6% vs. 58.3%, respectively) (Supplement 4). Inversely, adherence was higher among caregivers of patients with early-stage disease (83.3%) compared to caregivers of patients with late-stage disease (60.5%).
Discussion
In summary, timestamp metadata analysis showed that both adult cancer patients and their informal caregivers were highly adherent over the one-month study duration, with app activity completion at 86% for cancer patients and 84% for caregivers. The percentage of individual activities completed ranged from 91 to 100% for caregivers, and from 79.6% to 100% for patients. The activities that were the least completed by patients were the active physical assessments, gait speed and the sit to stand activity. We suspect these were difficult for patients to complete when they were not feeling well. Cancer patients often felt fatigued, 94% and 80% at the beginning and end of the study, respectively. The surveys that were least completed by caregivers over time were the “Every Other Day” surveys. We believe these surveys might have been too frequent, leading to caregiver burnout on completing them, or the caregivers saw reporting on their own status as less relevant than that of the cancer patient.
Patients and caregivers showed a comparable overall adherence rate (> 85%) for the weekly PRO-CTCAE and PROMIS Physical Function surveys, suggesting the importance of remote monitoring of patients’ symptoms and side effects to patients and caregivers alike. Adherence increased when we allowed 48 h between when the activity appeared in the app and completion, which suggests that whenever possible, it is important to provide a reasonable reporting time window to allow for activity completion. The average time it took caregivers to complete each app activity ranged from less than a minute to 5.8 min; the average time for patients ranged from less than a minute to 4.9 min. Our results do not support the perception that older patients are less compliant with digital health technologies, as the highest (90%) adherence in cancer patients was among the oldest group (≥ 70 years) of participants. This is consistent with a high level of adherence to using physical trackers found in elderly patients with atrial fibrillation32.
Interpreting our findings of results in light of known literature, cancer patient adherence with our app-based approach for treatment remote monitoring falls within the range of other studies of electronic clinical outcomes reporting, including 64.7% average patient adherence for weekly reporting during chemotherapy33, 74% median adherence for reporting three times per week during treatment with immune checkpoint inhibitors7, and 83% adherence for daily reporting in patients with breast or prostate cancer34. The United Kingdom electronic self-reporting of adverse events for patients undergoing cancer treatment eRAPID program found 72.2% patient adherence to adverse-event symptom reporting13.
The importance of our results is that this study enhances our knowledge about app engagement by demonstrating that metadata can be utilized to inform clinical teams about engagement behaviors of their study participants related to remote data capture and monitoring. This rapid assessment of study activities does not require accessing and analyzing clinical outcome data. Instead, metadata can be used to generate insights on risks of non-adherence during clinical trial data collection or during standard treatment.
There are several limitations of this study. First, we had a relatively small sample size (n = 50 patient–caregiver dyads), so the study was not powered for testing hypotheses about predictors of and barriers to app usage. In addition, only those patients who responded to the invitation to participate in the study were included, and these patients may have been more likely to be adherent. Second, notifications were sent on set dates and times, irrespective of whether activities were completed. The apps have since been updated to coordinate notifications with when activities are due. Third, this pilot study was not designed to test the clinical workflow related to ePROs collected by the apps. The app report was not integrated into the EHR and responses did not go to healthcare providers for review. Further research is needed to assess the feasibility and effect of transmitting urgent data from the apps to healthcare providers. Fourth, there are two recent studies comparing Kaiser membership with the general population as described in the California Health Interview Survey35, and one comparing with the US Census and Behavioral Risk Factor Surveillance System29. Both found that compared to the total population in the catchment area, the Kaiser members were similar in race, ethnicity, education and income. However, in this study, we did not attempt to achieve a representative sample of participants. The eligibility criteria for the study skewed the population towards females because breast cancer is frequently treated with IV chemotherapy. We suspect that recruitment via email and the requirement that study participants had to own an iPhone 6 or higher and be English speakers may have skewed the representativeness of participants by race/ethnicity. We have since developed an Android version. Finally, our month-long study was not long enough to demonstrate sustained adherence over longer courses of IV chemotherapy.
This study’s strengths included its demonstration of adherence with mobile health apps in a population undergoing active cancer treatment, with patients receiving IV chemotherapy or immunotherapy—therapies that frequently result in side effects. The sample included patients with a mix of cancer types. Observations were reported by patients through our DigiBioMarC app, and by their caregivers through the TOGETHERCare app. This provides the advantage of a second set of eyes on the patient’s symptoms, potentially providing an additive advantage in accuracy and completeness to improve remote monitoring abilities.
Future directions include incorporating more diverse participants across socio-economic, racial, ethnic, and health and digital literacy levels. This will help us examine whether remote monitoring of this type can reduce persistent health disparities by removing requirements to go to specific study locations, which may not be easily accessible with limited resources or from rural areas. We would also like to compare adherence to mobile apps with emailed or texted ePROs. In addition, based upon multiple responses in the semi-structured interviews, we hypothesize that adherence and engagement will be enhanced if clinicians receive the patient reported outcome data. Some of the team plans to explore how to implement this workflow in a frictionless manner within the EHR, with flags for PROs or activities of concern.
Conclusions
The high rate of participant adherence (86% for cancer patients and 84% for caregivers) in this timestamp metadata analysis suggests that the DigiBioMarC and TOGETHERCare apps can be used to collect patient- and caregiver-reported outcomes data during intensive treatment.
This study assessed user adherence using app metadata to look at completion of expected activities, time to complete activities, and completion of activities by participant characteristics. We suggest that metadata can be used to generate insights on risks of ePRO and remote monitoring of symptoms, health, and adverse events during clinical trial data collection or during standard treatment.
Data availability
This work was conducted, in part, with NIH Small Business Innovation Research (SBIR) funding. Due to the sensitive nature of the data captured, the data release requirements of the KPNC subcontract, and the intellectual property associated with this work, all requests for data should be directed to the Chair of the Publication Committee, Dr. Oakley-Girvan (oakley@stanford.edu or Ingrid@medable.com). Requests for data will be considered by the publication committee under a data use agreement and will follow strict guidelines, including HIPAA regulations, regarding data use and the purpose of the analysis.
References
Khozin, S. & Coravos, A. decentralized trials in the age of real-world evidence and inclusivity in clinical investigations. Clin. Pharmacol. Ther. 106, 25–27. https://doi.org/10.1002/cpt.1441 (2019).
Dorsey, E. R., Kluger, B. & Lipset, C. H. The new normal in clinical trials: Decentralized studies. Ann. Neurol. 88, 863–866. https://doi.org/10.1002/ana.25892 (2020).
Basch, E. et al. Effect of electronic symptom monitoring on patient-reported outcomes among patients with metastatic cancer: A randomized clinical trial. JAMA 327, 2413–2422. https://doi.org/10.1001/jama.2022.9265 (2022).
Basch, E. et al. Overall survival results of a trial assessing patient-reported outcomes for symptom monitoring during routine cancer treatment. JAMA 318, 197–198. https://doi.org/10.1001/jama.2017.7156 (2017).
Basch, E. et al. Symptom monitoring with patient-reported outcomes during routine cancer treatment: A randomized controlled trial. J. Clin. Oncol. 34, 557–565. https://doi.org/10.1200/jco.2015.63.0830 (2016).
Denis, F. et al. Two-year survival comparing web-based symptom monitoring vs routine surveillance following treatment for lung cancer. JAMA 321, 306–307. https://doi.org/10.1001/jama.2018.18085 (2019).
Msaouel, P. et al. Evaluation of technology-enabled monitoring of patient-reported outcomes to detect and treat toxic effects linked to immune checkpoint inhibitors. JAMA Netw. Open 4, e2122998. https://doi.org/10.1001/jamanetworkopen.2021.22998 (2021).
Trojan, A., Huber, U., Brauchbar, M. & Petrausch, U. Consilium smartphone app for real-world electronically captured patient-reported outcome monitoring in cancer patients undergoing anti-PD-L1-directed treatment. Case Rep. Oncol. 13, 491–496. https://doi.org/10.1159/000507345 (2020).
Min, Y. H. et al. Daily collection of self-reporting sleep disturbance data via a smartphone app in breast cancer patients receiving chemotherapy: A feasibility study. J. Med. Internet Res. 16, e135. https://doi.org/10.2196/jmir.3421 (2014).
Teckie, S. et al. A mobile patient-facing app for tracking patient-reported outcomes in head and neck cancer survivors: Single-Arm Feasibility Study. JMIR Form. Res. 5, e24667. https://doi.org/10.2196/24667 (2021).
Zini, E. M. et al. A pilot study of a smartphone-based monitoring intervention on head and neck cancer patients undergoing concurrent chemo-radiotherapy. Int. J. Med. Inform. 129, 404–412. https://doi.org/10.1016/j.ijmedinf.2019.06.004 (2019).
Kneuertz, P. J. et al. Improving patient engagement, adherence, and satisfaction in lung cancer surgery with implementation of a mobile device platform for patient reported outcomes. J. Thorac. Dis. 12, 6883–6891. https://doi.org/10.21037/jtd.2020.01.23 (2020).
Velikova, G. et al. Electronic self-reporting of adverse events for patients undergoing cancer treatment: The eRAPID research programme including two RCTs. Progr. Grants Appl. Res. https://doi.org/10.3310/fdde8516 (2022).
Aapro, M. et al. Digital health for optimal supportive care in oncology: Benefits, limits, and future perspectives. Support. Care Cancer 28, 4589–4612. https://doi.org/10.1007/s00520-020-05539-1 (2020).
Oakley-Girvan, I., Yunis, R., Longmire, M. & Ouillon, J. S. What works best to engage participants in mobile app interventions and e-Health: A scoping review. Telemed. J. E. Health 28, 768–780. https://doi.org/10.1089/tmj.2021.0176 (2022).
Davis, S. W. & Oakley-Girvan, I. Achieving value in mobile health applications for cancer survivors. J. Cancer Surviv. 11, 498–504. https://doi.org/10.1007/s11764-017-0608-1 (2017).
Barello, S. et al. eHealth for patient engagement: A systematic review. Front. Psychol. 6, 2013. https://doi.org/10.3389/fpsyg.2015.02013 (2015).
Pratap, A. et al. Indicators of retention in remote digital health studies: A cross-study evaluation of 100,000 participants. NPJ Digit. Med. 3, 21. https://doi.org/10.1038/s41746-020-0224-8 (2020).
Gebert, P., Schindel, D., Frick, J., Schenk, L. & Grittner, U. Characteristics and patient-reported outcomes associated with dropout in severely affected oncological patients: An exploratory study. BMC Med. Res. Methodol. 21, 77. https://doi.org/10.1186/s12874-021-01259-0 (2021).
O’Brien, H. L. & Toms, E. G. What is user engagement? A conceptual framework for defining user engagement with technology. J. Am. Soc. Inf. Sci. Technol. 59, 938–955 (2008).
Kelders, S. M., van Zyl, L. E. & Ludden, G. D. S. The concept and components of engagement in different domains applied to eHealth: A systematic scoping review. Front. Psychol. 11, 926. https://doi.org/10.3389/fpsyg.2020.00926 (2020).
Madujibeya, I., Lennie, T., Aroh, A., Chung, M. L. & Moser, D. Measures of engagement with mhealth interventions in patients with heart failure: Scoping review. JMIR Mhealth Uhealth 10, e35657. https://doi.org/10.2196/35657 (2022).
Pham, Q. et al. A library of analytic indicators to evaluate effective engagement with consumer mhealth apps for chronic conditions: Scoping review. JMIR Mhealth Uhealth 7, e11941. https://doi.org/10.2196/11941 (2019).
Oakley-Girvan, I. et al. Usability evaluation of mobile phone technologies for capturing cancer patient-reported outcomes and physical functions. Digit. Health 9, https://doi.org/10.1177/20552076231186515 (2023).
Oakley-Girvan, I. et al. A novel smartphone application for the informal caregivers of cancer patients: Usability study. PLOS Digit. Health 2, e0000173. https://doi.org/10.1371/journal.pdig.0000173 (2023).
Lattard, C. et al. Clinical and economic impact of clinical oncology pharmacy in cancer patients receiving injectable anticancer treatments: A systematic review. J. Cancer Res. Clin. Oncol. https://doi.org/10.1007/s00432-023-04630-4 (2023).
Öztürk, E. S. & Kutlutürkan, S. The effect of the mobile application-based symptom monitoring process on the symptom control and quality of life in breast cancer patients. Semin. Oncol. Nurs. 37, 151161. https://doi.org/10.1016/j.soncn.2021.151161 (2021).
Suchodolska, G. & Senkus, E. Mobile applications for early breast cancer chemotherapy-related symptoms reporting and management: A scoping review. Cancer Treat. Rev. 105, 102364. https://doi.org/10.1016/j.ctrv.2022.102364 (2022).
Davis, A. C., Voelkel, J. L., Remmers, C. L., Adams, J. L. & McGlynn, E. A. Comparing Kaiser permanente members to the general population: Implications for generalizability of research. Perm J. 27(2), 87–98. https://doi.org/10.7812/tpp/22.172 (2023).
Apple Research Kit. Active Tasks. http://researchkit.org/docs/docs/ActiveTasks/ActiveTasks.html (2018).
Centers for Disease Control and Prevention. Stopping elderly accidents, deaths and injuries (STEADI). https://www.cdc.gov/steadi/ (2021).
Alharbi, M., Straiton, N., Smith, S., Neubeck, L. & Gallagher, R. Data management and wearables in older adults: A systematic review. Maturitas 124, 100–110. https://doi.org/10.1016/j.maturitas.2019.03.012 (2019).
Absolom, K. et al. Phase III Randomized controlled trial of eRAPID: eHealth intervention during chemotherapy. J. Clin. Oncol. 39, 734–747. https://doi.org/10.1200/jco.20.02015 (2021).
Crafoord, M. T., Fjell, M., Sundberg, K., Nilsson, M. & Langius-Eklöf, A. Engagement in an interactive app for symptom self-management during treatment in patients with breast or prostate cancer: Mixed methods study. J. Med. Internet Res. 22, e17058. https://doi.org/10.2196/17058 (2020).
Gordon, N. P. Similarity of adult Kaiser Permanente members to the adult population in Kaiser Permanente’s Northern California service area: comparisons based on the 2017/2018 cycle of the California Health Interview Survey (Kaiser Permanente Division of Research, Oakland, CA, 2020).
Acknowledgements
Research reported in this publication was partially supported by the National Cancer Institute of the National Institutes of Health under contract numbers HHSN261201700030C and HHSN261201800010C. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. We thank members of the Medable Patient Caregiver Network for their valuable input and critical insights during early development of both DigiBioMarC and TOGETHERCare apps that was conducted with Duke and Stanford Universities prior to the research with KPNC. This study could not have been done without the able assistance of Elaine M. Kurtovich, KPNC Research Project Manager, and Maya E. Ramsey, KNPC Research Associate. At the time of this work, RY, IOG, and SWD were employed by Medable Inc., which developed the TOGETHERCare and DigiBioMarC apps with funding from the NCI. SJF was subcontracted to work with Medable to analyze the study data and AK and SA were subcontracted to work with Medable to enroll KPNC participants and conduct all study participant focused efforts. RL and EN did not receive any funding as part of this effort.
Author information
Authors and Affiliations
Contributions
I.O.G. and S.W.D. researched the literature and conceived the study. I.O.G. and S.W.D. obtained funding from NIH for the study. I.O.G. was the PI for both NIH fast track study awards. I.O.G. and R.Y. developed the protocol. E.N., R.L., and A.K. were involved in protocol development, IRB approval, and patient recruitment. I.O.G. and R.Y. completed the semi-structured interviews. S.A. prepared data files and data dictionaries. S.J.F. provided data analysis with input and direction from I.O.G., E.N., R.Y., R.L. and S.W.D. R.Y., I.O.G., S.W.D. wrote the first draft of the manuscript. All authors reviewed and edited the manuscript and approved the final version of the manuscript.
Corresponding author
Ethics declarations
Competing interests
RY, IOG, and SWD are employed by Medable Inc., which developed the TOGETHERCare and DigiBioMarC apps with funding from the National Institutes of Health. SJF is subcontracted to work with Medable to analyze the study data and AK and SA were subcontracted to work with Medable to enroll KPNC participants and conduct all study participant focused efforts. RL and EN have no competing interests as they were not paid for their roles on the project.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Yunis, R., Fonda, S.J., Aghaee, S. et al. Mobile app activity engagement by cancer patients and their caregivers informs remote monitoring. Sci Rep 14, 3375 (2024). https://doi.org/10.1038/s41598-024-53373-w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-53373-w
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.