Dissemination & implementation science and technology scalability
Despite the development of countless digital health technologies by researchers, few have made the leap from research to broad use at scale. The next era of digital health interventions should leverage commercialization opportunities, as well as dissemination and implementation (D&I) science to ensure adoption in real-world settings. D&I science offers 2 key directions for digital health research. First, it provides frameworks by which to consider which existing technologies with proven efficacy should be used in real-world settings (e.g., healthcare, education, employer systems) and how to evaluate implementation in those settings, with consideration of the multilevel factors that make them more or less likely to be adopted (Shelton et al.,
2020). Second, they challenge researchers to consider the context in which tools will be used and the resources available in those contexts, suggesting that future technologies should be designed for dissemination and real-world implementation from the start (see Table
2).
Table 2
New frontiers, challenges, and opportunities for digital health in behavioral medicine
D&I science and technology scalability | Most digital interventions are difficult to deliver at scale and are rarely commercialized, as there are barriers to integration with existing systems D&I represent distinct disciplines focused on the translation of efficacious approaches to real-world settings | Emphasis on scalability, dissemination, and implementation concerns from the development stage Use of tools such as no-code platforms and APIs to promote interoperability and facilitate broader use |
Remote data collection and telehealth | Collecting data and delivering treatment remotely increases accessibility, reduces transmission of infectious diseases, and shows similar treatment outcomes to in-person approaches | Additional work to establish best practices for remote data collection protocols and high-quality evidence for the efficacy of remote intervention Advocacy for the continued reimbursement of remote treatment delivery and parity with in-person services |
AI | Tools such as machine learning and large language models have been leveraged to summarize large datasets, generate intervention content, and provide two-way patient education | Continued exploration of opportunities to leverage these technologies in behavioral medicine |
Ethical considerations | Privacy, data security, and health equity continue to present challenges, in part due to shifting legal landscapes and identification of biases in digital systems | Increased attention to these issues, the limitations they present for behavioral medicine, and opportunities they present for behavioral medicine professionals to lead improvements Wider use of tools such as the Digital Health Checklist (ReCODE Health) |
Training and collaboration | Despite the availability of new, complex digital technologies and advanced research methods to generate needed evidence, behavioral medicine professionals rarely have support for staying up to date in these areas | Emphasis on digital technologies, D&I and ethical concerns, and advanced research methods in behavioral medicine training programs Greater availability of continuing education resources and the protected time to use them Frequent and close collaboration between professionals with and without expertise in these areas |
To date, few trials have focused on the implementation of digital health tools in real-world settings. A notable exception is the US Veterans Health Administration, where myriad digital tools have been adopted and routinely implemented for a variety of health issues (e.g., smoking cessation, weight management; Blok et al.,
2019; US Department of Veterans Affairs,
2024). The VA’s centralized electronic health record system and payment structures across this clinical context have greatly facilitated research and implementation (Jackson et al.,
2011). However, despite the wide availability of health apps specifically for veterans, a recent survey found that uptake is low and the strongest predictor of veteran use of VA-created apps is provider encouragement, which results in nearly 3 times higher odds of use (Hogan et al.,
2022). Thus, even in a large, established, and digitally integrated health system, health app uptake in routine practice is low. Additional research and support is are needed to encourage providers to prescribe these tools to patients.
Beyond the VA, implementation of digital health tools in real-world settings has been sparse. Notable examples include a study that examined implementation of digital referrals to web-assisted tobacco interventions in community-based primary care practices, which found that digital referrals produced similar referral rates to a paper system, but threefold greater conversion to intervention registrations (Houston et al.,
2015). Implementation facilitators included ease of using the system and the perceived intervention efficacy (Houston et al.,
2015). Similarly, the Home BP trial tested a digital intervention for hypertension management in primary care. Researchers first undertook a systematic intervention planning process to consider multilevel factors impacting potential implementation, including feedback from patients and health professionals. In a subsequent randomized trial, they tested their digital intervention versus usual care and collected implementation data (e.g., cost effectiveness) to inform clinical rollout, finding that the intervention led to better hypertension management than usual care and with minimal incremental costs (McManus et al.,
2021). Here too, more research is needed on barriers and facilitators to implementation of digital tools in routine practice.
Implementation trials that focus solely on testing strategies to implement digital tools in existing settings and structures are also needed. One such study is the ongoing DIGITIS Trial which tests strategies to integrate prescription digital therapeutics for substance use disorders (Glass et al.,
2023). Using a factorial design, clinics are randomized to receive different combinations of implementation techniques to identify the optimal overall approach. Additional insights to facilitate implementation, particularly its sustainability in the clinical setting, will come from the extensive ongoing work with digital mental health treatments (Meyerhoff et al.,
2023; Mohr et al.,
2021). For example, an interdisciplinary international group of healthcare experts convened in 2019 to consider the barriers and facilitators to broad adoption of digital mental health tools (Mohr et al.,
2021). They found that while there is consensus that the tools are effective and cost-effective, there are still complications with proper reimbursements and there is not an established way to evaluate the tools, preventing further clinical implementation (Mohr et al.,
2021).
Yet another limit to the implementation of digital behavioral medicine tools is that the quality of evidence thus far has not been strong enough to move many digital solutions to clinical application. For example, a recent review identified 721 studies that describe virtual reality technologies for mental health, yet weaknesses in study design have hindered progression toward clinical adoption (Wiebe et al.,
2022). Specifically, few studies use rigorous and evidence-based processes at both the technology development and initial clinical testing phases, resulting in data that cannot support broad implementation (Selaskowski et al.,
2024). A protocol-based dual publication model, which is similar to a registered report but specific to the development and evaluation of digital technologies for clinical application, has been proposed to improve methodological quality (Selaskowski et al.,
2024). This is a promising approach, as it would encourage more detailed description of the technology development methods and foster greater replicability of digital tools while requiring robust clinical studies to demonstrate the efficacy data needed to justify further use and testing.
Remote data collection and telehealth
The COVID-19 pandemic accelerated a shift to remote data collection and treatment delivery in an effort to reduce the transmission of infectious disease. Remote options are also more accessible than in-person approaches and may offer needed flexibility for hard-to-reach and underprivileged groups, as the burdens of transportation and childcare are minimized or removed altogether. Remote methods typically leverage Bluetooth or wireless connected devices, wearable devices, mobile apps, online platforms, and/or video teleconferencing software. Some of these are freely available (e.g., Zoom) and as noted, some are already in widespread use (e.g., Fitbit), and evidence to support the validity of remote methods is growing. Specifically, evidence shows that weight, waist circumference, and movement assessments can be conducted with cancer survivors (Hoenemeyer et al.,
2022), older adults (Villar et al.,
2024), and veterans (Ogawa et al.,
2021) via Zoom video call, with high reliability and high concurrence with in-person methods. Similar trials are in progress to assess the validity of remote assessments of physical performance and mobility among older adult cancer survivors (Blair et al.,
2020).
For those who do not already use these technologies, however, remote protocols may be expensive for researchers and clinics, and poor execution may result in suboptimal patient engagement. For example, Hoenemeyer et al. (
2022) note that high shipping costs for the equipment necessary to conduct remote arm curl and grip strength tests (i.e., mailing dumbbells to participants’ homes) prevented the team from including these typical tests in their remote trial. Even when technology or equipment are available (e.g., Zoom), technical difficulties such as poor internet connectivity and environmental conditions such as poor lighting, incorrect camera angles, and distractions in the home can result in low engagement and lower-quality data, relative to in-person procedures. Participants may also perceive researchers to be less directly engaged in remote meetings than in person, as researchers often have to manage multiple tasks simultaneously (e.g., screenshare, recording responses; McClelland et al.,
2024).
Studies conducted during the transition from in-person to remote treatment during the pandemic revealed additional challenges, such as declines in use of behavioral strategies such as self-monitoring (Bernhart et al.,
2022) and suboptimal acceptability of remote procedures (at least initially) in certain subgroups (e.g., older adults; Pisu et al.,
2021; Ross et al.,
2021), possibly due to low technology literacy. Recommendations to address such barriers include encouraging participants to have cameras on during video calls (to promote attention and engagement), training research staff to look at the camera rather than the screen (for more direct eye contact) and limit distractions such as electronic notifications, and encouraging both participants and research staff to log into meetings early to troubleshoot any technical problems (McClelland et al.,
2024). Researchers can also consider building in fallback options such as phone calls (if technical difficulties cannot be resolved) and offering breaks during longer sessions (McClelland et al.,
2024), and build these options into protocols from the start (rather than as deviations from the expected protocol).
Yet, recent evidence for treatment outcomes is highly encouraging: in 2 trials that pivoted from in-person to videoconference-delivered behavioral weight loss intervention, weight loss was comparable between groups who received hybrid in-person (pre-pandemic) followed by remote (during pandemic) and remote-only treatment (Ross et al.,
2022; Tchang et al.,
2022). Similarly, studies show little (if any) difference between in-person and remote (telehealth) treatment for mental health outcomes (Bulkes et al.,
2022; Lin et al.,
2022), and attrition does not differ between modalities (Giovanetti et al.,
2022). Thus, research increasingly demonstrates that fully remote treatment protocols can produce meaningful change in clinical outcomes with the potential for less burden on participants and patients.
Remotely delivered telehealth services were available before the COVID-19 pandemic, though adoption in routine clinical practice occurred mostly in rural and other settings where access was limited, and reimbursement policies varied greatly across states (Brotman & Kotloff,
2021). Reimbursement barriers were lifted during the pandemic to address the critical public health need. Federal programs such as Medicare have maintained reimbursement for telehealth services through 2024 (Department of Health & Human Services,
2024), but the future is uncertain: some insurance providers offer lower financial compensation for telehealth than in-person visits (Aremu et al.,
2022) and the sustainability of legislative and budgetary support for telehealth is unclear. The field of behavioral medicine should prioritize establishing the efficacy for remotely delivered interventions given that such data are needed to inform reimbursement policy, and should continue to advocate for telehealth reimbursement.
Artificial intelligence (AI)
AI, or “technology that simulates human intelligence and problem-solving capabilities” (Stryker & Kavlakoglu,
2024), is increasingly used in daily life and healthcare (e.g., GPS, digital assistants, social media algorithms, ChatGPT) and is a relatively new frontier for behavioral medicine. AI has myriad applications in behavioral medicine and has the potential to improve measurement and prediction of behavior and clinical outcomes with far greater precision than traditional methods (Bucher et al.,
2024). AI also has potential to help us design more personalized and effective interventions, which are urgently needed given the rapid evolution of data sources during the 21st century. Traditional data sources have included biological assays, surveys, and focus group/interviews, but in recent years, intensive longitudinal assessment methods, wearable devices, mobile applications, and online platforms have been used to collect high volumes of data (i.e., big data). AI is well-suited for handling big data and its use offers new ways to understand, predict, and intervene on health behavior. Machine learning, natural language processing, generative AI and large language models, and computer vision are four types of AI methods that have great potential to revolutionize behavioral medicine research.
Machine learning (ML) uses algorithms that learn from data to make predictions, identify patterns, and/or make decisions (Shalev-Shwartz & Ben-Davd,
2014). Applications in healthcare include predicting patient outcomes, risk for disease, and disease outbreaks; creating tailored treatment plans; and improving the efficiency of healthcare systems (Dixon et al.,
2024). In behavioral medicine, ML has been used to predict intervention outcomes (Khalilnejad et al.,
2024), diet lapses during weight loss treatment (Goldstein et al.,
2018), smoking behavior (Yu et al.,
2024), patient adherence (Masiero et al.,
2024), and depressive symptoms (De la Barrera et al.,
2024). ML has also been used to predict behavior in real time and to optimize and personalize behavioral interventions (Forman et al.,
2019; Presseller et al.,
2023; Rocha et al.,
2023; Scodari et al.,
2023).
Natural language processing (NLP) uses algorithms to understand, process, and analyze human language (AlShehri et al.,
2024; Vaniukov,
2024), and is leveraged in conversational agents and large language models. NLP has been used in medicine to analyze speech and text from a range of sources including clinic notes, patient comments in an electronic health record, and the research literature to aid in diagnosis, prevention, and patient engagement (AlShehri et al.,
2024; Petti et al.,
2020). Behavioral medicine researchers use NLP to study qualitative and/or social media data (Cha & Lee,
2024; Lau et al.,
2024; Patra et al.,
2023), assist with dietary self-monitoring via voice or text entry (Chikwetu et al.,
2023), and identify patient-reported outcomes via clinical notes in electronic health records (Ebrahimi et al.,
2024; Sim et al.,
2023).
Generative AI (genAI) is the use of AI to generate content of all types (e.g., text, images). Large language models (LLM) are a form of genAI that produces text (Yu et al.,
2023). LLMs perform tasks in response to text queries and generate human-like responses via a computer program that has been trained on vast amounts of data to interpret human language. They can generate content, engage in conversations, and answer questions, and they are used in conversational agents, text analysis, code generation, language translation, and text summarization (Thirunavukarasu et al.,
2023). Conversational agents, a popular type of LLM, use ML and NLP to engage in conversations with humans (Laranjo et al.,
2018). Chatbots are one type of conversational agent in which a software program automates specific conversational tasks, like answering a finite set of questions or providing information or assistance to a user (Tudor Car et al.,
2020). In behavioral medicine, chatbots have been used for weight management, vaccine communication, smoking cessation, and chronic disease management (Aggarwal et al.,
2023; Bak & Chin,
2024; Noh et al.,
2023; Passanante et al.,
2023). Conversational agents can be used to counsel and educate patients about these and other topics, which may reduce the intensity of behavioral interventions that rely on human delivery.
ChatGPT, often used as a conversational agent, is perhaps the most notable LLM, as it exploded in popularity after its release by OpenAI in 2023 (OpenAI,
2023). In medicine, ChatGPT and other LLMs assist with clinical notes and summaries, answer medical exam or patient questions, provide patient education, augment medical training (Omiye et al.,
2024), conduct cancer screening and genetic counseling, assess symptoms, and support caregivers (Jiang et al.,
2024). In behavioral medicine, researchers have examined the accuracy of LLMs in providing patient education (Kozaily et al.,
2024), debunking health misinformation, developing exercise programs, recommending evidence-based treatments, identifying motivational states, and even conducting systematic reviews (Amin et al.,
2023). Behavioral medicine researchers are also using LLMs to create intervention content (Willms et al., in press).
Computer vision may also have myriad uses in behavioral medicine, as it uses videos and/or images to train a computer program to recognize and interpret visual content. These data are gathered via sensors and/or cameras and ML algorithms perform object detection and image classification. For example, computer vision has been used to diagnose skin cancer by detecting characteristics of skin lesions not visible to the naked eye (Akilandasowmya et al.,
2024), to classify different types of back pain based on movements (Hartley et al.,
2024), to detect pain and acute patient deterioration via changes in facial expression, to monitor mobility in intensive care patients, to detect falls in the elderly, and to assist in the diagnosis of autism via head motion and facial expression data (Lindroth et al.,
2024). In behavioral medicine, computer vision has been used to identify foods based on pictures taken by users and use this information to tailor intervention messages (Chew et al.,
2024). Together, AI tools have enormous potential to increase the efficiency, accuracy, and reach of behavioral medicine interventions, making this an exciting new area for growth.
Although AI has extraordinary potential to improve individual and public health, it also has extraordinary potential to negatively impact health. For example, AI can be used to produce deepfakes and large volumes of disinformation, in the forms of authentic-appearing online articles with fictional medical references, social media posts and comments, and patient and physician testimonials (Menz et al.,
2024). The implications of the ability to rapidly develop and disseminate disinformation are likely to be far-reaching and evidence suggests that AI-powered disinformation campaigns are already infiltrating low- and middle-income nations (Hotez,
2024). In the past few years, the World Health Organization (WHO) has issued statements urging the judicious use of AI for health (World Health Organization,
2023) and laid out guidance on the ethical use of AI for health (World Health Organization,
2021). However, the WHO’s very own health chatbot, S.A.R.A.H. (Smart AI Resource Assistant for Health; World Health Organization,
2024) has come under fire for providing outdated information (Nix,
2024) and includes a disclaimer stating that “answers may not always be accurate” (World Health Organization,
2024). The WHO is calling for researchers to help them identify ways their chatbot can be used to disseminate accurate health information (World Health Organization,
2024). Given the chatbot’s focus on healthy lifestyle topics such as quitting smoking, physical activity, healthy diet, and stress reduction, behavioral medicine researchers are well-positioned to lead this charge. Generally, much research is needed on both the benefits and dangers of genAI to public health.
Ethical considerations
In 2019 we called attention to issues of privacy and data security, to promote the responsible use of digital health tools in behavioral medicine research and practice. Groups such as ReCODE Health (
2024) provide a wealth of resources on this topic, including the Digital Health Checklist (Nebeker et al.,
2021), which provides guidance to researchers and ethics committees with respect to the use of digital tools in behavioral research. Such resources are invaluable and support for their ongoing revision is essential as the digital health landscape continues to evolve. An important example comes from the 2022 US Supreme Court decision
Dobbs v. Jackson Women’s Health Organization, which overturned
Roe v. Wade. This decision had immediate implications for digital health. With states moving to outlaw abortion and related reproductive healthcare, data from self-monitoring tools such as Fitbit and menstrual cycle tracking apps could be subpoenaed by law enforcement to aid in the prosecution of those who perform and/or receive restricted services (Kim,
2022). In the wake of announcements that companies would make such data available, many users of these tools reported concerns about their use in research; some indicated that they would decline to participate in such research if use of these tools were required, particularly users who identified with minoritized groups (Salvatore et al.,
2024). Legislation restricting reproductive healthcare presents new ethical issues in women’s health research, and protecting the privacy and security of digital health data in behavioral medicine research is paramount.
The use of AI also comes with ethical considerations related to bias, privacy, and informed consent. Algorithmic bias can occur when data used to train AI are biased, when the AI is used in a different context or population than the one for which it was originally designed, or when the AI’s results are interpreted in a biased way (National Library of Medicine,
2024). Such biases may widen health inequities rather than help to close them. The Agency for Healthcare Research and Quality and the National Institute for Minority Health and Health Disparities recommend the following principles to reduce bias in AI: (1) promote health equity during all phases of the algorithm life cycle, (2) ensure algorithms and their use are transparent and explainable, (3) engage patients and communities during all phases, (4) identify algorithm fairness issues, and (5) establish accountability for equity and fairness in outcomes emanating from algorithms (Chin et al.,
2023).
Privacy is another ethical consideration in the use of AI, given the risk that protected health information ends up being used to train algorithms. AI uses must be HIPPA-compliant and privacy protections must be built-in and resistant to data breaches. In 2023, the American Psychiatric Association issued an advisory to clinicians and researchers against the use of patient information in AI systems (American Psychiatric Association,
2023). Researchers must be transparent about the data being used in AI systems and its potential biases and be mindful about informed consent, which can only be obtained when researchers can explain the technology used, how patient data will be used, and the potential limitations and biases of the technology (Diaz-Asper et al.,
2024). This is also relevant in clinical settings when AI is used for diagnosis and/or treatment decision making (Park,
2024). Because AI evolves faster than ethical best practices can be established, research is needed on potential harms and ethical issues emanating from the use of AI in behavioral medicine research and practice.