Joachim Roski, Booz Allen Hamilton; Wendy Chapman, University of Melbourne; Jaimee Heffner, Fred Hutchinson Cancer Research Center; Ranak Trivedi, Stanford University; Guilherme Del Fiol, University of Utah; Rita Kukafka, Columbia University; Paul Bleicher, OptumLabs; Hossein Estiri, Harvard Medical School; Jeffrey Klann, Harvard University; and Joni Pierce, University of Utah
This chapter explores the positive, transformative potential of artificial intelligence (AI) for health and health care. We discuss possible AI applications for patients and their families; the caregiving team of clinicians, public health professionals, and administrators; and health researchers (Roski et al., 2018). These solutions offer readers a glimpse of a possible future. We end by offering perspective about how AI might transform health care and by providing high-level considerations for addressing barriers to that future.
The health care industry has been investing for years in technology solutions, including AI. There have been some promising examples of health care AI solutions, but there are gaps in the evaluation of these tools in the peer-reviewed literature, and so it can be difficult to assess their impact. Also difficult to assess is the impact of combined solutions. Specific technology solutions when coupled may improve positive outcomes synergistically. For example, an AI solution may become exponentially more powerful if it is coupled with augmented reality, virtual reality, faster computing systems, robotics, or the Internet of Things (IoT). It is impossible to predict in advance.
This chapter presents the potential of AI solutions for patients and families, clinical care teams, public health program managers, business administrators, and researchers. Table 3-1 provides examples of the types of applications of AI for
TABLE 3-1 | Examples of Artificial Intelligence Applications for Stakeholder Groups
| Use Case or User Group | Category | Examples of Applications | Technology |
|---|---|---|---|
| Patients and families | Health monitoring Benefit/risk assessment Disease prevention and management |
|
Machine learning, natural language processing (NLP), speech recognition, chatbots Conversational artificial intelligence (AI), NLP, speech recognition, chatbots |
| Medication management |
|
Robotic home telehealth | |
| Rehabilitation |
|
Robotics | |
| Clinician care teams | Early detection, prediction, and diagnostics tools |
|
Machine learning |
| Surgical procedures |
|
Robotics, machine learning | |
| Precision medicine Patient safety |
|
Supervised machine learning, reinforcement learning Machine learning | |
| Public health program managers | Identification of individuals at risk Population health |
|
Deep learning (convolutional and recurrent neural networks) Deep learning, geospatial pattern mining, machine learning |
| Population health |
|
Deep learning, geospatial pattern mining, machine learning | |
| Business administrators | International Classification of Diseases, 10th Revision coding |
|
Machine learning, NLP |
| Fraud detection |
|
Supervised, unsupervised, and hybrid machine learning | |
| Cybersecurity |
|
Machine learning, NLP | |
| Physician management |
|
Machine learning, NLP | |
| Researchers | Genomics Disease prediction Discovery |
|
Integrated cognitive computing Neural networks Machine learning, computer-assisted synthesis |
these stakeholders. These examples are not exhaustive. The following sections explore the promise of AI in health care in more detail.
AI could soon play an important role in the self-management of chronic diseases such as cardiovascular diseases, diabetes, and depression. Self-management tasks can range from taking medications, modifying diet, and getting more physically active to care management, wound care, device management, and the delivery of injectables. Self-management can be assisted by AI solutions, including conversational agents, health monitoring and risk prediction tools, personalized adaptive interventions, and technologies to address the needs of individuals with disabilities. We describe these solutions in the following sections.
Conversational agents can engage in two-way dialogue with the user via speech recognition, natural language processing (NLP), natural language understanding, and natural language generation (Laranjo et al., 2018). AI is behind many of them. These interfaces may include text-based dialogue, spoken language, or both. They are called, variously, virtual agents, chatbots, or chatterbots. Some conversational agents present a human image (e.g., the image of a nurse or a coach) or nonhuman image (e.g., a robot or an animal) to provide a richer interactive experience. These are called embodied conversational agents (ECAs). These visible characters provide a richer and more convincing interactive experience than non-embodied voice-only agents such as Apple’s Siri, Amazon’s Alexa, or Microsoft’s Cortana. The imagistic entities can communicate nonverbally through hand gestures and facial expressions.
In the “self-management” domain, conversational agents already exist to address depression, smoking cessation, asthma, and diabetes. Although many chatbots and ECAs exist, evaluation of these agents has, unfortunately, been limited (Fitzpatrick et al., 2017).
The future potential for conversational agents in self-management seems high. While simulating a real-world interaction, the agent may assess symptoms, report back on outputs from health monitoring, and recommend a course of action based on these varied inputs. Most adults say they would use an intelligent virtual coach or an intelligent virtual nurse to monitor health and symptoms at home. There is somewhat lower enthusiasm for mental health support delivered via this method (Accenture, 2018).
Social support improves treatment outcomes (Hixson et al., 2015; Wicks et al., 2012). Conversational agents can make use of humans’ propensity to
treat computers as social agents. Such support could be useful as a means of combating loneliness and isolation (Stahl and Coeckelbergh, 2016; Wetzel, 2018). In other applications, conversational agents can be used to increase the engagement and effectiveness of interventions for health behavior change. Most studies of digital interventions for health behavior change have included support from either professionals or peers. It is worth noting that professionally supported interventions cost two to three times what technology interventions cost. A conversational agent could provide some social support and increased engagement while remaining scalable and cost-effective. Moreover, studies have shown that people tend to be more honest when interacting with technology than with humans (Borzykowski, 2016).
In the next decade, conversational AI will probably become more widely used as an extender of clinician support or as a stopgap where other options are not available (see Chapter 1). Now under development are new conversational AI strategies to infer emotion from voice analysis, computer vision, and other sources. We think it is likely that systems will thus become more conversant in the emotional domain and more effective in their communication.
AI can use raw data from accelerometers, gyroscopes, microphones, cameras, and other sensors, including smartphones. Machine learning algorithms can be trained to recognize patterns from the raw data inputs and then categorize these patterns as indicators of an individual’s behavior and health status. These systems can allow patients to understand and manage their own health and symptoms as well as share data with medical providers.
The current acceptance of wearables, smart devices, and mobile health applications has risen sharply. In just a 4-year period, between 2014 and 2018, the proportion of U.S. adults reporting that they use wearables increased from 9 percent to 33 percent. The use of mobile health apps increased from 16 percent to 48 percent (Accenture, 2018). Consumer interest is high (~50 percent) in using data generated by apps, wearables, and IoT devices to predict health risks (Accenture, 2018). Since 2013, AI startup companies with a focus on health care and wearables have raised $4.3 billion to develop, for example, bras designed for breast cancer risk prediction and smart clothing for cardiac, lung, and movement sensing (Wiggers, 2018).
AI-driven adaptive interventions are called JITAIs, or “just-in-time adaptive interventions.” These are learning systems that deliver dynamic, personalized
treatment to users over time (Nahum-Shani et al., 2015; Spruijt-Metz and Nilsen, 2014). The JITAI makes decisions about when and how to intervene based on response to prior intervention, as well as on awareness of current context, whether internal (e.g., mood, anxiety, blood pressure) or external (e.g., location, activity). JITAI assistance is provided when users are most in need of it or will be most receptive to it. These systems can also tell a clinician when a problematic pattern is detected. For example, a JITAI might detect when a user is in a risky situation for substance abuse relapse—and deliver an intervention against it.
These interventions rely on sensors, rather than a user’s self-report, to detect states of vulnerability or intervention opportunity. This addresses two key self-management challenges: the high user burden of self-monitoring and the limitations of self-awareness. As sensors become more ubiquitous in homes, in smartphones, and on bodies, the data sources for JITAIs are likely to continue expanding. AI can be used to allow connected devices to communicate with one another. (Perhaps a glucometer might receive feedback from refrigerators regarding the frequency and types of food consumed.) Leveraging data from multiple inputs can uniquely enhance AI’s ability to provide real-time behavioral management.
According to the Centers for Disease Control and Prevention, 16 million individuals are living with cognitive disability in the United States alone. Age is the single best predictor of cognitive impairments, and an estimated 5 million Americans more than 65 years old have Alzheimer’s disease. These numbers are expected to increase due to the growth of an aging population: currently nearly 9 percent of all adults are more than 65 years old, a percentage expected to double by 2050 (CDC, 2018; Family Caregiver Alliance, 2019). Critically, 15.7 million family members provide unpaid care and support to individuals with Alzheimer’s disease or other dementias (Alzheimers.net, 2019). The current system of care is unprepared to handle the current or future load of patient needs, or to allow individuals to “age in place” at their current homes rather than relocating to assisted living or nursing home facilities (Family Caregiver Alliance, 2019).
Smart home monitoring and robotics may eventually use AI to address these challenges (Rabbitt et al., 2015). Home monitoring has the potential to increase independence and improve aging at home by monitoring physical space, falls, and amount of time in bed. (Excessive time in bed can be both the cause and outcome of depression, and places the elderly at high risk for bedsores, loss of mobility, and increased mortality.) Currently available social robots such as PARO, Kabochan, and PePeRe provide companionship and stimulation for dementia patients. Recently, the use of robotic pets has been reported to be helpful in
reducing agitation in nursing home patients with dementia (Schulman-Marcus et al., 2019; YellRobot, 2018).
For example, PARO is a robot designed to look like a cute, white baby seal that helps calm patients in hospitals and nursing homes. Initial pilot testing of PARO with 30 patient–caregiver dyads showed that PARO improved affect and communication among those dementia patients who interacted with the robot. This benefit was especially seen among those with more cognitive deficits. Larger clinical trials have also demonstrated improvements in patient engagement, although the effects on cognitive symptoms remain ambiguous (Moyle et al., 2017).
Although socially assistive robots are designed primarily for older adult consumers, caregivers also benefit from them because they relieve caregiver burden a bit and thus improve their well-being. As the technology improves, it may be that robots will do increasingly sophisticated tasks. Future applications of robotics are being developed to provide hands-on care. Platforms for smart home monitoring may eventually incorporate caregiver and patient needs in one seamless experience to ensure a family-wide experience rather than individual experiences. Designers of smart home monitoring should consider the ethics of equitable access by designing AI for the elderly, the dependent, and the short- or long-term disabled (Johnson, 2018).
There are two main areas of opportunity for AI in clinical care: (1) enhancing and optimizing care delivery, and (2) improving information management, user experience, and cognitive support in electronic health records (EHRs). Strides have been made in these areas for decades, largely through rule-based, expert-designed applications typically focused on specific clinical areas or problems. AI techniques offer the possibility of improving performance further.
The amount of relevant data available for patient care is growing and will continue to grow in volume and variety. Data recorded digitally through EHRs only scratch the surface of the types of data that (when appropriately consented) could be leveraged for improving patient care. Clinicians are beginning to have access to data generated from wearable devices, social media, and public health records; to data about consumer spending, grocery purchase nutritional value, and an individual’s exposome; and to the many types of -omic data specific to an individual. AI will probably, we think, have a profound effect on the entire clinical care process, including prevention, early detection, risk/benefit identification, diagnosis, prognosis, and personalized treatment.
The area of prediction, early detection, and risk assessment for individuals is one of the most fruitful AI applications (Sennaar, 2018). In this chapter, we discuss examples of such use; Chapters 5 and 6 provide thoughts about external evaluation.
There are a number of demonstrations of AI in diagnostic imaging. Diagnostic image recognition can differentiate between benign and malignant melanomas, diagnose retinopathy, identify cartilage lesions within the knee joint (Liu et al., 2018), detect lesion-specific ischemia, and predict node status after positive biopsy for breast cancer. Image recognition techniques can differentiate among competing diagnoses, assist in screening patients, and guide clinicians in radiotherapy and surgery planning (Matheson, 2018). Automated image classification may not disrupt medicine as much as the invention of the roentgenogram did, but the roles of radiologists, dermatologists, pathologists, and cardiologists will likely change as AI-enabled diagnostic imaging improves and expands. Combining output from an AI diagnostic imaging prediction with prediction from the physician seems to decrease human error (Wang et al., 2016). Although some believe AI will replace physicians in diagnostic imaging, it is more likely that these techniques will mainly be assistive, sorting and prioritizing images for more immediate review, highlighting important findings that might have been missed, and classifying simple findings so that the humans can spend more time on complex cases (Parakh et al., 2019).
Histopathologic diagnosis has seen similar gains in cancer classification from tissue, in universal microorganism detection from sequencing data, and in analysis of a single drop of body fluid to find evidence of bacteria, viruses, or proteins that could indicate an illness (Best, 2017).
AI is becoming more important for surgical decision making. It brings to bear diverse sources of information, including patient risk factors, anatomic information, disease natural history, patient values, and cost, to help physicians and patients make better predictions regarding the consequences of surgical decisions. For instance, a deep learning model was used to predict which individuals with treatment-resistant mesial temporal lobe epilepsy would most likely benefit from surgery (Gelichgerrcht et al., 2018). AI platforms can provide roadmaps to aid the surgical team in the operating room, reducing risk and making surgery safer (Newmarker, 2018). In addition to planning and decision making, AI may be applied to change surgical techniques. Remote-controlled robotic surgery has
been shown to improve the safety of interventions where clinicians are exposed to high doses of ionizing radiation and makes surgery possible in anatomic locations not otherwise reachable by human hands (Shen et al., 2018; Zhao et al., 2018). As autonomous robotic surgery improves, it is likely that surgeons will in some cases oversee the movements of robots (Shademan et al., 2016).
Precision medicine allows clinicians to tailor medical treatment to the individual characteristics of each patient. Clinicians are testing whether AI will permit them to personalize chemotherapy dosing and map patient response to a treatment so as to plan future dosing (Poon et al., 2018). AI-driven NLP has been used to identify polyp descriptions in pathology reports that then trigger guideline-based clinical decision support to help clinicians determine the best surveillance intervals for colonoscopy exams (Imler et al., 2014). Other AI tools have helped clinicians select the best treatment options for complex diseases such as cancer (Zauderer et al., 2014).
The case of clinical equipoise—when clinical practice guidelines do not present a clear preference among care treatment options—also has significant potential for AI. Using retrospective data from other patients, AI techniques can predict treatment responses of different combinations of therapies for an individual patient (Brown, 2018). These types of tools may serve to help select a treatment immediately and may also provide new knowledge to future practice guidelines. Possibly useful will be dashboards demonstrating predicted outcomes along with cost of treatment and expected changes based on patient behavior, such as increased exercise. These may provide an excellent platform for shared decision making involving the patient, family, and clinical team. AI could also support a patient-centered medical homes model (Jackson et al., 2013).
As genome-phenome integration is realized, the use of genetic data in AI systems for diagnosis, clinical care, and treatment planning will probably increase. To truly impact routine care, though, genetic datasets will need to better represent the diversity of patient populations (Hindorff et al., 2018).
AI can also be used to find similar cases from patient records in an EHR to support treatment decisions based on previous outcomes (Schuler et al., 2018).
The following sections describe a few areas that could benefit from AI-supported tools integrated with EHR systems, including information management (e.g., clinical documentation, information retrieval), user experience, and cognitive support.
EHR systems and regulatory requirements have introduced significant clinical documentation responsibilities to providers, without necessarily supporting patient care decisions (Shanafelt et al., 2016). AI has the potential to improve the way in which clinicians store and retrieve clinical documentation. For example, the role of voice recognition systems in clinical documentation is well known. However, such systems have been used mostly to support clinicians’ dictation of narrative reports, such as clinical notes and diagnostic imaging reports (Hammana et al., 2015; Zick and Olsen, 2001). As mentioned previously, AI-enabled conversational and interactive systems (e.g., Amazon’s Alexa, Apple’s Siri) are now widespread outside of health care. Similar technology could be used in EHR systems to support various information management tasks. For example, clinicians could ask a conversational agent to find specific information in the patient’s record (e.g., “Show me the patient’s latest HbA1c results”), enter orders, and launch EHR functions. Instead of clicking through multiple screens to find relevant patient information, clinicians could verbally request specific information and post orders while still looking at and talking to the patient or caregivers (Bryant, 2018). In the near future, this technology has the potential to improve the patient–provider relationship by reducing the amount of time clinicians spend focused on a computer screen.
AI has the potential to not only improve existing clinical decision support (CDS) modalities but also enable a wide range of innovations with the potential to disrupt patient care. Improved cognitive support functions include smarter CDS alerts and reminders as well as better access to peer-reviewed literature.
A core cause for clinicians’ dissatisfaction with EHR systems is the high incidence of irrelevant pop-up alerts that disrupt the clinical workflow and contribute to “alert fatigue” (McCoy et al., 2014). This problem is partially caused by the low specificity of alerts, which are frequently based on simple and deterministic handcrafted rules that fail to consider the full clinical context. AI can improve the specificity of alerts and reminders by considering a much larger number of patient and contextual variables (Joffe et al., 2012). It can provide probability thresholds that can be used to prioritize alert presentation and determine alert format in the user interface (Payne et al., 2015). It can also continuously learn from clinicians’ past behavior (e.g., by lowering the priority of alerts they usually ignore).
Recent advances in AI show promising applications in clinical knowledge retrieval. For example, mainstream medical knowledge resources are already using machine learning algorithms to rank search results, including algorithms that continuously learn from users’ search behavior (Fiorini et al., 2018). AI-enabled clinical knowledge retrieval tools could also be accessed through the same conversational systems that allow clinicians to retrieve patient information from the EHR. Through techniques such as information extraction, NLP, automatic summarization, and deep learning, AI has the potential to transform static narrative articles into patient-specific, interactive visualizations of clinical evidence that could be seamlessly accessed within the EHR. In addition, “living systematic reviews” can continuously update clinical evidence as soon as the results of new clinical trials become available, with EHRs presenting evidence updates that may warrant changes to the treatment of specific patients (Elliott et al., 2014).
Next, we explore AI solutions for population and public health programs. These include solutions that could be implemented by health systems (e.g., accountable care organizations), health plans, or city, county, state, and federal public health departments or agencies. Population health examines the distribution of health outcomes within a population, the range of factors that influence the distribution of health outcomes, and the policies and interventions that affect those factors (Kindig and Stoddart, 2003). Population health programs are often implemented through nontraditional partnerships among different sectors of the community—public health, industry, academia, health care, local government entities, etc. On the other hand, public health is the science of protecting and improving the health of people and their communities (CDC Foundation, 2019). This work is achieved by promoting healthy lifestyles, researching disease and injury prevention, and detecting, preventing, and responding to infectious diseases. Overall, public health is concerned with protecting the health of entire populations. These populations can be as small as a local neighborhood or as big as an entire country or region of the world.
AI can help identify specific demographics or geographical locations where the prevalence of disease or high-risk behaviors exist. Researchers have successfully
applied convolutional neural network analytic approaches to quantify associations between the built environment and obesity prevalence. They have shown that physical characteristics of a neighborhood can be associated with variations in obesity prevalence across different neighborhoods (Maharana and Nsoesie, 2018). Shin et al. (2018) applied a machine learning approach that uses both biomarkers and sociomarkers to predict and identify pediatric asthma patients at risk of hospital revisits.
Without knowing specific symptom-related features, the sociomarker-based model correctly predicted two out of three patients at risk. Once identified, population or regions can be targeted with computational health campaigns that blur the distinction between interpersonal and mass influence (Cappella, 2017). However, the risks of machine learning in these contexts have also been described (Cabitza et al., 2017). They include (1) the risk that clinicians become unable to recognize when the algorithms are incorrect, (2) lack of an ability for the algorithms to address the context of care, or (3) the intrinsic lack of reliability of some medical data. However, many of these challenges are not intrinsic to machine learning or AI, but rather represent misuse of the technologies.
A spectrum of market-ready AI approaches to support population health programs already exists. They are used in areas of automated retinal screening, clinical decision support, predictive population risk stratification, and patient self-management tools (Contreras and Vehi, 2018; Dankwa-Mullan et al., 2019). Several solutions have received regulatory approval; for example, the U.S. Food and Drug Administration approved Medtronic’s Guardian Connect, marking the first AI-powered continuous glucose monitoring system. Crowd-sourced, real-world data on inhaler use, combined with environmental data, led to a policy recommendations model that can be replicated to address many public health challenges by simultaneously guiding individual, clinical, and policy decisions (Barrett et al., 2018).
There is an alternative approach to standard risk prediction models that applies AI tools. For example, predictive models using machine learning algorithms may facilitate recognition of clinically important unanticipated predictor variables that may not have previously been identified by “traditional” research approaches that rely on statistical methods testing a priori hypotheses (Waljee et al., 2014). Enabled by the availability of data from administrative claims and EHRs, machine learning can enable patient-level prediction, which moves beyond average population effects to consider personalized benefits and risks. Large-scale, patient-level
prediction models from observational health care data are facilitated by a common data model that enables prediction researchers to work across computer environments. An example can be found from the Observational Health Data Sciences and Informatics collaborative, which has adopted the Observational Medical Outcomes Partnership common data model for patient-level prediction models using observational health care data (Reps et al., 2018).
Another advantage of applying AI approaches to predictive models is the ability not only to predict risk but also the presence or absence of a disease in an individual. As an example, successful use of a memetic pattern-based algorithm approach was demonstrated in a broad risk spectrum of patients undergoing coronary artery disease evaluation and was shown to successfully identify and exclude coronary artery disease in a population instead of just predicting the probability of future events (Zellweger et al., 2018). In addition to helping health care organizations identify individuals with elevated risks of developing chronic conditions early in the disease’s progression, this approach may prevent unnecessary diagnostic procedures in patients where procedures may not be warranted and also support better outcomes.
Not all data elements needed to predict chronic disease can be found in administrative records and EHRs. Creating risk scores that include a blend of social, behavioral, and clinical data may help give providers the actionable, 360-degree insight necessary to identify patients in need of proactive, preventive care while meeting reimbursement requirements and improving outcomes (Kasthurirathne et al., 2018).
Public health professionals are focused on solutions for more efficient and effective administration of programs, policies, and services; disease outbreak detection and surveillance; and research. Relevant AI solutions are being experimented with in a number of areas.
The range of AI solutions that can improve disease surveillance is considerable. For a number of years, researchers have tracked and refined the options for tracking disease outbreaks using search engine query data. Some of these approaches rely on the search terms that users type into Internet search engines (e.g., Google Flu Trends). At the same time, caution is warranted with these approaches. Relying on data not collected for scientific purposes (e.g., Internet search terms) to predict flu outbreaks has been fraught with error (Lazer et al., 2014). Nontransparent search algorithms that change constantly cannot be easily replicated and studied.
These changes may occur due to business needs (rather than the needs of a flu outbreak detection application) or due to changes in search behavior of consumers. Finally, relying on such methods exclusively misses the opportunity to combine them and co-develop them in conjunction with more traditional methods. As Lazer et al. (2014) detail, combining traditional and innovative methods (e.g., Google Flu Trends) performs better than either method alone.
Researchers and solution developers have experimented with the integration of case- and event-based surveillance (e.g., news and online media, sensors, digital traces, mobile devices, social media, microbiological labs, and clinical reporting) to arrive at dashboards and analysis approaches for threat verification. Such approaches have been referred to as digital epidemiological surveillance and can produce timelier data and reduce labor hours of investigation (Kostokova, 2013; Zhao et al., 2015). Such analyses rely on AI’s capacities in spatial and spatiotemporal profiling, environmental monitoring, and signal detection (i.e., from wearable sensors). They have been successfully implemented to build early warning systems for adverse drug events, falls detection, and air pollution (Mooney and Pejaver, 2018). The ability to rely on unstructured data such as photos, physicians’ notes, sensor data, and genomic information, when enabled by AI, may lead to additional, novel approaches in disease surveillance (Figge, 2018).
Moreover, participatory systems such as social media and listservs could be relied on to solicit information from individuals as well as groups in particular geographic locations. For example, such approaches may encourage a reduction in unsafe behaviors that put individuals at risk for human immunodeficiency virus (HIV) infection (Rubens et al., 2014; Young et al., 2017). For example, it has been demonstrated that psychiatric stressors can be detected from Twitter posts in select populations through keyword-based retrieval and filters and the use of neural networks (Du et al., 2018). However, how such AI solutions could improve the health of populations or communities is less clear, due to the lack of context for some tweets and because tweets may not reflect the true underlying mental health status of a person who tweeted. Studies that retroactively analyze the tweeting behavior of individuals with known suicide attempts or ideation, or other mental health conditions, may allow refinement in such approaches.
Finally, AI and machine learning have been used to develop a dashboard to provide live insight into opioid usage trends in Indiana (Bostic, 2018). This tool enabled prediction of drug positivity for small geographic areas (i.e., hot spots), allowing for interventions by public health officials, law enforcement, or program managers in targeted ways. A similar dashboarding approach supported by AI solutions has been used in Colorado to monitor HIV surveillance and outreach interventions and their impact after implementation (Snyder et al., 2016). This tool integrated data on regional resources with near-real-time visualization of
complex information to support program planning, patient management, and resource allocation.
AI has already made inroads into environmental and occupational health by leveraging data generated by sensors, nanotechnology, and robots. For example, water-testing sensors with AI tools have been paired with microscopes to detect bacterial contamination in treatment plants through hourly water sampling and analysis. This significantly reduces the time traditionally spent sending water samples for laboratory testing and lowers the cost of certain automated systems (Leider, 2018).
In a similar fashion, remote sensing from meteorological sensors, combined with geographic information systems, has been used to measure and analyze air pollution patterns in space and over time. This evolving field of inquiry has been termed geospatial AI because it combines innovations in spatial science with the rapid growth of methods in AI, including machine learning and deep learning. In another approach to understanding environmental factors, images of Google Street View have been analyzed using deep learning mechanisms to analyze urban greenness as a predictor and enabler of exercise (e.g., walking, cycling) (Lu, 2018).
Robots enabled by AI technology have been successfully deployed in a variety of hazardous occupational settings to improve worker safety and prevent injuries that can lead to costly medical treatments or short- and long-term disability. Robots can replace human labor in highly dangerous or tedious jobs that are fatiguing and could represent health risks to workers—reducing injuries and fatalities. Similarly, the construction industry has relied on AI-enabled robots for handling hazardous materials, handling heavy loads, working at elevation or in hard-to-reach places, and completing tasks that require difficult work postures, risking injury (Hsiao et al., 2017). Some of these robotic deployments are replacing human labor; in other instances, humans collaborate with robots to carry out such tasks. Hence, the deployment of robots requires workers to develop new skills in directing and managing robots and managing interactions between different types of robots or equipment, all operating in dynamic work environments.
Coordination and payment for care in the United States is highly complex. It involves the patient, providers, health care facilities, laboratories, hospitals, pharmacies, benefit administrators, payers, and others. Before, during, and after a patient encounter, administrative coordination occurs around scheduling, billing,
and payment. Collectively, we call this the administration of care, or administrative workflow. AI might be used in this setting in the form of machine learning models that can work alongside administrative personnel to perform mundane, repetitive tasks in a more efficient, accurate, and unbiased fashion.
Newer methods in AI, known generally as deep learning, have advantages over traditional machine learning, especially in the analysis of text data. The use of deep learning is particularly powerful in a workflow where a trained professional reviews narrative data and makes a decision about a clear action plan. A vast number of prior authorizations exist, with decisions and underlying data; these data provide the ideal substrate to train an AI model. Textual information used for prior authorization can be used for training a deep learning model that reaches or even exceeds the ability of human reviewers, perhaps with even more consistency than human reviewers. AI for health care administration will likely be utilized extensively, even beyond those solutions deployed for direct clinical care. “While all AI solutions give some false positives and false negatives, in administration, these will mostly produce annoyances, but should be monitored closely to ensure that patient safety is never impacted.”
As AI solutions become more sophisticated and automated, it may happen that a variety of AI methodologies will be deployed for a given solution. For example, a request by a patient to refill a prescription might involve speech recognition or AI chatbots, a rules-based system to determine if prior authorization is required, automated provider outreach, and a deep learning system for prior authorization when needed. It is worth noting that deep learning systems already drive many of today’s speech recognition, translation, and chatbot programs.
We provide illustrative (non-exhaustive) examples in Table 3-2 for different types of applications of AI solutions to support routine health care administration processes.
Most health plans and pharmacy benefit managers require prior authorization of devices, durable equipment, labs, and procedures. The process includes the submission of patient information along with the proposed request, along with justification. Determinations require professional skill, analysis, and judgment. Automating this process can reduce biased decisions and improve speed, consistency, and quality of decisions.
There are a number of different ways that AI is applied today. For example, AI could simply be used to sort cases to the appropriate level of reviewer (e.g., nurse practitioner, physician advisor, medical director). Or, AI could identify and highlight the specific, relevant information in long documents or narratives to
| Topic | Example Opportunity | Value | Output/Intervention | Data |
|---|---|---|---|---|
| Prior authorization (Rowley, 2016; Wince, 2018; Zieger, 2018) | Automate decisions on drugs, labs, or procedures | Reduced cost, efficiency, improved quality, reduce bias | Authorization or rejection | Relevant patient electronic health record (EHR) data |
| Fraud, waste (Bauder and Khoshgoftaar, 2017; da Rosa, 2018; He et al., 1997) | Identify appropriate or fraudulent claims | Reduced cost, improved care | Identification of targets for investigation | Provider claims data |
| Provider directory management | Maintain accurate information on providers | Reduced patient frustration through accurate provider availability, avoid Medicare penalties | Accurate provider directory | Provider data from many sources |
| Adjudication | Determine if a hospital should be paid for an admission versus observation | Improved compliance, accurate payments | Adjudication decision | Relevant patient EHR record data |
| Automated coding (Huang et al., 2019; Li et al., 2018; Shi et al., 2017; Xiao et al., 2018) | Automate ICD-10a coding of patient encounters | Improved compliance, accurate payments | ICD-10 coding | Relevant patient EHR record data |
| Chart abstraction (Gehrmann et al., 2017) | Summarize redundant data into a coherent narrative or structured variables | Reduced cost, efficiency, improved quality | Accurate, clean narrative/problem list | Relevant patient EHR record data |
| Patient scheduling (Jiang et al., 2018; Nelson et al., 2019; Sharma, 2016) | Identify no-shows and optimize scheduling | Improved patient satisfaction, faster appointments, provider efficiency | Optimized physician schedule | Scheduling history, EHR data |
a The ICD-10-CM (International Classification of Diseases, 10th Revision, Clinical Modification) is a system used by physicians and other health care providers to classify and code all diagnoses, symptoms, and procedures recorded in conjunction with hospital care in the United States.
produce information regarding estimated costs or benefit/risk assessment to aid a consumer in a decision.
Automation of prior authorization could reduce administrative costs, frustration, and idle time for provider and payer alike. Ultimately, such a process could lead to fewer appeals as well, which is a costly outcome of any prior authorization decision. A prior authorization model would need to work in near real time, because the required decisions are typically time sensitive.
The use of prior authorization limits liability, but AI implementation could create some liability risk. Some AI models are or become biased, and examples of this have been featured in the news recently. There has been coverage of AI models discriminating on names associated with particular ethnic groups, mapping same-day delivery routes to avoid high-crime areas, discriminating in lending practices, etc. (Hamilton, 2019; Ingold and Soper, 2016; Ward-Foxton, 2019; Williamson-Lee, 2018).
Coding is an exacting, expert-driven process that extracts information from EHRs for claims submissions and risk adjustment. These are called ICD-10 codes, from the International Classification of Diseases, 10th Revision (ICD-10). It is a human labor–intensive process that requires an understanding of language, expertise in clinical terminology, and a nuanced, expert understanding of administrative coding of medical care. Of note, codes are often deleted and added, and their assignment to particular medical descriptions often changes. Computer-assisted coding has existed for more than a decade; it typically has used more traditional, semantic-based NLP. Proximity and other methods are used to identify appropriate codes to assist or pre-populate manual coding.
The accuracy of coding is very important, and the process of assigning an unspecified number of multiple labels to an event is a complex one. It can lead to false negatives and false positives. False negatives in coding may result in a denial of reimbursement. False positives may lead to overcharges, compliance issues, and excess cost to payers.
There are opportunities for AI techniques to be applied to this administrative coding. Notes and patient records can be vectorized within this space, using tools such as Word2vec, so that they might be used in deep learning and other AI predictive models along with a wide variety of structured data, such as medication orders, laboratory tests, and vital signs.
Because of the complexity of multilabel prediction, humans will have to supervise and review the process for the foreseeable future. This also increases the need for transparency in the algorithmic outputs as part of facilitating human review. This is especially important in the review of coding in long EHR narratives. Transparency will also be helpful for monitoring automated processes because treatments and medical standards change over time and algorithms have to be retrained. This is a topic discussed in more detail in Chapter 6.
In the short term, AI coding solutions may help coders and create checks for payers. In the long term, increasing automation may be achieved for some or many types of encounters/hospitalizations. This automation will be reliant on
data comprehensiveness, public acceptance, algorithm accuracy, and appropriate regulatory frameworks.
The uses of AI technologies in research are broad; they frequently drive the development of new machine learning algorithms. To narrow this massive landscape, we focus our discussion on research institutions with medical training facilities. These medical school–affiliated health care providers often house massive and multiple large-scale data repositories (such as biobanks, Digital Imaging and Communications in Medicine or DICOM systems, and EHR systems).
Research with EHR data offers promising opportunities to advance biomedical research and improve health care by interpreting structured, unstructured (e.g., free text), genomic, and imaging data.
Extracting practical information from EHR data is challenging because such data are highly dimensional, heterogeneous, sparse, and often of low quality (Jensen et al., 2012; Zhao et al., 2017). Nevertheless, AI technologies are being applied to EHR data. AI techniques used on these data include a vast array of data mining approaches, from clustering and association rules to deep learning. We focus our discussion in the sections below on areas of present key importance.
Deep learning algorithms rely on the large quantities of data and massive computer resources, both of which are newly possible in this era. Deep learning can identify underlying patterns in data well beyond the pattern-perceiving capacities of humans. Deep learning and its associated techniques have become popular in many data-driven fields of research. The principal difference between deep and traditional (i.e., shallow) machine learning paradigms lies in the ability of deep learning algorithms to construct latent data representations from a large number of raw features, often through deep architectures (i.e., many layers) of artificial neural networks. This “unsupervised feature extraction” sometimes permits highly accurate predictions. Recent research on EHR data has shown that deep learning predictive models can outperform traditional clinically used predictive
models for predicting early detection of heart failure onset (Choi et al., 2017), various cancers (Miotto et al., 2016), and onset of and weaning from intensive care unit interventions (Suresh et al., 2017).
Nevertheless, the capacity of deep learning is a double-edged sword. The downside of deep learning comes from exactly where its superiority to other learning paradigms originates—that is, its ability to build and learn features. Model complexity means that human interpretability of deep learning models is almost nonexistent, because it is extremely hard to infer how the model makes its predictions so well. Deep learning models are “black box” models, where the internal workings of the algorithms remain unclear or mysterious to users of these models. As in other black box AI approaches, there is significant resistance to implementing deep learning models in the health care delivery process.
Detecting abnormal brain structure is much more challenging for humans and machines than detecting a broken bone or a fracture. Exciting things are being done in this area with deep learning. One recent study predicted age from brain images (Cole et al., 2017). Multimodal image recognition analysis has discovered novel impairments not visible from a single view of the brain (e.g., structural MRI versus functional MRI) (Plis et al., 2018). Companies such as Avalon AI1 are commercializing this type of work.
Effective AI use does not always require new modeling techniques. Some work at Massachusetts General Hospital in Boston uses a large selection of images and combines established machine learning techniques with mature brain-image analysis tools to explore what is normal for a child’s developing brain (NITRC, 2019; Ou et al., 2017). Other recent applications of AI to radiology data include using machine learning on electrocardiogram data to characterize types of heart failure (Sanchez-Martinez et al., 2018). In addition, AI can aid in reducing noise in real images (e.g., endoscopy) via “adversarial training.” It can smooth out erroneous signals in images to enhance prediction accuracy (Mahmood et al., 2018). AI is also being applied to moving images; gait analysis has long been done by human observation alone, but it now can be performed with greater accuracy by AI that uses video and sensor data. These techniques are being used to detect Parkinson’s disease, to improve geriatric care, for sports rehabilitation, and in other areas (Prakash et al., 2018). AI can also improve video-assisted surgery, for example, by detecting colon polyps in real time (Urban et al., 2018)
___________________
AI can assist in analyzing clinical practice patterns from EHR data to develop clinical practice models before such research can be distilled into literature or made widely available in clinical decision support tools. This notion of “learning from the crowd” stems from Condorcet’s jury theorem, which states that the average decisions of a crowd of unbiased experts are more correct than any individual’s decisions. (Think of the “jellybeans in a jar” challenge—the average of everyone’s guesses is surprisingly close to the true number.)
The most straightforward approach uses association rule mining to find patterns, but this tends to find many false associations (Wright et al., 2010). Therefore, some researchers have attempted to use more AI approaches such as Bayesian network learning and probabilistic topic modeling (Chen et al., 2017; Klann et al., 2014).
A phenotype refers to an observable trait of an organism, resulting from its genetic code and surrounding environment, and the interactions between them. It is becoming increasingly popular to identify patient cohorts by trait for clinical and genomic research. Although EHR data are often incomplete and inaccurate, they do convey enough information for constructing clinically relevant sets of observable characteristics that define a disease, or a phenotype.
EHR phenotyping uses the information in a patient’s health records to infer the presence of a disease (or lack thereof). This is done by using algorithms that apply predetermined rules, machine learning, and statistical methods to derive phenotypes.
Rule-based phenotyping is time-consuming and expensive, and so applying machine learning methods to EHR data makes sense. The principal mechanism in this approach transforms raw EHR data (e.g., diagnostic codes, laboratory results, clinical notes) into meaningful features that can predict the presence of a disease. Machine learning–based solutions mine both structured and unstructured data stored in EHRs for phenotyping. NLP algorithms have been used in research to extract relevant features from EHRs. For example, Yu et al. (2015, 2017) used NLP to identify candidate clinical features from a pool of comprehensive medical concepts found in publicly available online knowledge sources.
When the set of features is extracted, different classification algorithms can be used to predict or classify the phenotype. Choice of the classification algorithm in supervised learning relies on the characteristics of the data on which the algorithm will be trained and tested. Feature selection and curation of gold-standard training sets includes two rate-limiting factors. Curating annotated
datasets to train supervised algorithms requires involvement from domain experts, which hampers generalizability and scalability of phenotyping algorithms. As a result, the classification algorithms for phenotyping research has been limited so far to regularized algorithms that can address overfitting, which is what happens when an algorithm that uses many features is trained on small training sets.
Regularized classifiers penalize more features in favor of model parsimony. To overcome this limitation, new research is investigating the application of hybrid approaches (known as semi-supervised) to create semi-automatically annotated training sets and the use of unsupervised methods to scale up EHR phenotyping. The massive amounts of data in EHRs, if processed through deep neural networks, may soon permit the computation of phenotypes from a wider vector of features.
As mentioned previously, large national initiatives are now combining biobanks of genomic information with these phenotypes. For example, eMerge frequently uses genomic analyses such as genome-wide association studies in combination with phenotyping algorithms to define the gold-standard cohort or to study genetic risk factors in the phenotyped population (Gottesman et al., 2013; McCarty et al., 2011).
Machine learning has the capacity to make drug discovery faster, cheaper, and more effective. Drug designers frequently apply machine learning techniques to extract chemical information from large compound databases and to design drugs with important biological properties. Machine learning can also improve drug discovery by permitting a more comprehensive assessment of cellular systems and potential drug effects. With the emergence of large chemical datasets in recent years, machine and deep learning methods have been used in many areas (Baskin et al., 2016; Chen et al., 2018; Lima et al., 2016; Zitnik et al., 2018). These include
Large chemical databases have made drug discovery faster and cheaper. EHR databases have brought millions of patients’ lives into the universe of statistical learning. Research initiatives to link structured patient data with biobanks,
radiology images, and notes are creating a rich and robust analytical playground for discovering new knowledge about human disease. Deep learning and other new techniques are creating solutions that can operate on the scale required to digest these multiterabyte datasets. The accelerating pace of discovery will probably challenge the research pipelines that translate new knowledge back into practice.
Although it is difficult to predict the future in a field that is changing so quickly, we offer the following ideas about how AI will be used and considerations for optimizing the success of AI for health.
AI will change health care delivery less by replacing clinicians than by supporting or augmenting clinicians in their work. AI will support clinicians with less training in performing tasks currently relegated to specialists. It filters out normal or noncomplex cases so that specialists can focus on a more challenging case load.
AI will support humans in tasks that suffer from inattention, cause fatigue, and are physically difficult to perform. AI will substitute for humans by facilitating screening and evaluation in areas with limited access to medical expertise. Some AI tools, like those for self-management or population health support, will be useful in spite of lower accuracy.
When an AI tool assists human cognition, it will initially need to explain the connections it has drawn, allowing for an understanding of a pathway to effects. With sufficient accuracy, humans will begin to trust the AI output and will require less transparency and explanation. In situations where AI substitutes for medical expertise, the workflow should include a human in the loop to identify misbehavior and provide accountability (Rahwan, 2018).
The central focus of health care will continue to expand from health care delivery systems to a dispersed model that aggregates information about behavior, traits, and environment in addition to medical symptoms and test results.
Market forces and privacy concerns or regulations may impede data sharing and analysis (Roski et al., 2014). Stakeholders will need to creatively balance
competing demands. More national-level investments similar to the National Institutes of Health’s All of Us program can facilitate and accelerate these partnerships. More ethical and legal guidelines are needed for successful data sharing and analysis (see Chapters 1 and 7).
AI tools will continue to be developed by industry, research, government, and individuals. With emerging standards such as SMART on FHIR (Substitutable Medical Apps, Reusable Technology on Fast Healthcare Interoperability Resource), these tools will increasingly be implemented across platforms regardless of the EHR vendor, brand of phone, etc. This will most likely speed the adoption of AI.
AI will probably discover associations that have not yet been detected by humans and make predictions that differ from prevailing knowledge and expertise. As a result, some currently accepted practices may be abandoned, and best practice guidelines will be adjusted.
If the output of AI systems is going to influence international guidelines, developers of the applications will require fuller and more representative datasets for training and testing.
The dissemination of innovation will occur rapidly, which on the one hand may advance the adoption of new scientific knowledge but on the other may encourage the rushed adoption of innovation without sufficient evidence.
As AI increasingly infiltrates the field of health, biases inherent in clinical practice will appear in the datasets used in AI models. The discovery of existing bias will open the door to changing practices, but it may also produce public disillusionment and mistrust.
Important growth areas for AI include platforms designed for and accessible by the people most in need of additional support. This includes older adults, people living with multiple comorbid conditions, and people in low-resource settings.
Accenture. 2018. Consumer survey on digital health: US results. https://www.accenture.com/t20180306T103559Z__w__/us-en/_acnmedia/PDF-71/accenture-health-2018-consumer-survey-digital-health.pdf (accessed November 12, 2019).
Alzheimers.net. 2017. Alzheimer’s statistics. https://www.alzheimers.net/resources/alzheimers-statistics (accessed March 26, 2019).
Barrett, M., V. Combs, J. G. Su, K. Henderson, M. Tuffli, and AIR Louisville Collaborative. 2018. AIR Louisville: Addressing asthma with technology, crowdsourcing, cross-sector collaboration, and policy. Health Affairs (Millwood) 37(4):525–534.
Baskin, I. I., D. Winkler, and I. V. Tetko. 2016. A renaissance of neural networks in drug discovery. Expert Opinion on Drug Discovery 11(8):785–795.
Bauder, R. A., and T. M. Khoshgoftaar. 2017. Medicare fraud detection using machine learning methods. In 2017 16th IEEE International Conference on Machine Learning and Applications. https://doi.org/10.1109/icmla.2017.00-48.
Best, J. 2017. AI that knows you’re sick before you do: IBM’s five-year plan to remake healthcare. ZDNet. https://www.zdnet.com/article/ai-that-knows-youre-sick-before-you-do-ibms-five-year-plan-to-remake-healthcare (accessed November 12, 2019).
Borzykowski, B. 2016. Truth be told, we’re more honest with robots. BBC WorkLife. https://www.bbc.com/worklife/article/20160412-truth-be-told-were-more-honest-with-robots (accessed November 12, 2019).
Bostic, B. 2018. Using artificial intelligence to solve public health problems. Beckers Hospital Review. https://www.beckershospitalreview.com/healthcare-information-technology/using-artificial-intelligence-to-solve-public-health-problems.html (accessed November 12, 2019).
Brown, D. 2018. RSNA 2018: Researchers use AI to predict cancer survival, treatment response. AI in Healthcare News. https://www.aiin.healthcare/topics/research/research-ai-cancer-survival-treatment-response (accessed November 12, 2019).
Bryant, M. 2018. Hospitals turn to chatbots, AI for care. Healthcare Dive. https://www.healthcaredive.com/news/chatbots-ai-healthcare/516047 (accessed November 12, 2019).
Cabitza, F., R. Rasoini, and G. F. Gensini. 2017. Unintended consequences of machine learning in medicine. JAMA 318(6):517–518.
Cappella, J. N. 2017. Vectors into the future of mass and interpersonal communication research: Big data, social media, and computational social science. Human Communication Research 43(4):545–558.
CDC (Centers for Disease Control and Prevention). 2018. Prevalence of disabilities and health care access by disability status and type among adults—United States, 2016. Morbidity and Mortality Weekly Report 67(32):882–887.
CDC Foundation. 2019. What is public health? https://www.cdcfoundation.org/what-public-health (accessed November 12, 2019).
Chen, H., O. Engkvist, Y. Wang, M. Olivecrona, and T. Blaschke. 2018. The rise of deep learning in drug discovery. Drug Discovery Today 23(6):1241–1250.
Chen, J. H., M. K. Goldstein, S. M. Asch, L. Mackey, and R. B. Altman. 2017. Predicting inpatient clinical order patterns with probabilistic topic models vs conventional order sets. Journal of the American Medical Informatics Association 24(3):472–480.
Choi, E., S. Biswal, B. Malin, J. Duke, W. F. Stewart, and J. Sun. 2017. Generating multi-label discrete electronic health records using generative adversarial networks. arXIV. http://arxiv.org/abs/1703.06490 (accessed December 7, 2019).
Cole, J. H., R. P. K. Poudel, D. Tsagkrasoulis, M. W. A. Caan, C. Steves, T. D. Spector, and G. Montana. 2017. Predicting brain age with deep learning from raw imaging data results in a reliable and heritable biomarker. NeuroImage 163(December):115–124.
Contreras, I., and J. Vehi. 2018. Artificial intelligence for diabetes management and decision support: Literature review. Journal of Medical Internet Research 20(5):e10775.
da Rosa, R. C. 2018. An Evaluation of Unsupervised Machine Learning Algorithms for Detecting Fraud and Abuse in the US Medicare Insurance Program. Ph.D. dissertation, Florida Atlantic University, Boca Raton. https://pqdtopen.proquest.com/doc/2054014362.html?FMT=ABS (accessed November 12, 2019).
Dankwa-Mullan, I., M. Rivo, M. Sepulveda, Y. Park, J. Snowdon, and K. Rhee. 2019. Transforming diabetes care through artificial intelligence: The future is here. Population Health Management 22(3). http://doi.org/10.1089/pop.2018.0129.
Du, J., Y. Zhang, J. Luo, Y. Jia, Q. Wei, C. Tao, and H. Xu. 2018. Extracting psychiatric stressors for suicide from social media using deep learning. BMC Medical Informatics and Decision Making 18(Suppl 2):43.
Elliott, J. H., T. Turner, O. Clavisi, J. Thomas, J. P. Higgins, C. Mavergames, and R. L. Gruen. 2014. Living systematic reviews: An emerging opportunity to narrow the evidence-practice gap. PLoS Medicine 11(2):e1001603.
Family Caregiver Alliance. 2019. Caregiver Statistics: Demographics. National Center on Caregiving. https://www.caregiver.org/caregiver-statistics-demographics (accessed March 26, 2019).
Figge, H. 2018. Deploying artificial intelligence against infectious disease. U.S. Pharmacist 43(3):21–24.
Fiorini, N., K. Canese, G. Starchenko, E. Kireev, W. Kim, V. Miller, M. Osipov, M. Kholodov, R. Ismagilov, S. Mohan, J. Ostell, and Z. Lu. 2018. Best Match: New relevance search for PubMed. PLoS Biology 16(8):e2005343.
Fitzpatrick, K. K., A. Darcy, and M. Vierhile. 2017. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agenda (Woebot): A randomized controlled trial. JMIR Mental Health 4(2):e19.
Gehrmann, S., F. Dernoncourt, Y. Li, E. T. Carlson, J. T. Wu, J. Welt, J. Foote, E. T. Moseley, D. W. Grant, P. D. Tyler, and L. A. Celi. 2017. Comparing rule-based and deep learning models for patient phenotyping. arXiv preprint. 1703.08705.
Gelichgerrcht, E., B. Munsell, S. Bhatia, W. Vandergrift, C. Rorden, C. McDonalid, J. Edwards, R. Kuzniecky, L. Bonilha. 2018. Deep learning applied to whole-brain connectome to determine seizure control after epilepsy surgery. Epilepsia 59(9):1643–1654.
Gottesman, O., H. Kuivaniemi, G. Tromp, W. A. Faucett, R. Li, T. A. Manolio, S. C. Sanderson, J. Kannry, R. Zinberg, M. A. Basford, M. Brilliant, D. J. Carey, R. L. Chisholm, C. G. Chute, J. J. Connolly, D. Crosslin, J. C. Denny, C. J. Gallego, J. L. Haines, H. Hakonarson, J. Harley, G. P. Jarvik, I. Kohane, I. J. Kullo, E. B. Larson, C. McCarty, M. D. Ritchie, D. M. Roden, M. E. Smith, E. P. Böttinger, M. S. Williams, and eMERGE Network. 2013. The electronic medical records and genomics (eMERGE) network: Past, present, and future. Genetics in Medicine 15(10):761–771.
Hamilton, E. 2019. AI perpetuating human bias in the lending space. https://www.techtimes.com/articles/240769/20190402/ai-perpetuating-human-bias-in-the-lending-space.htm (accessed November 12, 2019).
Hammana, I., L. Lepanto, T. Poder, C. Bellemare, and M. S. Ly. 2015. Speech recognition in the radiology department:A systematic review. Health Informatics Management 44(2):4–10.
He, H., J. Wang, W. Graco, and S. Hawkins. 1997. Application of neural networks to detection of medical fraud. Expert Systems with Applications 13(4):329–336.
Hindorff, L. A., V. L. Bonham, L. C. Brody, M. E. C. Ginoza, C. M. Hutter, T. A. Manolio, and E. D. Green. 2018. Prioritizing diversity in human genomics research. Nature Reviews Genetics 19:175–185.
Hixson, J., D. Barnes, K. Parko, T. Durgin, S. Van Bebber, A. Graham, and P. Wicks. 2015. Patients optimizing epilepsy management via an online community. Neurology 85(2):129–136. https://doi.org/10.1212/WNL.0000000000001728.
Hsiao, H., H. Choi, J. Sammarco, S. Earnest, D. Castillo, and G. Hill. 2017. NIOSH presents: An occupational safety and health perspective on robotics applications in
the workplace. https://blogs.cdc.gov/niosh-science-blog/2017/12/05/robot_safety_conf (accessed December 13, 2019).
Huang, J., C. Osorio, and L. W. Sy. 2019. An empirical evaluation of deep learning for ICD-9 code assignment using MIMIC-III clinical notes. Computer Methods and Programs in Biomedicine 177:141–153.
Imler, T. D., J. Morea, T. F. Imperiale. 2014. Clinical decisions support with natural language processing facilitates determination of colonscopy surveillance intervals. Clinical Gastroenterology and Hepatology 12(7):1130–1136.
Ingold, D., and S. Soper. 2016. Amazon doesn’t consider the race of its customers. Should it? Bloomberg. https://www.bloomberg.com/graphics/2016-amazon-same-day (accessed November 12, 2019).
Jackson, G. L., B. J. Powers, R. Chatterjee, J. P. Bettger, A. R. Kemper, V. Hasselblad, R. J. Dolor, J. Irvine, B. L. Heidenfelder, A. S. Kendrick, R. Gray, and J. W. Williams. 2013. The patient-centered medical home: A systematic review. Annals of Internal Medicine 158:169–178.
Jensen, P. B., L. J. Jensen, and S. Brunak. 2012. Mining electronic health records: Towards better research applications and clinical care. Nature Reviews Genetics 13(6):395–405.
Jiang, S., K. S. Chin, and K. L. Tsui. 2018. A universal deep learning approach for modeling the flow of patients under different severities. Computer Methods and Programs in Biomedicine 154:191–203.
Joffe, E., O. Havakuk, J. R. Herskovic, V. L. Patel, and E. V. Bernstam. 2012. Collaborative knowledge acquisition for the design of context-aware alert systems. Journal of the American Medical Informatics Association 219(6):988–994.
Johnson, J. 2018. Designing technology for an aging population. Presentation at Stanford Center on Longevity meeting at Tresidder Oak Lounge. http://longevity.stanford.edu/2018/10/17/designing-technology-for-an-aging-population (accessed December 7, 2019).
Kasthurirathne, S. N., J. R. Vest, N. Menachemi, P. K. Halverson, and S. J. Grannis. 2018. Assessing the capacity of social determinants of health data to augment predictive models identifying patients in need of wraparound social services. Journal of the American Medical Informatics Association 25(1):47–53.
Kindig, D., and G. Stoddart. 2003. What is population health? American Journal of Public Health 93(3):380–383.
Klann, J. G., P. Szolovits, S. M. Downs, and G. Schadow. 2014. Decision support from local data: Creating adaptive order menus from past clinician behavior. Journal of Biomedical Informatics 48:84–93.
Kostokova, P. 2013. A roadmap to integrated digital public health surveillance: The vision and the challenges. In WWW ‘13 Companion Proceedings of the
22nd International Conference on World Wide Web. New York: CMS. Pp. 687–694. https://www.researchgate.net/publication/250963354_A_roadmap_to_integrated_digital_public_health_surveillance_The_vision_and_the_challenges (accessed November 12, 2019).
Laranjo, L., A. G. Dunn, H. L. Tong, A. B. Kocaballi, J. Chen, R. Bashir, D. Surian, B. Gallego, F. Magrabi, and A. Coiera. 2018. Conversational agents in healthcare: A systematic review. Journal of the American Medical Informatics Association 25(9):1248–1258.
Lazer, D., R. Kennedy, G. King, and A. Vespignani. 2014. The parable of Google flu: Traps in big data analysis. Science 343(6176):1203–1205.
Leider, N. 2018. AI could protect public health by monitoring water treatment systems. AI in Healthcare News. https://www.aiin.healthcare/topics/artificial-intelligence/ai-public-health-monitoring-water-treatment (accessed November 12, 2019).
Li, M., Z. Fei, M. Zeng, F. Wu, Y. Li, Y. Pan, and J. Wang. 2018. Automated ICD-9 coding via a deep learning approach. IEEE/ACM Transactions on Computational Biology and Bioinformatics 16(4):1193–1202. https://doi.org/10.1109/TCBB.2018.2817488.
Lima, A. N., E. A. Philot, G. H. G. Trossini, L. P. B. Scott, V. G. Maltarollo, and K. M. Honorio. 2016. Use of machine learning approaches for novel drug discovery. Expert Opinion on Drug Discovery 11(3):225–239.
Liu, F., Z. Zhou, A. Samsonov, D. Blankenbaker, W. Larison, A. Kanarek, K. Lian, S. Kambhampati, and R. Kijowski. 2018. Deep learning approach for evaluating knee MR images: Achieving high diagnostic performance for cartilage lesion detection. Radiology 289(1):160–169.
Lu, Y. 2018. The association of urban greenness and walking behavior: Using Google Street View and deep learning techniques to estimate residents’ exposure to urban greenness. International Journal of Environmental Health Research and Public Health 15:1576.
Maharana, A., and E. O. Nsoesie. 2018. Use of deep learning to examine the association of the built environment with prevalence of neighborhood adult obesity. JAMA Network Open 1(4):e181535. https://doi.org/10.1001/jamanetworkopen.2018.1535.
Mahmood, F., R. Chen, and N. J. Durr. 2018. Unsupervised reverse domain adaptation for synthetic medical images via adversarial training. IEEE Transactions on Medical Imaging 37(12):2572–2581. https://doi.org/10.1109/TMI.2018.2842767.
Matheson, R. 2018. Machine-learning system determines the fewest, smallest doses that could still shrink brain tumors. MIT News. http://news.mit.edu/2018/artificial-intelligence-model-learns-patient-data-cancer-treatment-less-toxic-0810.
McCarty, C. A., R. L. Chisholm, C. G. Chute, I. J. Kullo, G. P. Jarvik, E. B. Larson, R. Li, D. R. Masys, M. D. Ritchie, D. M. Roden, J. P. Struewing, W. A. Wolf, and eMERGE Team. 2011. The eMERGE network: A consortium of biorepositories linked to electronic medical records data for conducting genomic studies. BMC Medical Genomics 4:13.
McCoy, A. B., E. J. Thomas, M. Krousel-Wood, and D. F. Sittig. 2014. Clinical decision support alert appropriateness: A review and proposal for improvement. Ochsner Journal 14(2):195–202.
Miotto, R., L. Li, B. A. Kidd, and J. I. T. Dudley. 2016. Deep patient: An unsupervised representation to predict the future of patients from the electronic health records. Scientific Reports 6:26094.
Mooney, S. J., and V. Pejaver. 2018. Big data in public health: Terminology, machine learning, and privacy. Annual Review of Public Health 39:95–112.
Moyle, W., C. J. Jones, J. E. Murfield, L. Thalib, E. R. A. Beattie, D. K. H. Shum, S. T. O’Dwyer, M. C. Mervin, and B. M. Draper. 2017. Use of a robotic seal as a therapeutic tool to improve dementia symptoms: A cluster-randomized controlled trial. Journal of the American Medical Directors Association 18(9):766–773.
Nahum-Shani, I., E. B. Hekler, and D. Spruijt-Metz. 2015. Building health behavior models to guide the development of just-in-time adaptive interventions: A pragmatic framework. Health Psychology 34S:1209–1219. https://doi.org/10.1037/hea0000306.
Nelson, A., D. Herron, G. Rees, and P. Nachev. 2019. Predicting scheduled hospital attendance with artificial intelligence. NPJ Digital Medicine 2(1):26.
Newmarker, C. 2018. Digital surgery touts artificial intelligence for the operating room. Medical Design & Outsourcing. https://www.medicaldesignandoutsourcing.com/digital-surgery-touts-artificial-intelligence-for-the-operating-room (accessed November 12, 2019).
NITRC (Neuroimaging Tools & Resources Collaboratory). 2019. MGH Neonatal/Pediatric ADC Atlases. https://www.nitrc.org/projects/mgh_adcatlases (accessed October 18, 2019).
Ou, Y., L. Zöllei, K. Retzepi, V. Castro, S. V. Bates, S. Pieper, K. P. Andriole, S. N. Murphy, R. L. Gollub, and P. E. Grant. 2017. Using clinically acquired MRI to construct age-specific ADC atlases: Quantifying spatiotemporal ADC changes from birth to 6-year old. Human Brain Mapping 38(6):3052–3068.
Parakh, A., H. Lee, J. H. Lee, B. H. Eisiner, D. V. Sahani, and S. Do. 2019. Urinary stone detection on CT images using deep convolutional neural networks: Evaluation of model performance and generalization. Radiology: Artificial Intelligence 1(4). https://doi.org/10.1148/ryai.2019180066.
Payne, T. H., L. E. Hines, R. C. Chan, S. Hartman, J. Kapusnik-Uner, A. L. Russ, B. W. Chaffee, C. Hartman, V. Tamis, B. Galbreth, and P. A. Glassman. 2015. Recommendations to improve the usability of drug-drug interaction clinical decision support alerts. Journal of the American Medical Informatics Association 22(6):1243–1250.
Plis, S. M., F. Amin, A. Chekroud, D. Hjelm, E. Damaraju, H. J. Lee, J. R. Bustillo, K. Cho, G. D. Pearlson, and V. D. Calhoun. 2018. Reading the (functional) writing on the (structural) wall: Multimodal fusion of brain structure and function via a deep neural network based translation approach reveals novel impairments in schizophrenia. NeuroImage 181:734–747.
Poon, H., C. Quirk, K. Toutanova, and S. Wen-tau Yih. 2018. AI for precision medicine. Project Hanover. https://hanover.azurewebsites.net/#machineReading (accessed November 12, 2019).
Prakash, C., R. Kumar, and N. Mittal. 2018. Recent developments in human gait research: Parameters, approaches, applications, machine learning techniques, datasets and challenges. Artificial Intelligence Review 49(1):1–40.
Rabbitt, S. M., A. E. Kazdin, and B. Scassellati. 2015. Integrating socially assistive robotics into mental healthcare interventions: Applications and recommendations for expanded use. Clinical Psychology Review 35:35–46.
Rahwan, I. 2018. Society-in-the-loop: Programming the algorithmic social contract. Ethics and Information Technology 20(1):5–14. https://doi.org/10.1007/s10676-017-9430-8.
Reps, J. M., M. J. Schuemie, M. A. Suchard, P. B. Ryan, and P. R. Rijnbeek. 2018. Design and implementation of a standardized framework to generate and evaluate patient-level prediction models using observational healthcare data. Journal of the American Medical Informatics Association 25(8):969–975.
Roski, J., G. Bo-Linn, and T. Andrews. 2014. Creating value in healthcare through big data: Opportunities and policy implications. Health Affairs 33(7):1115–1122.
Roski, J., B. Gillingham, E. Juse, S. Barr, E. Sohn, and K. Sakarcan. 2018. Implementing and scaling artificial intelligence solutions: Considerations for policy makers and decision makers. Health Affairs (Blog). https://www.healthaffairs.org/do/10.1377/hblog20180917.283077/full (accessed November 12, 2019).
Rowley, R. 2016. Can AI reduce the prior authorization burden in healthcare? Health IT. https://hitconsultant.net/2016/07/11/34693/#.Xd6pD25Fw2w (accessed November 12, 2019).
Rubens, M., A. Ramaamoorthy, A. Saxena, and N. Shehadeh. 2014. Public health in the 21st century: The role of advanced technologies. Frontiers in Public Health 2:1–4.
Sanchez-Martinez, S., N. Duchateau, T. Erdei, G. Kunszt, S. Aakhus, A. Degiovanni, P. Marino, E. Carluccio, G. Piella, A. G. Fraser, and B. H. Bijnens. 2018. Machine learning analysis of left ventricular function to characterize heart failure with preserved ejection fraction. Circulation: Cardiovascular Imaging 11(4):e007138.
Schuler, A., A. Callahan, K. Jung, and N. Shah. 2018. Performing an informatics consult: Methods and challenges. Journal of the American College of Radiology 15:563–568.
Schulman-Marcus, J., S. Mookherjee, L. Rice, and R. Lyubarova. 2019. New approaches for the treatment of delirium: A case for robotic pets. American Journal of Medicine 132(7):781–782. https://doi.org/10.1016/j.amjmed.2018.12.039.
Sennaar, K. 2018. Machine learning medical diagnostics—4 current applications. Emerj Artificial Intelligence Research. https://emerj.com/ai-sector-overviews/machine-learning-medical-diagnostics-4-current-applications (accessed November 12, 2019).
Shademan, A., R. S. Decker, J. D. Opfermann, S. Leonard, A. Krieger, and P. C. W. Kim. 2016. Supervised autonomous robotic soft tissue surgery. Science Translational Medicine 8(337):337ra64.
Shanafelt, T. D., L. N. Dyrbye, C. Sinsky, O. Hasan, D. Satele, J. Sloan, and C. P. West. 2016. Relationship between clerical burden and characteristics of the electronic environment with physician burnout and professional satisfaction. Mayo Clinic Proceedings 91(7):836–848.
Sharma, M. 2016. Benefits of an AI-based patient appointments service for hospitals. Medium. https://medium.com/@HCITExpert/benefits-of-an-ai-based-patient-appointments-service-for-hospitals-by-msharmas-617fdb2498e0 (accessed November 12, 2019).
Shen, H., C. Wang, L. Xie, S. Zhou, L. Gu, and H. Xie. 2018. A novel remote-controlled robotic system for cerebrovascular intervention. International Journal of Medical Robotics and Computer Assisted Surgery 14(6):e1943.
Shi, H., P. Xie, Z. Hu, M. Zhang, and E. P. Xing. 2017. Towards automated ICD coding using deep learning. arXiv preprint. 1711.04075.
Shin, E. K., R, Mahajan, O. Akbilgic, and A. Shaban-Nejad. 2018. Sociomarkers and biomarkers: Predictive modeling in identifying pediatric asthma patients at risk of hospital revisits. Nature Medicine 50(1). doi: 10.1038/s41746-018-0056-y.
Snyder, L., D. McEwen, M. Thrun, and A. Davidson. 2016. Visualizing the local experience: HIV Data to Care Tool. Online Journal of Public Health Informatics 8(1):e39.
Spruijt-Metz, D., and W. Nilsen. 2014. Dynamic models of behavior for just-in-time adaptive interventions. IEEE Pervasive Computing 13(3):13–17.
Stahl, B. C., and M. Coeckelbergh. 2016. Ethics of healthcare robotics: Towards responsible research and innovation. Robotics and Autonomous Systems 86:152–161.
Suresh, H., N. Hunt, A. Johnson, L. A. Celi, P. Szolovits, and M. Ghassemi. 2017. Clinical intervention prediction and understanding with deep neural networks. In Proceedings of Machine Learning for Healthcare 2017, JMLR W&C Track Vol. 68, edited by F. Doshi-Velez, J. Fackler, D. Kale, R. Ranganath, B. Wallace, and J. Wiens. Pp. 322–337. http://mucmd.org/CameraReadySubmissions/65%5CCameraReadySubmission%5Cclinical-intervention-prediction%20(4).pdf (accessed May 13, 2020).
Urban, G., P. Tripathi, T. Alkayali, M. Mittal, F. Jalali, W. Karnes, and P. Baldi. 2018. Deep learning localizes and identifies polyps in real time with 96% accuracy in screening colonoscopy. Gastroenterology 155(4):1069–1078
Waljee, A. K., P. D. R. Higgins, and A. G. Singal. 2014. A primer on predictive models. Clinical and Translational Gastroenterology 5:e44.
Wang, D., A. Khosla, R. Gargeya, H. Irshad, and A. H. Beck. 2016. Deep learning for identifying metastatic breast cancer. arXIV. https://arxiv.org/abs/1606.05718 (accessed November 12, 2019).
Ward-Foxton, S. 2019. Reducing bias in AI models for credit and loan decisions. EE Times. https://www.eetimes.com/document.asp?doc_id=1334632# (accessed November 12, 2019).
Wetzel, R. C. 2018. Is it ethical to let patients develop relationships with robots? AI Medicine. http://ai-med.io/ethical-patients-relationships-robots (accessed November 12, 2019).
Wicks, P., D. Keininger, M. Massagli, C. de la Loge, C. Brownstein, J. Isojärvi, and J. Heywood. 2012. Perceived benefits of sharing health data between people with epilepsy on an online platform. Epilepsy & Behavior 23(1):16–23. https://doi.org/10.1016/j.yebeh.2011.09.026.
Wiggers, K. 2018. CB Insights: AI health care startups have raised $4.3 billion since 2013. VentureBeat. https://venturebeat.com/2018/09/13/cb-insights-ai-health-care-startups-have-raised-4-3-billion-since-2013 (accessed November 12, 2019).
Williamson-Lee, J. 2018. How machines inherit their creators’ biases: A. I. doesn’t have to be conscious to be harmful. Medium. https://medium.com/coinmonks/ai-doesnt-have-to-be-conscious-to-be-harmful-385d143bd311 (accessed November 12, 2019).
Wince, R. 2018. Why AI is the future of prior auths. insideBIGDATA. https://insidebigdata.com/2018/12/21/ai-future-prior-auths (accessed November 12, 2019).
Wright, A., E. S. Chen, and F. L. Maloney. 2010. An automated technique for identifying associations between medications, laboratory results and problems. Journal of Biomedical Informatics 43(6):891–901.
Xiao, C., E. Choi, and J. Sun. 2018. Opportunities and challenges in developing deep learning models using electronic health records data: A systematic review. Journal of the American Medical Informatics Association 25(10):1419–1428.
YellRobot. 2018. Robot pets for elderly and dementia patients. https://yellrobot.com/robot-pets-for-elderly (accessed December 7, 2019).
Young, S. D., W. Yu, and W. Wang. 2017. Toward automating HIV identification: Machine learning for rapid identification of HIV-related social media data. Journal of Acquired Immune Deficiency Syndrome 74(Suppl 2):128–131.
Yu, S., P. K. Liao, S. Y. Shaw, V. S. Gainer, S. E. Churchill, P. Szolovits, S. N. Murphy, I. S. Kohane, and T. Cai. 2015. Toward high-throughput phenotyping: Unbiased automated feature extraction and selection from knowledge sources. Journal of the American Medical Informatics Association 22(5):993–1000.
Yu, S., A. Chakrabortty, K. P. Liao, T. Cai, A. N. Ananthakrishnan, V. S. Gainer, S. E. Churchill, P. Szolovits, S. N. Murphy, I. S. Kohane, and T. Cai. 2017. Surrogate-assisted feature extraction for high-throughput phenotyping. Journal of the American Medical Informatics Association 24(e1):e143–e149.
Zauderer, M. G., A. Gucalp, A. S. Epstein, A. D. Seidman, A. Caroline, S. Granovsky, J. Fu, J. Keesing, S. Lewis, H. Co, J. Petri, M. Megerian, T. Eggebraaten, P. Bach, and M. G. Kris. 2014. Piloting IBM Watson Oncology within Memorial Sloan Kettering’s regional network. Journal of Clinical Oncology 32(15 Suppl):e17653.
Zellweger, M. J., A. Tsirkin, V. Vasilchenko, M. Failer, A. Dressel, M. E. Kleber, P. Ruff, and W. März. 2018. A new non-invasive diagnostic tool in coronary artery disease: Artificial intelligence as an essential element of predictive, preventive, and personalized medicine. EPMA Journal 9(3):235–247.
Zhao, J., P. Papapetrou, L. Asker, and H. Boström. 2017. Learning from heterogeneous temporal data in electronic health records. Journal of Biomedical Informatics 65(January):105–119.
Zhao, J., G. Wang, Z. Jiang, C. Jiang, J. Liu, J. Zhou, and J. Li. 2018. Robotic gastrotomy with intracorporeal suture for patients with gastric gastrointestinal stromal tumors located at cardia and subcardiac region. Surgical Laparoscopy Endoscopy Percutaneous Technology 28(1):e1–e7.
Zhao, L., J. Chen, F. Chen, W. Wang, C. T. Lu, and N. Ramakrishnan. 2015. SimNest: Social media nested epidemic simulation via online semi-supervised learning. In Proceedings of the IEEE International Conference on Data Mining. Pp. 639–648. doi:10.1109/ICDM.2015.39.
Zick, R. G., and J. Olsen. 2001. Voice recognition software versus a traditional transcription service for physician charting in the ED. American Journal of Emergency Medicine 19(4):295–298.
Zieger, A. 2018. Will payers use AI to do prior authorization? And will these AIs make things better? Healthcare IT Today. https://www.healthcareittoday.com/2018/12/27/will-payers-use-ai-to-do-prior-authorization-and-will-these-ais-make-things-better (accessed November 12, 2019).
Zitnik, M., M. Agrawal, and J. Leskovec. 2018. Modeling polypharmacy side effects with graph convolutional networks. Bioinformatics 34(13):i457–i466.
Suggested citation for Chapter 3: Roski, J., W. Chapman, J. Heffner, R. Trivedi, G. Del Fiol, R. Kukafka, P. Bleicher, H. Estiri, J. Klann, and J. Pierce. 2020. How artificial intelligence is changing health and health care. In Artificial intelligence in health care: The hope, the hype, the promise, the peril. Washington, DC: National Academy of Medicine.