I Introduction
Artificial intelligence (AI), particularly its subcategory, machine learning, is changing our daily lives and the way we receive health care. The digital health apps market is booming, with over 350,000 health apps available to patients and consumers, ranging from wellness and fitness apps to disease management apps.Footnote 1 In particular, many direct-to-consumer medical AI apps for “self-diagnosis” (DTC medical self-diagnosing AI apps) are emerging that help individuals to identify a disease or other condition based on entering, for example, symptoms.Footnote 2 DTC medical self-diagnosing AI apps offer new opportunities, but they also raise issues. While the current legal debate has mainly focused on the poor accuracy of DTC medical self-diagnosing apps,Footnote 3 this chapter will discuss the labeling challenges associated with these apps that have received little attention in the literature.
This chapter will first explore the current and future landscape of DTC medical self-diagnosing AI apps. It will then focus on their regulation and discuss whether DTC medical self-diagnosing AI apps are medical devices under section 201(h)(1) of the Federal Food, Drug, and Cosmetic Act (FDCA). This will be followed by a discussion of two labeling challenges raised by DTC medical self-diagnosing AI apps: First, the concern of labeling DTC medical self-diagnosing AI apps as what I call “information-only” tools, and second, particular issues associated with the use of AI, ranging from bias to adaptive algorithms.
This chapter concludes that the labeling of DTC medical self-diagnosing AI apps as “information-only” rather than “diagnostic” tools is unknown to most consumers. The Food and Drug Administration (FDA) should create user-friendly labeling standards for AI-based medical devices, including those that are DTC. For example, these standards should ensure that consumers are adequately informed about the indications for use, model characteristics, and the risks and limitations of the respective DTC medical self-diagnosing AI apps. Based on a risk-based approach, some of these apps should also be prescribed by physicians rather than being offered directly to consumers over the counter. Physicians can help direct the use of the app in question and point out material facts, such as the risk of false positives and negatives, in the patient–physician conversation. In the long run, it may also be helpful to set up a new federal entity responsible for (at least for the coordination of) all issues raised by mobile health apps, from regulation to privacy to reimbursement. While this chapter focuses on FDA regulation for DTC medical self-diagnosing AI apps, some of the suggested solutions here may also have implications for other DTC apps.
II The Current and Future Landscape of DTC Medical Self-Diagnosing AI Apps
The US mobile health market is expected to grow continuously over the next decade, with medical apps (compared to fitness apps) representing the bulk of the market.Footnote 4 Before, or instead of, visiting a doctor’s office, consumers are trying more than ever before to self-diagnose their conditions by putting keywords of their symptoms into search engines like Google or using DTC medical self-diagnosing AI apps.Footnote 5 Approximately 80 percent of patients use the Internet for health-related searches.Footnote 6 According to a 2017 US survey, only 4 percent (ages 61 and older) to 10 percent (ages 18 to 29) of adults used apps for self-diagnosis, but 32 percent (ages 18 to 29) to 62 percent (ages 61 and older) of adults said that they could imagine using them.Footnote 7 Since the COVID-19 pandemic, digital health technologies have gained popularity to mitigate the spread of the virus,Footnote 8 and the use of medical self-diagnosing apps, including those based on AI, has become a reality for more adults in the USA.Footnote 9
In 2021, Google announced the planned launch of a pilot study of its “AI-powered dermatology tool” to help consumers find answers to their skin, nail, and hair condition questions.Footnote 10 With their phone’s camera, consumers simply need to take three photos of their skin, nail, or hair concerns from different perspectives and answer a few questions, such as their skin type and other symptoms.Footnote 11 The app will then offer a list of possible conditions.Footnote 12 Google’s app, dubbed DermAssist, is currently CE-marked as a low-risk (so-called class I) medical device in the European Union (EU) but is being further tested via a limited market release.Footnote 13 The CE marking indicates that the device conforms with the applicable legal requirements.Footnote 14 DermAssist is not yet available in the USA and has not undergone an FDA review for safety and effectiveness.Footnote 15
But Google is not the only company that is investing in dermatology apps. Indeed, a quick search in mobile app stores like Apple and Google Play reveals that there are already similar apps available to download for US consumers, such as AI Dermatologist: Skin Scanner by the developer Acina. Once consumers download this AI app, they can check their skin by taking a photo of, for example, their mole with their phone’s camera.Footnote 16 Within one minute, consumers will receive a risk assessment from AI Dermatologist, including some advice concerning the next steps.Footnote 17 It appears that AI Dermatologist is CE-marked as a medical device in the EU but has not undergone premarket review by the FDA.Footnote 18
There are also other DTC medical self-diagnosing AI apps already available on the US market. A classic example is Apple’s electrocardiogram (ECG) and irregular rhythms notification feature apps.Footnote 19 Both apps are moderate-risk (so-called class II) medical devices that received marketing authorization from the FDA in September 2018.Footnote 20 They are used with the Apple Watch and are addressed directly to consumers. While Apple’s ECG app is intended to store, create, transfer, record, and display a single channel ECG,Footnote 21 Apple’s irregular rhythms notification feature app detects irregular heart rhythm episodes suggestive of atrial fibrillation.Footnote 22 Another example is the AI symptom checker Ada. Consumers can manage their health by answering Ada’s health questions about their symptoms, such as headaches and stomach problems.Footnote 23 Ada’s AI will then use its medical dictionary of medical conditions and disorders to deliver possible causes for the symptoms and offer advice.Footnote 24 Ada’s consumer app is currently CE-marked as a class I medical device in the European Economic Area,Footnote 25 but, similar to AI Dermatologist, it does not appear that the app has undergone a premarket review by the FDA.Footnote 26
III DTC Medical Self-Diagnosing AI Apps as Medical Devices
Can the FDA regulate DTC medical self-diagnosing AI apps? The answer is yes, if they are classified as medical devices under FDCA section 201(h)(1). This section will discuss the definition of a medical device, the FDA’s enforcement discretion, and a relevant exception to the medical device definition.
A The Medical Device Definition and the FDA’s Enforcement Discretion
Under FDCA section 201(h)(1), a “device” is
an instrument, apparatus, implement, machine, contrivance, implant, in vitro reagent, or other similar or related article, including any component, part, or accessory, which is … intended for use in the diagnosis of disease or other conditions, or in the cure, mitigation, treatment, or prevention of disease, in man …, and which does not achieve its primary intended purposes through chemical action within or on the body of man … and which is not dependent upon being metabolized for the achievement of its primary intended purposes….Footnote 27
From the outset, the FDA can only regulate software functions that are classified as medical devices under FDCA section 201(h)(1) (so-called “device software functions”).Footnote 28 In other words, the FDA has no statutory authority to regulate software functions that are not considered medical devices under FDCA section 201(h)(1).Footnote 29 There are different types of software classifications. A relevant one is “Software as a Medical Device” (SaMD), which is standalone software and, as such, counts as a medical device.Footnote 30 The International Medical Device Regulators Forum defines SaMD as “software intended to be used for one or more medical purposes that perform these purposes without being part of a hardware medical device.”Footnote 31 For example, Apple’s ECG and irregular rhythms notification feature apps are both SaMD because they are software-only apps intended for a medical purpose.Footnote 32
Only recently, in September 2022, the FDA updated its Guidance for Device Software Functions and Mobile Medical Applications (Mobile Medical App Guidance) to reflect recent changes, such as the issuance of the FDA’s final Guidance on Clinical Decision Support Software.Footnote 33 Although the Mobile Medical App Guidance contains nonbinding recommendations, it represents the FDA’s current thinking on its regulatory approach to device software functions, including mobile medical apps.Footnote 34 The FDA defines “mobile medical apps” as mobile apps that incorporate device software functionalities that meet the medical device definition in the FDCA, and either are “intended … to be used as an accessory to a regulated medical device; or … to transform a mobile platform into a regulated medical device.”Footnote 35
The “intended use” is relevant for determining whether a mobile app is considered a medical device.Footnote 36 The term means “the objective intent of the persons legally responsible for the labeling of devices.”Footnote 37 Such persons are usually the manufacturers whose expressions determine the intent.Footnote 38 The intent can also be shown by the circumstances surrounding the product’s distribution.Footnote 39 For instance, the objective intent can be derived from advertising materials, labeling claims, and written or oral statements by the product’s manufacturer or its representatives.Footnote 40
In its Mobile Medical App Guidance, the FDA clarifies that it intends to focus its regulatory oversight on only those device software functions whose functionality could present a risk to the safety of patients if they were not to function as intended.Footnote 41 This means that the FDA intends to exercise enforcement discretion over those software functions that are or may be medical devices under FDCA section 201(h)(1) but present a lower risk to the public.Footnote 42 Enforcement discretion means that the agency does not aim to enforce requirements under the FDCA.Footnote 43
For example, the FDA intends to apply its regulatory oversight to device software functions that analyze images of skin lesions using mathematical algorithms and provide users with risk assessments of the lesions.Footnote 44 In contrast, for instance, the FDA considers apps exclusively intended for patient education, such as an app that helps guide patients to ask the right questions to their physician concerning their disease, as not being medical devices, and, thus, those apps fall outside of the FDA’s statutory authority.Footnote 45 An example of a mobile app that may meet the medical device definition, but for which the FDA intends to exercise enforcement discretion because it poses a lower risk to the public, is an app that provides a “Skill of the Day” behavioral technique that patients with diagnosed psychiatric conditions can access when experiencing increased anxiety.Footnote 46
When applying the FDA’s current thinking in the Mobile Medical App Guidance to DTC medical self-diagnosing AI apps, some of these apps are considered device software functions that are the focus of the agency’s regulatory oversight. Take as an example Apple’s ECG and irregular rhythms notification feature apps. Both apps are considered class II (moderate-risk) medical devices and had to undergo a premarket review by the FDA via the so-called De Novo process before being placed on the US market.Footnote 47
However, even if DTC medical self-diagnosing AI apps are considered medical devices because they help individuals identify a disease or other condition and are considered to be “intended for use in the diagnosis of disease or other conditions, or in the cure, mitigation, treatment, or prevention of disease,”Footnote 48 the FDA may exercise enforcement discretion over some of them if they are considered to pose a low risk to the public. For example, as mentioned previously, the consumer app Ada is currently CE-marked as a class I (low-risk) medical device in the European Economic Area.Footnote 49 However, it seems that Ada has not undergone a premarket review by the FDA.Footnote 50 One option why this is likely the case is that Ada (may) meet(s) the medical device definition in FDCA section 201(h)(1),Footnote 51 but falls within the FDA’s enforcement discretion because it is considered to pose a lower risk to the public. This analysis also seems to be consistent with the Mobile Medical App Guidance. In Appendix B of its Guidance, the FDA lists examples of software functions that may meet the medical device definition but for which the agency exercises enforcement discretion, including:
– “Software functions that use a checklist of common signs and symptoms to provide a list of possible medical conditions and advice on when to consult a health care provider” and
– “Software functions that guide a user through a questionnaire of signs and symptoms to provide a recommendation for the type of health care facility most appropriate to their needs.”Footnote 52
In addition, most class I medical devices under the FDCA are also a priori exempt from premarket notification (510(k)) requirements.Footnote 53
B The Medical Device Exception, FDCA Section 520(o)(1)(B)
Section 3060(a) of the 21st Century Cures Act introduced five exceptions to the medical device definition for certain software functions. One of these exceptions is particularly relevant for DTC AI apps – namely FDCA section 520(o)(1)(B), which states that “the term device, as defined in section 201(h), shall not include a software function that is intended … for maintaining or encouraging a healthy lifestyle and is unrelated to the diagnosis, cure, mitigation, prevention, or treatment of a disease or condition; … .”
In 2019, the FDA issued nonbinding Guidance on Changes to Existing Medical Software Policies Resulting from Section 3060 of the 21st Century Cures Act (Cures Act Guidance), in which the agency, among other things, expresses its current interpretation of FDCA section 520(o)(1)(B).Footnote 54 In particular, the FDA clarifies that FDCA section 520(o)(1)(B) means software functions that belong to the first category of general wellness intended uses, as defined in the FDA’s nonbinding Guidance on General Wellness: Policy for Low Risk Devices (General Wellness Guidance),Footnote 55 and that are “unrelated to the diagnosis, cure, mitigation, prevention, or treatment of a disease or condition.”Footnote 56 Software functions that fall within the first category of general wellness intended uses are intended for “maintaining or encouraging a general state of health or a healthy activity.”Footnote 57 For example, an app that assists users with weight loss goals and does not make any reference to diseases or conditions falls under FDCA section 520(o)(1)(B), and, thus, is not considered a medical device under FDCA section 201(h)(1).Footnote 58
In its Cures Act Guidance, the FDA also clarifies that software functions that fall within the second category of general wellness intended uses, as defined in the General Wellness Guidance, are not covered by FDCA section 520(o)(1)(B).Footnote 59 Software functions that fall within the second category of general wellness intended uses have “an intended use that relates the role of healthy lifestyle with helping to reduce the risk or impact of certain chronic diseases or conditions and where it is well understood and accepted that healthy lifestyle choices may play an important role in health outcomes for the disease or condition.”Footnote 60
In contrast to the first category of general wellness intended uses, this second category relates to the prevention or mitigation of a disease or condition, and, thus, software functions that fall within this second category are not excluded from the medical device definition.Footnote 61 For example, if the app in the previous example makes reference to diseases or conditions – for instance, if it claims that maintaining a healthy weight will aid living well with type 2 diabetes – this app falls outside of the scope of FDCA section 520(o)(1)(B).Footnote 62
As understood here, DTC medical self-diagnosing AI apps help users to identify a disease or other condition based on entering, for example, symptoms. They are related “to the diagnosis, cure, mitigation, prevention, or treatment of a disease or condition” and, thus, do not fall under the medical device exception in FDCA section 520(o)(1)(B).Footnote 63 To sum up, DTC medical self-diagnosing AI apps are medical devices under FDCA Section 201(h)(1) that are either the focus of the FDA’s regulatory oversight or for which the agency exercises its enforcement discretion. Figure 10.1 summarizes the regulation of mobile health apps, including DTC medical self-diagnosing AI apps.
IV Labeling Challenges for DTC Medical Self-Diagnosing AI Apps
As established above, DTC medical self-diagnosing AI apps, as understood here, are medical devices that are either the focus of the FDA’s regulatory oversight or for which the agency exercises its enforcement discretion. This section will focus on the labeling challenges for DTC medical self-diagnosing AI apps. It will first give an overview of medical device labeling and the relevant terminology. It will then focus on labeling challenges for DTC medical self-diagnosing AI apps and make suggestions on how to overcome them.
A Labeling
Device software functions are organized into one of three classes based on their risk level, ranging from class I (lowest risk) to class III (highest risk).Footnote 64 Depending on the device classification, manufacturers must follow the associated controls – that is, General Controls, Special Controls, and/or Premarket Approval.Footnote 65 In principle, General Controls apply to all device software functions.Footnote 66 For instance, the General Device Labeling Requirements in Part 801 of Title 21 of the Code of Federal Regulations (CFR) are General Controls.Footnote 67 21 CFR Part 801 includes, among other things, general labeling provisions, such as the name and place of business, adequate directions for use, and the use of symbols, as well as special requirements for specific devices, such as hearing aids, and labeling requirements for unique device identification and over-the-counter devices.Footnote 68
Labeling is defined in FDCA section 201(m) as “all labels and other written, printed, or graphic matter (1) upon any article or any of its containers or wrappers, or (2) accompanying such article.” It is a generic term that also includes all labels.Footnote 69 Under FDCA section 201(k), the term “label” means
a display of written, printed, or graphic matter upon the immediate container of any article; and a requirement made by or under authority of this Act that any word, statement, or other information appear on the label shall not be considered to be complied with unless such word, statement, or other information also appears on the outside container or wrapper, if any there be, of the retail package of such article, or is easily legible through the outside container or wrapper.
In the context of DTC medical self-diagnosing AI apps, the label will usually be available in non-physical form through the app itself.
It is also worth noting that if the “labeling is false or misleading in any particular,” the device is considered misbranded under FDCA section 502(a)(1). The term “misleading” means that the labeling proves deceptive to device users and creates or leads to a false impression in their minds.Footnote 70 For example, this can be the case if the label contains exaggerated claims or if it fails to inform users about relevant facts.Footnote 71
B Challenges
DTC medical self-diagnosing AI apps raise labeling challenges. This section will discuss two: First, the concern of labeling DTC medical self-diagnosing AI apps as what I call “information-only” tools, and second, particular issues associated with the use of AI, ranging from bias to adaptive algorithms. It will also make suggestions on how to address these challenges. While the following remarks focus on medical devices, they may also have implications for those DTC apps that fall outside the FDA’s statutory authority.
i Labeling as “Information-Only” Tools
Apple’s ECG and irregular rhythms notification feature apps used with the Apple Watch are both over-the-counter class II medical devices that received marketing authorization from the FDA in September 2018.Footnote 72 As previously mentioned, Apple’s ECG app is intended to store, create, transfer, record, and display a single channel ECG.Footnote 73 The indications for use, however, also include, among other things, the following sentences:
The user is not intended to interpret or take clinical action based on the device output without consultation of a qualified healthcare professional. The ECG waveform is meant to supplement rhythm classification for the purposes of discriminating AFib [atrial fibrillation] from normal sinus rhythm and not intended to replace traditional methods of diagnosis or treatment.Footnote 74
The FDA created a new device type, namely “electrocardiograph software for over-the-counter use,” regulated in 21 CFR 870.2345, for Apple’s ECG app and substantially equivalent devices.Footnote 75 Interestingly, 21 CFR 870.2345(a) also states that “this device is not intended to provide a diagnosis.”
Moreover, as mentioned, Apple’s irregular rhythms notification feature app detects irregular heart rhythm episodes suggestive of atrial fibrillation.Footnote 76 But, much like Apple’s ECG app, this app’s indications for use include, inter alia, the following phrases:
It is not intended to provide a notification on every episode of irregular rhythm suggestive of AFib and the absence of a notification is not intended to indicate no disease process is present; rather the feature is intended to opportunistically surface a notification of possible AFib when sufficient data are available for analysis. These data are only captured when the user is still. Along with the user’s risk factors, the feature can be used to supplement the decision for AFib screening. The feature is not intended to replace traditional methods of diagnosis or treatment.Footnote 77
The FDA also created a new device type, namely “photoplethysmograph analysis software for over-the-counter use,” laid down in 21 CFR 870.2790, for Apple’s irregular rhythms notification feature app and substantially equivalent devices.Footnote 78 Similar to 21 CFR 870.2345, this regulation also clarifies that “this device is not intended to provide a diagnosis.”Footnote 79
But Apple’s apps are not the only DTC medical self-diagnosing AI apps that articulate that their device “is not intended to provide a diagnosis.” For example, Google’s 2021 announcement of its AI-powered dermatology tool says:Footnote 80 “The tool is not intended to provide a diagnosis nor be a substitute for medical advice as many conditions require clinician review, in-person examination, or additional testing like a biopsy. Rather we hope it gives you access to authoritative information so you can make a more informed decision about your next step.”Footnote 81
In addition, Google’s website states: “DermAssist is intended for informational purposes only and does not provide a medical diagnosis.”Footnote 82 The same is also true for the AI Dermatologist: Skin Scanner app.Footnote 83 When looking up information about the app in an app store, the preview states: “It is essential to understand that an AI-Dermatologist is not a diagnostic tool and cannot replace or substitute a visit to your doctor.”Footnote 84 App store previews of Ada say something similar: “CAUTION: The Ada app cannot give you a medical diagnosis…. The Ada app does not replace your healthcare professional’s advice or an appointment with your doctor.”Footnote 85
Consequently, DTC medical self-diagnosing AI apps are labeled as “information-only” rather than “diagnostic” tools.Footnote 86 Irrespective of whether DTC medical self-diagnosing AI apps are medical devices that are the focus of the FDA’s regulatory oversight or for which the agency exercises its enforcement discretion, these apps seem to have in common that their manufacturers claim they are “not intended to provide a diagnosis.” This is likely due to their over-the-counter nature, although Apple’s clinical study of the ECG app, for example, showed that the app correctly diagnosed atrial fibrillation with 98.3 percent sensitivity and 99.6 percent specificity.Footnote 87 As a comparison, a prescription device is a “device which, because of any potentiality for harmful effect, or the method of its use, or the collateral measures necessary to its use is not safe except under the supervision of a practitioner licensed by law to direct the use of such device.”Footnote 88 But do patients and consumers really understand that Apple’s ECG app and similar apps are not intended to replace traditional diagnosis and treatment methods, let alone that some have been FDA reviewed and others have not?
There appears to be a significant discrepancy between the user’s perception of the intended use of DTC medical self-diagnosing AI apps and their actual intended use (i.e., not to diagnose). Indeed, a recent study on AI-assisted symptom checkers revealed that 84.1 percent of respondents perceive them as diagnostic tools.Footnote 89 In addition, 33.2 percent of respondents use symptom checkers for deciding whether to seek care, and 15.8 percent of respondents said they use them to receive medical advice without seeing a doctor.Footnote 90 However, as seen above, apps like the ones from Apple and other companies have clear indications for use, and, thus, are likely not considered deceptive to device users and, thus, not “misleading” under FDCA section 502(a)(1).Footnote 91 Nevertheless, even if one cannot establish misleading labeling under the FDCA, there is this misperception among users that these apps are diagnostic tools.
This misperception can also be due, among other things, to the fact that many users may not read the labels. Labeling has many benefits, including helping patients and consumers to make more informed decisions, such as by informing them about the potential limitations of an app. But if users do not read the labels and accompanying statements and language like “this device is not intended to provide a diagnosis” is buried somewhere within them, using DTC medical self-diagnosing AI apps can become risky and jeopardize patient health. For example, imagine a patient uses an app like AI Dermatologist and screens herself for skin cancer. What if the AI misses a melanoma, and the patient does not see a doctor because she perceives the app as a diagnostic tool and believes everything is alright?
Regulators and stakeholders, such as app developers, need to better educate users of DTC medical self-diagnosing AI apps, for example, about the indications for use, whether the app has undergone FDA review, and its risks. With the right design, labels could help to achieve these goals. Several groups have already shown the benefits of “eye-popping” label designs, such as with the help of “nutrition” or “model facts” labels.Footnote 92 In particular for apps, there is a multitude of possible design options (e.g., pop-up notifications in plain language) to make users more effectively aware of important information.Footnote 93 Thus, regulators like the FDA could – with the help of stakeholders and label designers – develop user-friendly label design options for DTC medical self-diagnosing AI apps.Footnote 94 Once created, additional educational campaigns could be used to promote the proper reading of the label.Footnote 95 Human factors testing would also be helpful, particularly to see whether users understand when to seek medical help.Footnote 96
In addition, as part of its regulatory review, the FDA should consider whether some of these apps should be prescribed by doctors rather than being offered directly to consumers over the counter.Footnote 97 The advantage could be that physicians could assist patients with the use of the app in question and point out material facts in the patient–physician conversation. A risk-based approach may likely be useful here to determine such “prescription apps.”
Moreover, there is a general question of whether the FDA’s current approach to practice enforcement discretion over many DTC medical self-diagnosing AI apps is convincing. Other countries have come up with different regulatory designs to better protect consumers. For example, Germany incentivizes manufacturers of even low-risk apps (i.e., class I devices) to produce high-quality apps that comply with specific standards (e.g., safety, privacy, etc.) by offering insurance coverage for their apps in return.Footnote 98 While the FDA does not currently seem to have the resources to execute a similar approach and review all DTC medical self-diagnosing AI apps, the flood of mobile health apps and all the associated issues,Footnote 99 ranging from poorly designed products to inadequate data protection, to labeling issues and misperceptions concerning their use, requires a new regulatory approach in the long run. A better option might be to create a new federal entity in the future that would be responsible for (at least the coordination of) all issues raised by mobile health apps, including DTC medical self-diagnosing AI apps, from regulation over privacy to enforcement actions and reimbursement.
ii Particular Issues of AI: From Bias to Adaptive Algorithms
Another labeling challenge that DTC medical self-diagnosing AI apps raise is that they are not only directly addressed to consumers without a licensed practitioner’s supervision, but that they also operate using AI. Indeed, AI-based medical devices, including DTC medical self-diagnosing AI apps, are very different from traditional medical devices, such as simple tongue depressors.Footnote 100
First, DTC medical self-diagnosing AI apps may use methods like deep learning that make them opaque (often dubbed “black boxes”).Footnote 101 This means that the end users of the DTC medical self-diagnosing AI app (and likely even the app developers) cannot understand how the AI reached its recommendations and/or decisions. Second, DTC medical self-diagnosing AI apps may be biased. AI tools are prone to different types of bias, ranging from biased data fed to them (e.g., a skin cancer screening app that is largely trained on white skin images) to label choice biases (e.g., the algorithm uses an ineffective proxy for ground truth).Footnote 102 Third, DTC medical self-diagnosing AI apps may continuously learn from new data (e.g., health information, images, etc.) supplied by consumers using such apps (so-called “adaptive algorithms”).Footnote 103 These apps are, thus, much more unpredictable in terms of their reliability and would preferably need constant monitoring to avoid introducing new biases, for example.Footnote 104 Lastly, the human–AI interaction is complex. In particular, DTC medical self-diagnosing AI apps that have unique characteristics as their outputs are often probabilistic and, thus, require consumers to incorporate the information received into their own beliefs.Footnote 105 In addition, DTC medical self-diagnosing AI apps are usually available for little money or even for free.Footnote 106 They can easily be used as often as consumers wish.Footnote 107 For example, consumers of a skin scanner app may decide to scan their moles many times (rather than just once), which increases the chance of false-positive results – that is, the app detects a potential disease that is not actually present.Footnote 108 Because consumers are typically risk-averse about their health outcomes, they may seek medical help when it is not needed, further overburdening the health care system and taking away limited resources from patients who are more likely to need them.Footnote 109
Despite the differences between AI-based medical devices, such as DTC medical self-diagnosing AI apps, and traditional medical devices, such as simple tongue depressors, there are currently no labeling requirements for medical devices specifically aimed at AI (see Title 21 of the CFR).Footnote 110 The FDA has not yet developed any labeling standards for AI-based medical devices, let alone those directly addressed to consumers.Footnote 111 Thus, when creating the optimal design labels for DTC medical self-diagnosing AI apps,Footnote 112 the FDA should also focus on the content and use this opportunity to develop labeling standards for AI-based medical devices, including those that are DTC.Footnote 113
It is crucial that consumers know and understand, among other things, the indications for use, model characteristics, and the risks and limitations of AI-based medical devices.Footnote 114 For example, users of DTC medical self-diagnosing AI apps should be made aware of the type of AI used (e.g., a black box, an adaptive algorithm, etc.) and the risks associated with using the app in question. They should also be informed about the various risks of bias and warned against blindly relying on the app’s outputs. Moreover, consumers should be alerted to the fact that increased testing can lead to an increased chance of false positives and generally be educated about the risks of false-positive and false-negative results, including when to see a doctor. A discussion with stakeholders needs to occur as soon as possible on the content of the labels of AI-based medical devices, including DTC medical self-diagnosing AI apps.Footnote 115 In particular, the language used for the labeling of these devices will need to be plain when they are DTC.Footnote 116
V Conclusion
The digital health apps market is booming, and DTC medical self-diagnosing AI apps are emerging that help users to identify a disease or other condition based on entering, for instance, symptoms. Examples of such apps include Apple’s ECG and irregular rhythm notification feature apps, Google’s AI-powered dermatology tool, the AI Dermatologist: Skin Scanner app, and the symptom checker Ada. DTC medical self-diagnosing AI apps raise a multitude of challenges, including questions of labeling. What should labels directly addressed to consumers look like? What information should be included in such a label?
This chapter has argued that the FDA should develop user-friendly labeling standards for AI-based medical devices, including DTC medical self-diagnosing AI apps. For example, consumers need to be effectively informed about the type of AI used (e.g., a black box, an adaptive algorithm, etc.), the various risks of bias, the risks of false-positive and negative results, and when to seek medical help. In particular, the design of such labels needs to promote their reading so that users are made aware that the DTC medical self-diagnosing AI app in question is an “information-only” tool and is “not intended to provide a diagnosis.” Additionally, some of these apps should be prescribed by a doctor, not offered over the counter, based on a risk-based approach so that the doctor can point out key facts. In the long run, it may also be helpful to create a new federal entity responsible for (at least the coordination of) all issues raised by mobile health apps, ranging from regulation to privacy to reimbursement.