Skip to main content

Thank you for visiting You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

Evaluation of an artificial intelligence-based medical device for diagnosis of autism spectrum disorder


Autism spectrum disorder (ASD) can be reliably diagnosed at 18 months, yet significant diagnostic delays persist in the United States. This double-blinded, multi-site, prospective, active comparator cohort study tested the accuracy of an artificial intelligence-based Software as a Medical Device designed to aid primary care healthcare providers (HCPs) in diagnosing ASD. The Device combines behavioral features from three distinct inputs (a caregiver questionnaire, analysis of two short home videos, and an HCP questionnaire) in a gradient boosted decision tree machine learning algorithm to produce either an ASD positive, ASD negative, or indeterminate output. This study compared Device outputs to diagnostic agreement by two or more independent specialists in a cohort of 18–72-month-olds with developmental delay concerns (425 study completers, 36% female, 29% ASD prevalence). Device output PPV for all study completers was 80.8% (95% confidence intervals (CI), 70.3%–88.8%) and NPV was 98.3% (90.6%–100%). For the 31.8% of participants who received a determinate output (ASD positive or negative) Device sensitivity was 98.4% (91.6%–100%) and specificity was 78.9% (67.6%–87.7%). The Device’s indeterminate output acts as a risk control measure when inputs are insufficiently granular to make a determinate recommendation with confidence. If this risk control measure were removed, the sensitivity for all study completers would fall to 51.6% (63/122) (95% CI 42.4%, 60.8%), and specificity would fall to 18.5% (56/303) (95% CI 14.3%, 23.3%). Among participants for whom the Device abstained from providing a result, specialists identified that 91% had one or more complex neurodevelopmental disorders. No significant differences in Device performance were found across participants’ sex, race/ethnicity, income, or education level. For nearly a third of this primary care sample, the Device enabled timely diagnostic evaluation with a high degree of accuracy. The Device shows promise to significantly increase the number of children able to be diagnosed with ASD in a primary care setting, potentially facilitating earlier intervention and more efficient use of specialist resources.


Autism spectrum disorder (ASD) is one of the most common developmental disorders, with a US prevalence of 1.9%1. The importance of timely ASD diagnosis, possible as early as 18 months, is underscored by studies linking earlier intervention during the critical neurodevelopmental window to enhanced long-term outcomes. These include greater improvements in social and communication skills2,3, cognitive abilities4,5, verbal abilities3,6, and adaptive behavior7,8. Despite documented benefits of early intervention, the mean age of ASD diagnosis in the US remains high at over 4 years1,9,10,11. The estimated three-year delay between initial caregiver concern and an ASD diagnosis is even longer for children who are non-white, female, of lower socioeconomic status, or rural residing12,13,14,15. Roughly 27% of children with ASD remain undiagnosed at age 815.

One factor contributing to diagnostic delay is the rapid increase in demand for ASD evaluations that has outpaced specialist capacity and led to prolonged wait times16,17,18,19. Gender, race, and socioeconomic biases in access and diagnosis have also created additional delays for certain subpopulations1,14. ASD diagnostic practices in the US are currently fragmented and heavily reliant on a limited number of pediatric subspecialists and team-based behavioral evaluations20. These assessments are time-intensive, and families may wait as long as 18 months between initial screening by their healthcare provider (HCP) in the primary care setting and final diagnosis by the specialist16.

Low levels of ASD diagnosis in primary care settings present another barrier to timely ASD diagnosis and initiation of interventions. Currently, only ~1% of patients with ASD are diagnosed by primary care HCPs in the US21,22. The American Academy of Pediatrics (AAP) recommends that upon a failed ASD screen in primary care, providers comfortable with the Diagnostic and Statistical Manual of Mental Disorders, 5th edition (DSM-5) criteria diagnose ASD or refer to a specialist for further evaluation18. Yet, ~60% of children who fail screenings are neither referred to a specialist nor diagnosed in the primary care setting22. Common barriers to primary care diagnosis include low confidence in using ASD diagnostic tools due to lack of specialist training and/or lack of time to administer, lack of perceived self-efficacy in making the diagnosis, and lack of time to properly review results with caregivers and discuss treatment recommendations23,24,25. In addition, common screening tools used in primary care settings may miss many cases of ASD. For example, in a cohort of over 20,000 children with ASD outcome data, 61% (278/454) of children who received an outcome diagnosis of ASD screened negative on a common screener26.

The complexity of making an accurate ASD diagnostic determination may also add to primary care HCP reluctance to diagnose. ASD has varied clinical presentations and heterogeneous etiology20. Additionally, ASD can co-occur with and/or share many overlapping features of other disorders diagnosed in childhood such as attention deficit hyperactivity disorder (ADHD), intellectual disability, speech and language delay, or a variety of psychiatric conditions27. Providers are also tasked with making an assessment within a time-constrained clinical encounter where the child may not display behavior characteristic of that seen in the home environment or may become newly behaviorally reactive due to change in environment28. These complexities underscore the importance of tailored diagnostic aids to support primary care HCPs in making timely and accurate diagnoses when the diagnosis is straightforward, and to advise further evaluation when the presentation is complex or unclear. However, prior to 2 June 2021, no diagnostic devices had market authorization from the Food and Drug Administration (FDA) to aid in the diagnosis of ASD in the primary care setting29.

Existing ASD diagnostic tools20,30,31 are used almost exclusively in specialty care settings in the US. Several of these tools, especially when used in combination, have shown good diagnostic accuracy20 and consistent performance across trained examiners32,33. However, specialized training requirements, the time needed to administer, and insufficient reimbursement rates to justify primary care HCP effort34 make them challenging for use in primary care settings. Overdiagnosis can also occur in populations with a lower prevalence of ASD, as is the case in primary care35. Reliability may also be reduced if an ASD diagnostic tool intended for combination use is instead used as a stand-alone instrument by time-pressured clinicians35. In addition, many of these tools were not designed for remote administration. Geographic and logistical hurdles to finding a trained healthcare professional and appropriate clinical setting may contribute to an imbalance in coverage for some populations15.

In order to increase primary care HCP's capacity to promptly diagnose ASD and/or refer complex cases for specialist review, innovative diagnostic aids suitable for use in the primary care setting are urgently needed. Recent research has highlighted the potential for artificial intelligence-based tools to augment various aspects of ASD care36. The objective of this study was to test the accuracy of one such tool, an artificial intelligence-based Device that produces recommendations for the HCP after analyzing behavioral features from three distinct inputs: a caregiver questionnaire, an analysis of two short home videos, and an HCP questionnaire.

The Device is a Software as a Medical Device (SaMD)37 that deploys a gradient boosted decision tree algorithm. The algorithm uses behavioral features selected through machine-learning techniques as maximally predictive of ASD across a variety of phenotypic presentations38,39,40,41,42,43. The device’s underlying machine learning algorithm was initially developed using patient record data from thousands of children with diverse conditions, presentations, and comorbidities who were either diagnosed with ASD or confirmed not to have ASD based on standardized diagnostic tools and representing both genders across the supported age range. The algorithm was iteratively improved, supplemented with ASD-expert input, and prospectively validated for 7 years prior to this study38,39,40,41,42,43. Use of multimodular Device inputs is consistent with current guidelines and recommendations for an evaluation of ASD that include having both caregiver and clinician input, as well as a structured observation of the child44. Previously published analysis of earlier algorithm iterations demonstrated that combining inputs significantly improved the performance of the algorithm38.

The Device reports that a subject is positive for ASD, negative for ASD, or “indeterminate” (Fig 1). An indeterminate output is given when Device inputs are insufficiently granular for the algorithm to render a highly predictive output. For example, a patient may exhibit an insufficient number and/or severity of features to be confidently classified by the algorithm as being either ASD negative or ASD positive. The Device’s indeterminate output also referred to in the literature as an “abstention” or “no result” output, is a standard method of risk control in machine-learning algorithms45,46. We worked in conjunction with the FDA to establish minimum thresholds for PPV and NPV, which were the primary endpoints for this pivotal study. Once established, we used these values (PPV greater than 65% and NPV greater than 85%) as the boundaries to govern the model’s range of acceptable outcomes during model hyperparameter tuning. We tuned on training/testing data to evaluate combinations of PPV and NPV at or above the FDA thresholds with variable abstention rates. We used cross-validation to maintain the PPV and NPV minimums and to arrive at the current Device abstention thresholds. Abstention in cases of high clinical uncertainty provides a safeguard against known machine-learning failure modes, fosters clinician trust through model transparency, and helps flag cases where additional human expertise or data may be needed46. In the primary care setting, and specifically in relation to neurodevelopmental disorders and ASD where symptoms appear on a spectrum and may co-occur with multiple other phenotypically overlapping conditions27, ambiguous and complex presentations are expected.

Fig. 1: Graphical representation of the Device and its major components.
figure 1

a Caregiver uses a smartphone to answer a brief questionnaire in approximately 5 min, b Caregiver uploads two short (1 min, 30 s up to 5 min) home videos of their child to be scored by trained video analysts, and c their primary care physician (or other qualified healthcare provider) independently answers a short clinical question set in approximately 10 min. These inputs are securely transmitted to the d trained analysts where video features are extracted in approximately 11 min. e The caregiver, healthcare provider, and video analyst inputs are combined into a mathematical vector for machine-learning analysis and classification. f The Device provides an “ASD positive” or “ASD negative” or “no result (indeterminate)” output. The indeterminate output indicates information was insufficiently granular to make a determination at that time.

In this double-blinded, active comparator cohort study conducted across six states, we evaluated the ability of the Device to aid in the diagnosis of ASD in children aged 18–72-months for whom a caregiver or HCP had a concern for developmental delay, but were not yet diagnosed. We compared the Device output to the clinical reference standard, consisting of a diagnosis made by a specialist clinician, based on DSM-5 criteria and validated by one or more blinded reviewing specialist clinicians.


Subject enrollment

A total of 711 participants were enrolled and 425 completed the study between August 2019 and June 2020. Completers and non-completers were similar in terms of demographics (See Table 1). In March 2020, when a national state of emergency was declared in response to COVID-19, a total of 711 participants had enrolled in the study; 585 participants had completed all Device inputs, and 328 of these participants had also completed the specialist evaluation. COVID-19 control measures led to changes in study visit schedules, missed visits, patient discontinuations, and site closures (9 out of 14 sites). Sites that remained open did so with reduced availability to see participants. We estimate that 100 specialist clinician visits could not be completed due to COVID-19. Following the introduction of COVID-19 measures, an additional 97 specialist evaluations were carried out at the sites that remained open. In total, 425 participants completed both the Device input and specialist evaluation component of the study and were included in the final analysis. Data from participants who did not complete both study components were not included in the final analysis. The estimated drop-out rate without the impact of COVID-19 is 26.2% (Fig. 2).

Fig. 2: Participant flowchart.
figure 2

A total of 711 participants enrolled in the study between August 2019 and June 2020. Of these participants, 126 dropped out prior to completing all Device inputs. A further 160 dropped out after completing all Device inputs but prior to completing the specialist evaluation. In total, 425 participants completed all Device inputs and the specialist evaluation and were counted as study completers.

Baseline characteristics and demographics


The mean age of all study participants was 3.36 years (SD = 1.19) and of completers was 3.33 years (SD = 1.15). The mean age of study completers with ASD was 2.96 years (SD = 1.06) and the mean age of completers for whom the Device rendered an ASD positive result was 2.81 years (SD = 0.94). Additional participant baseline characteristics and demographics are presented in Table 1. Among the study completers, specialist clinicians determined that 61.9% (263/425) had one or more non-ASD developmental or behavioral conditions, 28.7% (122/425) were ASD positive, and 9.4% (40/425) were ASD negative and neurotypical. Table 2 shows all behavioral and developmental concerns for each completer, as listed by caregivers at study commencement, and key comorbidities, as determined by the site specialist as part of their assessment.

Table 1 Baseline characteristics and demographics.
Table 2 Participant’s clinical profiles- caregiver and specialist reported.


The HCPs who completed input 3 (n = 15) were physicians who completed residency training in either general pediatrics or family medicine. They worked across a variety of primary care settings and represented both genders (males 47%). Years in practice post-residency ranged from 1 to 38 with a mean of 16 years in practice. HCPs were not recruited on the basis of any special interest or experience of ASD diagnosis, beyond what was delivered as part of their standard training.

Device output summary

The Device provided a determinate output (ASD positive or negative) for 31.8% (135/425) of participants. In other words, approximately 1 in 3 subjects received a determinate result from the Device. Of the participants who received an ASD diagnosis by a clinical specialist, 52.5% (64/122) received a determinate result and all were correctly classified by the Device with the exception of a single false negative. Among participants given an ASD negative and neurotypical diagnosis by a clinical specialist, 35.0% (14/40) received an ASD negative Device result and none were misclassified as ASD positive. If the indeterminate safety mechanism were removed, the sensitivity for all study completers would fall to 51.6% (63/122) (95% CI 42.4%, 60.8%) and specificity would fall to 18.5% (56/303) (95% CI 14.3%, 23.3%), due in part to the high number of participants with other neurodevelopmental conditions. Table 3 compares Device output to clinical reference standard diagnosis for all study completers.

Table 3 Confusion matrix showing Device output compared to clinical reference standard diagnosis.

Determinate category analysis

For participants for whom the Device rendered a determinate output, 18.4% (78/425) were classified ASD positive and 13.4% (57/425) ASD negative. The Device yielded these ASD positive or negative results with a PPV of 80.8% (95% CI, 70.3%–88.8%), NPV of 98.3% (95% CI, 90.6%–100%), sensitivity of 98.4% (95% CI, 91.6%–100%), and specificity of 78.9% (95% CI, 67.6%–87.7%).

Indeterminate category analysis

Specialist diagnosis found 91.0% (264/290) of this group had at least one neurodevelopmental disorder. Specifically, 71.0% (206/290) were ASD negative and had at least one other non-ASD neurodevelopmental or behavioral condition, 20.0% (58/290) were ASD positive, and the remaining 9.0% (26/290) received no developmental delay diagnosis.

Covariate analysis

Race and ethnicity covariate analysis was based on non-exclusive categories. Therefore, participants who reported multiple races or ethnicities (13.4% of study completers) were included in each race or ethnicity category they identified. While the study was not powered for statistical inference on covariates, we detected no difference in Device performance across participants’ sex, race/ethnicity, income, or education level as determined by examining the overlap of corresponding 95% CIs (see Table 4). The Device provided a higher determinate rate for participants under 3 years old (39%) compared to participants 3 years and over (26%; p = 0.006) and higher specificity for participants 3 years and over (88%) compared to participants under 3 years old (67%; p = 0.03).

Table 4 Device performance metrics by covariate.

Severity scores for both the “social communication” and the “restricted and repetitive behavior” categories for all children who received a positive ASD reference diagnosis were recorded by diagnosing specialists per the DSM-5 criteria (Level 3: “Requiring very substantial support”; Level 2: “Requiring substantial support”; Level 1: “Requiring support”). We break these scores down by Device output (positive, indeterminate, negative) in Table 5.

Table 5 ASD severity level scores—social communication and restricted and repetitive behavior by device output.

Table 6 shows the frequency of key comorbidities across Device outputs.

Table 6 Specialist diagnosed comorbidities broken down by device output.

In-person vs remote HCP questionnaire assessment

Of the study completers, 366 (86.1%) of the HCP assessments (Device input 3) were completed via an in-person visit, while 59 (13.9%) were completed via a remote visit. No evidence of performance degradation was found when assessments were performed remotely.

Time associated with Device use

Completion of input 1 (the caregiver questionnaire) took a median time of 4 min and 56 s. Input 2, completion of video analyst scoring, took a median time of 10 min and 54 s from the time video review began to submission of scores. Input 3 time is self-reported since the HCPs completed a hardcopy of the questionnaire that was only later uploaded to the portal. Qualitatively, HCPs involved in the study reported it took roughly 10 min to complete input 3. Upon completion of inputs, results were immediately produced (Fig 1).

Diagnostic certainty—agreement amongst clinicians

A diagnosing specialist clinician completed the patient assessment. The patient's case was then independently assessed by a reviewing specialist clinician. If these two specialists disagreed about the ASD diagnosis, then the case was referred to a second reviewing specialist. The first reviewing specialist agreed with the diagnosing specialist 79% of the time. The remaining 21% of subjects required a review by a second independent specialist clinician, of which 43% of the time, they agreed with the diagnosing specialist, and 57% of the time, they agreed with the first reviewing specialist. Of the 425 study completers, the diagnosing specialist was either “somewhat” or “completely certain” of the diagnosis 95% of the time. The diagnosing specialist clinician was “completely certain” 67% of the time when diagnosing ASD, 74% of the time when ruling out ASD and suspecting another non-ASD condition, and 95% of the time when ruling out ASD and when the subject was neurotypical (without ASD or other non-ASD condition suspected). In subjects with ASD, when the Device was positive for ASD, the diagnosing specialist clinician was “completely certain” of the diagnosis 71% (95% CI 59–82%) of the time. In subjects with ASD, when the Device abstained from rendering a diagnostic output, the diagnosing specialist clinician was “completely certain” 64% (95% CI 50%–76%) of the time. A test of the significance of the difference between those two proportions returns a p value of 0.40.


The Device evaluated in this study was designed to aid HCPs to diagnose ASD in 18–72-month-olds flagged by clinicians or caregivers as having a potential developmental delay. The Device’s user-friendly inputs, timely result provision, and indeterminate output were all designed to maximize its utility, safety, and trustworthiness in the primary care setting. Compared to assessment tools currently available in specialty settings20,30,31, the Device tested in this study requires less time to administer and less specialty training. It also captures video data that provides rich insight into the child’s natural behavior outside of the clinic setting. Of note, as a mobile system, the Device is amenable to administration via telemedicine, making it adaptable for use in remote and rural settings as well as during public health emergencies such as the current COVID-19 pandemic.

For nearly a third of this primary care study sample, the Device supported efficient and highly accurate diagnostic evaluations in conjunction with clinical judgment. This is significant since, currently, only ~1% of children with ASD in the US are diagnosed in primary care21,22. Of the children for whom the Device made a determinate diagnostic evaluation, 98.4% with ASD received an ASD positive Device result and 78.9% without ASD received an ASD negative Device result. Moreover, 80.8% of children who received an ASD positive Device result were true positives and 98.3% of children with an ASD negative Device result were true negatives. None of the 15 children who received a false-positive Device result were clinically assessed as being neurotypical (without ASD or other non-ASD conditions suspected) and all had a non-ASD developmental-behavioral pediatric condition that could potentially benefit from similar early interventions as ASD. A third of these false-positive cases had one specialist clinician determine that they met diagnostic criteria for ASD. The Device was designed to minimize false negatives to safeguard against missing a diagnosis of ASD, which could have profound consequences such as delayed treatment initiation. There was only a single false negative in this study.

The results of this double-blind active comparator study are promising for future clinical practice and timely in light of the most recent AAP clinical report that advocates for the development of tools to aid in the diagnosis of ASD18. Replication and follow-on work are needed, but study results support the potential of the Device to enable a larger portion of children to be diagnosed in a primary care setting than is currently occurring22. The Device also has the potential to reduce the mean age of diagnosis and time to diagnosis for a subset of children. For example, the mean age of completers for whom the Device rendered an ASD positive result was 2.81 years, which is 1.5 years earlier than the current average age of diagnosis1,9,10. Time burden associated with Device use was also considerably lower than that of existing assessment tools, making it potentially more practical to deploy in time-pressured primary care environments.

The finding of Device performance consistency across sex, race/ethnicity, income, and parental education level is also encouraging, providing preliminary evidence for the Device’s potential to address some well-known ASD diagnostic disparities. When the Device provided a result, for example, it correctly identified 92.3% of girls with ASD. This finding is important given that gender and racial/ethnic biases exist in the current standard of care, such that females, African American, and Hispanic children are less often diagnosed, misdiagnosed more often, and when diagnosed, are diagnosed later on average47,48,49,50. There was also no evidence of performance degradation when the HCP questionnaire assessments were performed remotely versus in-person. This study finding is reassuring and speaks to the potential for the Device to reduce lags in diagnosis for some vulnerable populations. For example, those who live in rural and remote areas, or low-income families for whom taking time off work to bring their child to in-person assessments may prove challenging.

Although the sample population was ethnically and racially diverse, the study was not powered for statistical inference on covariates such as comorbidities, gender, race/ethnicity, education level, or income level. Future studies that include larger samples of subpopulations are needed to build upon the initial finding of consistent Device performance across these covariates. Formal IQ testing was not conducted as part of the study as a large portion of participants were below the age threshold where non-verbal cognitive abilities can be reliably tested with traditional measures of intelligence51. Appropriately powered follow-up studies are needed to confirm Device performance across a range of intellectual levels. Additionally, the Device is currently available only in the English language, and non-English speaking families were excluded from the study. In order to address this bias, multiple language options for non-English speakers are currently in development including Mandarin and Spanish. Compared to the general US population the study cohort also had a higher level of parental education, which is common among clinical trials52, and a lower household income. Lower-income may reflect reduced parental participation in the workforce as a result of participants’ developmental delays53.

Cost and reimbursement data are also needed to clarify the extent to which the Device could be utilized equitably by primary care HCPs in practice. In light of growing pressures on Medicaid programs to cover the early diagnosis of ASD and concurrent budgetary challenges, better use of existing primary care infrastructure, such as that offered by this Device, may support Medicaid programs to remain financially sustainable while adhering to laws and regulations. Additional costing data are needed, however, to fully understand the likelihood of the Device being widely adopted in US primary care settings, and the extent to which it would be accessible to low-income families or families without health insurance. Additionally, Device use requires access to a smartphone which not all low-income American families may have54.

As a risk control measure to minimize the likelihood of false negatives, an indeterminate output that sacrificed coverage was built into the Device. Given the often complex presentation of ASD in primary care settings, including multiple phenotypically overlapping comorbidities and behavioral features that may unfold over time, the finding that the Device abstained from making a recommendation for two out of every three children in this cohort is unsurprising. ASD has degrees of diagnostic complexity that depend on the presentation. Even among experienced specialists, considerable diagnostic uncertainties persist, for example, when diagnosing girls55 or children with moderate (vs high or low) levels of observable ASD symptoms56. Many children with developmental delays are not autistic or have co-occurring conditions which confound the diagnosis. The drop in sensitivity and specificity that would have been observed in this study were the indeterminate output removed, highlights the importance of abstention as a safety mechanism when deploying AI within complex clinical scenarios46. The goal of the Device is not to enable diagnosis of all presentations of ASD, but to aid primary care HCPs to diagnose the subset of children the Device can confidently determine a recommendation for, potentially precluding the need to refer all children for tertiary center evaluation. Reducing tertiary referral loads by even a third could significantly shorten wait times for interventions to begin.

Feasibility data57 suggest the indeterminate rate will also vary depending on the makeup of the population the Device is being applied to. In our study population, there was a high rate of complex determinations and a relatively low neurotypical rate (9.4%). In a more neurotypical population, we would anticipate seeing fewer Device abstentions and higher Device specificity. In populations with a higher ASD prevalence than that observed in the study cohort, we would also expect fewer abstentions.

Areas of focus for future research include improving the scalability of the video component of the Device by decreasing reliance on human video analysts in subsequent generations of the algorithm. Research exploring the training and education requirements of HCPs in primary care settings such that they would feel confident using the Device as a diagnostic aid in practice is also needed. Future Device studies currently planned include a registrational study to monitor the stability of Device results as compared to diagnosis over time, including establishing a diagnosis later in children who at the time of initial evaluation did not meet diagnostic criteria for ASD.

Reducing the age of ASD diagnosis and time to diagnosis is essential if early intervention is to commence in the window of brain development where it is most effective. Limited diagnostic capacity in primary care settings, including a lack of tailored diagnostic aids, together with long wait times for specialist evaluations, contribute to current diagnostic bottlenecks. In this double-blind active comparator study of 18–72-month-olds with identified developmental delay concerns, the accuracy of a Device designed to aid HCPs to diagnose ASD was assessed. The AI-based Device was found to facilitate timely and accurate ASD diagnostic evaluation in nearly a third of children in the clinical trial setting while minimizing false negatives across the cohort to maintain clinical safety. While future research is needed, the Device shows the potential to expand primary care diagnostic capacity, thereby enabling earlier intervention for a subset of children and more efficient use of limited specialist resources.


This double-blinded active comparator cohort pivotal study was conducted to support the FDA Market Authorization for the Device, a novel ASD diagnosis aid designed for use in primary care settings.

Clinical trial registration

This study was registered on (NCT04151290) prior to study initiation.


The study protocol and informed consent forms were reviewed and approved by a centralized Institutional Review Board (IntegReview IRB). Protocol Number: Q170886. IntegReview IRB granted approval of the study (protocol version 1.0) on 19 July 2019. IntegReview was subsequently purchased by Advarra IRB. Informed consent was obtained from all caregivers whose children participated in the study.

Recruitment and screening

Children 18–72 months old with identified concerns for developmental delay by an HCP or caregiver were recruited for the study from 14 sites across six US states. Participants represented a population of patients that HCPs would see in their primary care practice. All participants had a caregiver with functional English capability. Caregiver-reported demographic data in the form of participant age, sex, race/ethnicity, and parental education and income were collected to evaluate performance variability among subgroups due to well-documented disparities in ASD diagnosis14.

Device description

The diagnosis aid is a SaMD that uses a machine learning algorithm modeled after standard medical evaluation methodologies. The Device comprises the following: a caregiver-facing mobile application, a video analyst portal, a healthcare provider portal, the underlying machine learning algorithm that drives the Device outputs, and several supporting software and backend services and infrastructure, including privacy and security encryption and infrastructure in compliance with HIPAA and other best practices.

Device interfaces were designed to be user-centric and easy-to-navigate. Relevant Apple, Google, and general Web content accessibility guidelines were followed across all Device interfaces to maximize user inclusivity. Human factors and usability validation testing were conducted throughout the Device design process to ensure external users were able to interact with the Device interfaces without use errors or patterns of use errors that could lead to harm to a patient. Additionally, a risk management process was utilized to identify potential use errors and ensure they were adequately mitigated to a safe and acceptable level. The Caregiver-facing app functions across both iOS and Android platforms. The HCP portal supports the most recent and previous versions of Chrome and Safari browsers running on the most recent and previous versions of Mac and Windows operating systems. The Video Analyst Portal supports the most recent Safari browser running on the most recent iPad operating systems (Fig. 1).

Each Device input has two age-dependent questionnaire versions (18–47 months or 48–71 months). Versions are tailored to reflect developmentally relevant features. The child is automatically assigned the relevant version based on their age. Across the three inputs, each of the two age groups receives 64 questions, however, the breakdown of the number of questions per input varies (number of questions for 18–47-month-olds: input 1 = 18 questions, input 2 = 33 questions, input 3 = 13 questions. Number of questions for 48–71-month-olds: input 1 = 21 questions, input 2 = 28 questions, input 3 = 15 questions). The Device was trained to handle this variation and perform with equal accuracy for both age groups.

Questionnaire content stems from feature ranking experiments conducted previously to identify behavioral, executive functioning, and language and communication features that are maximally predictive of an ASD diagnosis38,39,40,41,42,43. Input questions elicit information about these core behaviors. Behavioral features used by the classifiers vary only slightly between the two age groups. For example, a few questions seek different information in the domains of social interaction and communication in order to best capture the ASD phenotype among children based on developmental trajectories. Core behaviors probed by Device questionnaires are as follows: the ability to integrate different forms of communication, joint attention, pretend play with toys, anger or aggression, language and communication (non-verbal, expressive, receptive and speech), quality of social responses, anxiety level, range and amount of facial expressions, appropriate play, reciprocal communication, appropriateness of eye contact, repetitive mannerisms, creative play, level of engagement, responsive smile, directed gaze, negative response to stimuli, self-injury, group play, obsessive-compulsive, sensory interests, hyperactivity, offering comfort to others, shared interests, imitation, overall developmental challenges/delays, socially directed smile, initiation of activities or interactions, overall quality of interactions with others, unusual interests, interest in others, pretend play with others.

Study flow

Study treatment (ASD assessment using the device)

After providing written informed consent, subject caregivers used the Device Application on their smartphone to complete the caregiver assessment (Device input 1) and record two brief videos of their child (to be used in completion of Device input 2). A HCP completed Device input 3. Results were rapidly available upon completion of the three inputs. The caregivers, video analysts, and HCPs were blinded to each other’s input to the Device and to the Device output (Fig. 3).

  • Device input 1: Caregivers completed an 18 or 21-item age-dependent questionnaire via a mobile application.

  • Device input 2: Caregivers used the mobile application to upload two, short videos of their child interacting, playing, or talking in natural settings. Videos must be a minimum of 1 min and 30 s, up to 5 min each. The application instructed caregivers on how to take high-quality videos (e.g., showing child’s hands and face with sufficient lighting). In order to ensure strong inter-rater reliability, all video analysts received standardized training, performance analysis, and on-going performance reviews during the study. Prior to receiving video analyst certification, all video analysts were tested on a set of previously unseen training video submissions. These submissions represented a mix of each age group. Analyst performance was required to meet or exceed the operational minimum performance guarantees for PPV, NPV, as well as the abstention rate in order for them to receive certification. In addition to rigorous testing and training standards, all video analysts involved in the study had to meet stringent educational and clinical eligibility requirements. These requirements included: at least a master’s degree from professional backgrounds including psychology, occupational therapy, physical therapy, speech-language pathology, special education, or a related field with specific training in ASD diagnosis and/or treatment, and; at least 5 years of professional and/or clinical experience working with children with ASD. The video analyst scores served as input 2 to the Device.

  • Device input 3: A HCP met with the caregiver and child during a 30-min in-person or virtual visit and completed a 13 or 15-item age-dependent questionnaire. Remote visits using a telemedicine platform were held in a manner equivalent to those done in-person, while protecting the subject’s safety and privacy. There was oversight to ensure the methods and conduct of remote assessments were consistent across sites and study subjects to minimize variability in the data.

Fig. 3: Study flow.
figure 3

Consenting participants meeting inclusion criteria will complete both the Device inputs and the clinical reference standard evaluation.

Time burden associated with Device use was captured electronically for inputs 1 and 2. As HCPs completed a paper version of their questionnaire that was later uploaded to the portal, the time burden associated with input 3 was obtained qualitatively by asking HCPs to estimate the number of minutes questionnaires took to complete. After the caregiver assessment was completed and scorable videos submitted, the caregiver was contacted by a research coordinator to schedule an appointment for a diagnostic evaluation by the diagnosing specialist clinician.

Specialist diagnostic assessments

Specialist assessments were conducted by board-certified child and adolescent psychiatrists, child neurologists, developmental-behavioral pediatricians, or child psychologists with more than five years of experience diagnosing ASD. Specialists used structured clinical observation, clinician interview and examination, medical/developmental review, and standardized assessment instruments to provide a diagnosis based on DSM-5 criteria. The structured observational assessment included conversation or free play, symbolic interactive play, and sensory stimulation components. This robust diagnostic process is aligned with best practice recommendations for ASD evaluation44. Standardized medical history, physical examination findings, and a video recorded portion of the initial specialist’s assessment were provided to a blinded reviewing specialist clinician who independently evaluated whether DSM-5 criteria for ASD were met. When the first two specialists disagreed, a third specialist was consulted, and the majority decision determined the clinical reference standard diagnosis. Blinded consensus diagnosis introduced an additional level of rigor to the study design, reducing risks associated with confirmation bias, deindividuation, and homogenized group thinking 35.

If a specialist clinician: (1) diagnosed co-morbid conditions, (2) determined that the patient was negative for ASD and provided a diagnosis other than ASD, or (3) determined that the patient was neurotypical, those data were also captured. Diagnosis of ASD alone or ASD plus co-morbid condition(s) constituted a positive ASD clinical reference standard diagnosis. Clinicians were also asked to self-report their degree of certainty of the ASD diagnostic conclusion on a Likert scale, with 1 being, “completely uncertain”, 2 being “somewhat uncertain”, 3 being “somewhat certain”, and 4 being “completely certain”.

Statistical analyses

Statistical analyses were conducted using R 4.0.2 with the PropCIs and Exact packages. Endpoints (PPV, NPV, sensitivity, specificity, and determinate result rate) were evaluated using contingency tables. The selection of predictive values as primary endpoints reflects the utility of the Device in a broad spectrum of subjects, as these values measure how often in the study population the Device correctly identifies a patient with ASD and how frequently it correctly determines that a patient does not have ASD. The Clopper-Pearson interval was used to calculate two-sided 95% CI. The descriptive statistics for sex, race/ethnicity, income, and education were tabulated separately; Device performance was reported for each group and separate two-sided 95% CI were computed. For age analyses, Boschloo’s Exact Test was used to examine statistical differences in performance among subgroups.

Reporting summary

Further information on research design is available in the Nature Research Reporting Summary linked to this article.

Data availability

Data are not publicly available because they contain sensitive patient information. Individual, de-identified, participant data that underlie the results reported in this article and study protocol may be made available to qualified researchers upon reasonable request. Proposals should be directed to to gain access. Data requestors will be required to sign a data-sharing agreement prior to access. The full study protocol is available on

Code availability

The code used in the study is proprietary and is not publicly available.


  1. Maenner, M. J., Shaw, K. A. & Baio, J. Prevalence of autism spectrum disorder among children aged 8 years—autism and developmental disabilities monitoring network, 11 sites, United States, 2016. MMWR Surveill. Summ. 69, 1 (2020).

    PubMed  PubMed Central  Article  Google Scholar 

  2. Mazurek, M. O., Kanne, S. M. & Miles, J. H. Predicting improvement in social–communication symptoms of autism spectrum disorders using retrospective treatment data. Res. Autism Spectr. Disord. 6, 535–545 (2012).

    Article  Google Scholar 

  3. MacDonald, R., Parry-Cruwys, D., Dupere, S. & Ahearn, W. Assessing progress and outcome of early intensive behavioral intervention for toddlers with autism. Res. Dev. Disabil. 35, 3632–3644 (2014).

    PubMed  Article  Google Scholar 

  4. Ben Itzchak, E. & Zachor, D. A. Who benefits from early intervention in autism spectrum disorders? Res. Autism Spectr. Disord. 5, 345–350 (2011).

    Article  Google Scholar 

  5. Flanagan, H. E., Perry, A. & Freeman, N. L. Effectiveness of large-scale community-based intensive behavioral intervention: a waitlist comparison study exploring outcomes and predictors. Res. Autism Spectr. Disord. 6, 673–682 (2012).

    Article  Google Scholar 

  6. Vivanti, G. & Dissanayake, C., The Victorian ASELCC Team. Outcome for children receiving the early start denver model before and after 48 months. J. Autism Dev. Disord. 46, 2441–2449 (2016).

    PubMed  Article  Google Scholar 

  7. Smith, T., Klorman, R. & Mruzek, D. W. Predicting outcome of community-based early intensive behavioral intervention for children with autism. J. Abnorm. Child Psychol. 43, 1271–1282 (2015).

    PubMed  Article  Google Scholar 

  8. Dawson, G. et al. Randomized, controlled trial of an intervention for toddlers with autism: the Early Start Denver Model. Pediatrics 125, e17 (2010).

    PubMed  Article  Google Scholar 

  9. van’t Hof, M. et al. Age at autism spectrum disorder diagnosis: a systematic review and meta-analysis from 2012 to 2019. Autism 25, 862–873 (2021).

    Article  Google Scholar 

  10. Pierce, K. et al. Evaluation of the diagnostic stability of the early autism spectrum disorder phenotype in the general population starting at 12 months. JAMA Pediatr. 173, 578–587 (2019).

    PubMed  PubMed Central  Article  Google Scholar 

  11. Maenner, M. J. et al. Prevalence and characteristics of autism spectrum disorder among children aged 8 years—autism and developmental disabilities monitoring network, 11 sites, United States, 2018. MMWR Surveill. Summ. 70, 1 (2021).

    PubMed  PubMed Central  Article  Google Scholar 

  12. Delobel-Ayoub, M. et al. Socioeconomic disparities and prevalence of autism spectrum disorders and intellectual disability. PloS One 10, e0141964 (2015).

    PubMed  PubMed Central  Article  CAS  Google Scholar 

  13. Oswald, D. P., Haworth, S. M., Mackenzie, B. K. & Willis, J. H. Parental report of the diagnostic process and outcome: ASD compared with other developmental disabilities. Focus Autism Other Dev. Disabil. 32, 152–160 (2017).

    Article  Google Scholar 

  14. Wiggins, L. D. et al. Disparities in documented diagnoses of autism spectrum disorder based on demographic, individual, and service factors. Autism Res. 13, 464–473 (2020).

    PubMed  Article  Google Scholar 

  15. Shattuck, P. T. et al. Timing of identification among children with an autism spectrum disorder: findings from a population-based surveillance study. J. Am. Acad. Child Adolesc. Psychiatry 48, 474–483 (2009).

    PubMed  PubMed Central  Article  Google Scholar 

  16. Gordon-Lipkin, E., Foster, J. & Peacock, G. Whittling down the wait time: exploring models to minimize the delay from initial concern to diagnosis and treatment of autism spectrum disorder. Pediatr. Clin. 63, 851–859 (2016).

    Google Scholar 

  17. Elsabbagh, M. et al. Global prevalence of autism and other pervasive developmental disorders. Autism Res. 5, 160–179 (2012).

    PubMed  PubMed Central  Article  Google Scholar 

  18. Hyman, S. L., Levy, S. E. & Myers, S. M. Council on Children with Disabilities S on D and BP. Identification evaluation, and management of children with autism spectrum disorder. Pediatrics 1, e20193447 (2019).

  19. Bridgemohan, C. et al. A workforce survey on developmental-behavioral pediatrics. Pediatrics 141, e20172164 (2018).

  20. Falkmer, T., Anderson, K., Falkmer, M. & Horlin, C. Diagnostic procedures in autism spectrum disorders: a systematic literature review. Eur. Child Adolesc. Psychiatry 22, 329–340 (2013).

    PubMed  Article  Google Scholar 

  21. Rhoades, R. A., Scarpa, A. & Salley, B. The importance of physician knowledge of autism spectrum disorder: results of a parent survey. BMC Pediatrics 7, 37 (2007).

    PubMed  PubMed Central  Article  Google Scholar 

  22. Monteiro, S. A., Dempsey, J., Berry, L. N., Voigt, R. G. & Goin-Kochel, R. P. Screening and referral practices for autism spectrum disorder in primary pediatric care. Pediatrics 144, e20183963 (2019).

  23. Carbone, P. S., Norlin, C. & Young, P. C. Improving early identification and ongoing care of children with autism spectrum disorder. Pediatrics 137, e20151850 (2016).

    PubMed  Article  Google Scholar 

  24. Fenikilé, T. S., Ellerbeck, K., Filippi, M. K. & Daley, C. M. Barriers to autism screening in family medicine practice: a qualitative study. Prim. Health Care Res. Dev. 16, 356–366 (2015).

    PubMed  Article  Google Scholar 

  25. Self, T. L., Parham, D. F. & Rajagopalan, J. Autism spectrum disorder early screening practices: a survey of physicians. Commun. Disord. Q. 36, 195–207 (2015).

    Article  Google Scholar 

  26. Guthrie, W. et al. Accuracy of autism screening in a large pediatric network. Pediatrics 144, e20183963 (2019).

  27. Lai, M.-C. et al. Prevalence of co-occurring mental health diagnoses in the autism population: a systematic review and meta-analysis. Lancet Psychiatry 6, 819–829 (2019).

    PubMed  Article  Google Scholar 

  28. Nazneen, N. et al. Supporting parents for in-home capture of problem behaviors of children with developmental disabilities. Personal. Ubiquitous Comput. 16, 193–207 (2012).

    Article  Google Scholar 

  29. Cognoa. Cognoa Receives FDA Marketing Authorization for First-of-its-kind Autism Diagnosis Aid. (2021).

  30. Lord, C. et al. Autism diagnostic observation schedule, (ADOS-2) modules 1–4. (2012).

  31. Lord, C., Rutter, M. & Le Couteur, A. Autism diagnostic interview-revised: a revised version of a diagnostic interview for caregivers of individuals with possible pervasive developmental disorders. J. Autism Dev. Disord. 24, 659–685 (1994).

    CAS  PubMed  Article  Google Scholar 

  32. Cicchetti, D. V., Lord, C., Koenig, K., Klin, A. & Volkmar, F. R. Reliability of the ADI-R: Multiple examiners evaluate a single case. J. Autism Dev. Disord. 38, 764–770 (2008).

    PubMed  Article  Google Scholar 

  33. Hill, A. et al. Stability and interpersonal agreement of the interview-based diagnosis of autism. Psychopathology 34, 187–191 (2001).

    CAS  PubMed  Article  Google Scholar 

  34. Randall, M. et al. Diagnostic tests for autism spectrum disorder (ASD) in preschool children. Cochrane Database Syst. Rev. (2018).

  35. Kaufman, N. K. Rethinking “gold standards” and “best practices” in the assessment of autism. Appl. Neuropsychol. Child 1–12 (2020).

  36. Hyde, K. K. et al. Applications of supervised machine learning in autism spectrum disorder research: a review. Rev. J. Autism Dev. Disord. 6, 128–146 (2019).

    Article  Google Scholar 

  37. FDA. Software as a Medical Device (SaMD) (2018).

  38. Abbas, H., Garberson, F., Liu-Mayo, S., Glover, E. & Wall, D. P. Multi-modular AI approach to streamline autism diagnosis in young children. Sci. Rep. 10, 1–8 (2020).

    CAS  Article  Google Scholar 

  39. Abbas, H., Garberson, F., Glover, E. & Wall, D. P. Machine learning approach for early detection of autism by combining questionnaire and home video screening. J. Am. Med. Inform. Assoc. 25, 1000–1007 (2018).

    PubMed  PubMed Central  Article  Google Scholar 

  40. Levy, S., Duda, M., Haber, N. & Wall, D. P. Sparsifying machine learning models identify stable subsets of predictive features for behavioral detection of autism. Mol. Autism 8, 1–17 (2017).

    Article  Google Scholar 

  41. Tariq, Q. et al. Mobile detection of autism through machine learning on home video: a development and prospective validation study. PloS Med. 15, e1002705 (2018).

    PubMed  PubMed Central  Article  Google Scholar 

  42. Wall, D. P., Dally, R., Luyster, R., Jung, J.-Y. & DeLuca, T. F. Use of artificial intelligence to shorten the behavioral diagnosis of autism. 7, e43855 (2012).

  43. Kosmicki, J. A., Sochat, V., Duda, M. & Wall, D. P. Searching for a minimal set of behaviors for autism detection through feature selection-based machine learning. Transl. Psychiatry 5, e514 (2015).

    CAS  PubMed  PubMed Central  Article  Google Scholar 

  44. Brian, J. A., Zwaigenbaum, L. & Ip, A. Standards of diagnostic assessment for autism spectrum disorder. Paediatr Child Health 24, 444–451 (2019).

    PubMed  PubMed Central  Article  Google Scholar 

  45. Cortes, C., DeSalvo, G., Gentile, C., Mohri, M. & Yang, S. Online learning with abstention. International conference on machine learning 1059–1067 (2018).

  46. Kompa, B., Snoek, J. & Beam, A. L. Second opinion needed: communicating uncertainty in medical machine learning. NPJ Digit. Med. 4, 1–6 (2021).

    Article  Google Scholar 

  47. Constantino, J. N. et al. Timing of the diagnosis of autism in african american children. Pediatrics 146, e20193629 (2020).

    PubMed  Article  Google Scholar 

  48. Yingling, M. E., Hock, R. M. & Bell, B. A. Time-lag between diagnosis of autism spectrum disorder and onset of publicly-funded early intensive behavioral intervention: do race–ethnicity and neighborhood matter? J. Autism Dev. Disord. 48, 561–571 (2018).

    PubMed  Article  Google Scholar 

  49. Becerra, T. A. et al. Autism spectrum disorders and race, ethnicity, and nativity: a population-based study. Pediatrics 134, e63 (2014).

    PubMed  PubMed Central  Article  Google Scholar 

  50. Ros-Demarize, R. et al. ASD symptoms in toddlers and preschoolers: an examination of sex differences. Autism Res. 13, 157–166 (2020).

    PubMed  Article  Google Scholar 

  51. Hus, Y. & Segal, O. Challenges surrounding the diagnosis of autism in children. Neuropsychiatr. Dis. Treat. 17, 3509 (2021).

    PubMed  PubMed Central  Article  Google Scholar 

  52. Waldron, E. M., Hong, S., Moskowitz, J. T. & Burnett-Zeigler, I. A systematic review of the demographic characteristics of participants in us-based randomized controlled trials of mindfulness-based interventions. Mindfulness 9, 1671–1692 (2018).

    Article  Google Scholar 

  53. Liao, X. & Li, Y. Economic burdens on parents of children with autism: a literature review. CNS Spectr. 25, 468–474 (2020).

    PubMed  Article  Google Scholar 

  54. Pew Research Centre. Mobile Fact Sheet. (2021).

  55. Westman Andersson, G., Miniscalco, C. & Gillberg, C. Autism in preschoolers: does individual clinician’s first visit diagnosis agree with final comprehensive diagnosis? Sci. World J. 2013, 716267 (2013).

    Google Scholar 

  56. McDonnell, C. G. et al. When are we sure? predictors of clinician certainty in the diagnosis of autism spectrum disorder. J. Autism Dev. Disord. 49, 1391–1401 (2019).

    PubMed  Article  Google Scholar 

  57. Cognoa. Feasibility Data (unpublished). (2019).

Download references


We thank the following individuals for their contribution to the study: John N. Constantino, MD; Clara Lajonchere, Ph.D.; Nauman Ahmad, MD; Rogelio Amisola, MD; Judith Aronson-Ramos, MD; Sheila Balog, Ph.D.; Thomas Barela, MD; Jacques Benun, MD; Laura Carpenter, Ph.D.; Chantal Caviness, MD, Ph.D.; Molly Chartrand, MD; Omelia Elizabeth Danford, MD; Rebecca Garfinkle, DO; Addison Gearhart, MD; Urmila Gupta, MD; Anna Elizabeth Holliman, MD; Rosemarie Manfredi, PsyD; Andrew McIntosh, MD; Iqbal Memon, MD; Ann Moell, MD; Megan Norris, Ph.D.; Melissa J. Przeklasa Auth, MD; Linda Ribnik, MD; Sonam Sidhu, MD; Gilbert Templeton, MD; Karen Toth, Ph.D.; Ashley Toler, DO; Duc Vo, MD; Jennifer L. Hugunin, B.A.; Rachel Davis, Ph.D; Carmela Salomon, Ph.D. This study was sponsored by Cognoa Inc. The study design, statistical methods, and endpoints were collaboratively determined by the sponsor, its clinical and scientific advisors, independent clinicians and researchers, study investigators, and the FDA under Breakthrough Designation. The study was conducted with a clinical research organization; the sponsor was blinded and unable to influence the results.

Author information

Authors and Affiliations



Dr. Taraman substantially contributed to the conception and design of the study, served as the study’s primary investigator, interpreted the data, and drafted the article. Drs. Megerian, Coury, Sohl, and Wall substantially contributed to the conception and design of the study, interpretation of the data, and revised the manuscript critically for important intellectual content. Drs. Dey, Melmed, Nicholls, Rouhbakhsh, Narasimhan, Shareef, Lerner, Romain, and Golla contributed substantially to the acquisition and interpretation of data and revised the manuscript critically for important intellectual content. Drs. Ostrovsky, Shannon, Kraft, and Gal-Szabo contributed substantially to the analysis and interpretation of data and revised the manuscript critically for important intellectual content. Mr. Liu-Mayo and Mr. Abbas contributed to the conception and design of the study, conducted an analysis of the data, drafted, reviewed, and revised the statistical methods and results in portions of the manuscript. All authors approved the final manuscript as submitted and agree to be accountable for all aspects of the work.

Corresponding author

Correspondence to Sharief Taraman.

Ethics declarations

Competing interests

Dr. Megerian and Dr. Taraman had full access to all the data in the study following the breaking of the blinding and take responsibility for the integrity of the data and the accuracy of the data analysis. Dr. Carpenter, Dr. Lajonchere, Dr. Lerner, Dr. Megerian, Dr. Nicholls, and Dr. Ostrovsky have received consulting fees from Cognoa. Dr. Constantino serves on the Advisory Board for Cognoa. Dr. Coury has received consulting fees from BioRosa, Cognoa, Quadrant Biosciences, and Stalicla and receives research support from Autism Speaks and Quadrant Biosciences. Dr. Sohl has received consulting fees from Cognoa, is on the Medical Advisory Board for Quadrant Biosciences, and provides research support for Autism Speaks. Mr. Abbas, Dr. Kraft, Mr. Liu-Mayo, Dr. Shannon, and Dr. Taraman are employees of Cognoa and have Cognoa stock options. Dr. Taraman additionally receives consulting fees for Cognito Therapeutics, volunteers as a board member of the AAP—OC chapter and AAP—California, is a paid advisor for MI10 LLC, and owns stock for NTX, Inc., and HandzIn. Dr. Kraft is also a consultant for SOBI, Inc. and Happiest Baby, Inc., and serves on the Advisory Board for DotCom Therapy. Dr. Wall is the co-founder of Cognoa, is on the board of directors, and holds Cognoa stock. Dr. Gal-Szabo is an independent contractor for Cognoa. Drs. Dey, Melmed, Rouhbakhsh, Narasimhan, Romain, Golla, and Shareef have no conflicts of interest to disclose.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Megerian, J.T., Dey, S., Melmed, R.D. et al. Evaluation of an artificial intelligence-based medical device for diagnosis of autism spectrum disorder. npj Digit. Med. 5, 57 (2022).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI:


Quick links

Nature Briefing

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing