Suicide, depression, and anxiety screening using an automated virtual assistant - Summary: Each year in the US, more than 47,000 people die by suicide and many of these tragedies could be prevented with early identification and treatment. At-risk individuals rarely receive screening until depression, anxiety, and suicide symptoms are severe. Universal screening tools deployed in schools and physicians’ offices could address this problem, but the lack of an objective tool along with a shortage of mental health clinicians in these settings is a major barrier to screening on a public health scale. Current methods are subject to inter-rater variability, require experience for interpretation, and are somewhat cumbersome to administer. Meta-analyses have indicated that predicting suicide death using current methods does no better than random chance. There is a need for a singular, easy-to-use, accurate, reliable, objective clinical decision support tool that can be used as a clinical and public health approach to identify patients at risk of suicide, depression, and anxiety to guide at-risk individuals to life-saving resources and interventions. The complexity and broad variability of so- called “thought markers” such as verbal and non-verbal language cues resist easy classification by a clinician, but artificial intelligence (AI) approaches can identify consistent, meaningful patterns. By using a natural language processing AI approach, we have developed Clairity™, an AI-enabled app that automates the identification of at-risk individuals based on speech and vocal markers. Initial development has demonstrated Clairity’s potential for predicting suicide risk, depression, or anxiety, but it currently relies on a live interviewer. Perpetual health care worker shortages, exacerbated by the pandemic, present a major barrier to scalability. Interactive voice response (IVR) may be a solution to health system strain that has no signs of abatement. An IVR system could also reduce inter-interviewer variability and, for some patients, entice better engagement. We will assess the effects of user perception of IVR as well as demographics on model performance when using an IVR system in comparison to human interviews. We will validate modified models to inform key decisions regarding our commercialization path and future trial designs. Aim 1. Characterize the effects of IVR data collection on model performance. Factors that decrease model performance by ≥ 10% AUC versus live interviewer will be identified for model improvement. Aim 2. Validate updated algorithms with a holdout data set. Successful models will correctly classify risk categories with an AUC ≥ 80%. We will also analyze demographic differences to inform tuning of the IVR system (e.g., longer pause times, allowing participants to select a specific virtual assistant). Aim 3. Assess user perspective on engagement with IVR. Patient perceptions of virtual versus live interviewers may affect model performance. We will survey patients regarding their perceptions and emotions in dealing with a virtual agent and assess correlation with model performance. Impact: Advances from this project are expected to enable a scalable, real-time clinical decision support tool for risk screening that could improve treatment referral and ultimately reduce suicides.