Abstract
Mental health services and treatment are unfortunately subject to sociodemographic disparities. To address this issue, recent studies have begun to apply analytics methods—that is, artificial intelligence in general, machine learning and deep learning in particular—toward the identification of such disparities and, where possible, mitigation of bias within models used in mental health research. However, it is difficult to understand the scope and status of such research as it is spread across many journals and contexts of study. Here we conducted an analysis of articles in this area. We identified 40 articles from 2017 to July 2023 related to the use of analytics in the context of sociodemographic disparities in mental health. We find that prediction, clustering/grouping and fairness models were most often applied in the articles analyzed. A number of mental health-related sociodemographic disparities were identified in these articles, for example, associated with race/ethnicity, gender, age and socioeconomic status, but such findings were typically context dependent. Thus, we also provide suggestions in this Analysis on how to both enhance generalizability and embrace context-dependent findings, especially via the identification of heterogeneous treatment effects, model bias mitigation, use of generative artificial intelligence, incorporation of data from devices, and translation of findings into practice.
This is a preview of subscription content, access via your institution
Access options
Subscribe to this journal
Receive 12 digital issues and online access to articles
$79.00 per year
only $6.58 per issue
Buy this article
- Purchase on SpringerLink
- Instant access to the full article PDF.
USD 39.95
Prices may be subject to local taxes which are calculated during checkout


Similar content being viewed by others
Data availability
The data are journal articles obtained from the searches described earlier of Web of Science (https://www.webofscience.com/wos/woscc/basic-search), PubMed (https://pubmed.ncbi.nlm.nih.gov/), IEEE Xeplore (https://ieeexplore.ieee.org/Xplore/home.jsp), PsycInfo (https://psycinfo.apa.org) and Google Scholar (https://scholar.google.com). All articles used in our analyses are referenced in this analysis.
Code availability
Code is available via GitHub at https://github.com/ProfAaronBaird/JournalArticleRepo/tree/main/NatureMHAnalyticsDisparities.
References
Safran, M. A. et al. Mental health disparities. Am. J. Public Health 99, 1962–1966 (2009).
Braveman, P. Health disparities and health equity: concepts and measurement. Annu. Rev. Public Health 27, 167–194 (2006).
Alang, S. M. Sociodemographic disparities associated with perceived causes of unmet need for mental health care. Psychiatr.Rehab. J. 38, 293–299 (2015).
Barr, P. B., Bigdeli, T. B. & Meyers, J. L. Prevalence, comorbidity and sociodemographic correlates of psychiatric disorders reported in the All of Us Research program. JAMA Psychiatry 79, 622–628 (2022).
Joynt Maddox, K. E. et al. Adjusting for social risk factors impacts performance and penalties in the hospital readmissions reduction program. Health Serv. Res. 54, 327–336 (2019).
Kleindorfer, D. Sociodemographic groups at risk: race/ethnicity. Stroke 40, S75–S78 (2009).
Edwards, A. M. et al. Health disparities among rural individuals with mental health conditions: a systematic literature review. J. Rural Ment. Health 47, 163 (2023).
Cook, B. L. et al. A review of mental health and mental health care disparities research: 2011–2014. Med. Care Res. Rev. 76, 683–710 (2019).
Ralston, A. L., Andrews, A. R. III & Hope, D. A. Fulfilling the promise of mental health technology to reduce public health disparities: review and research agenda. Clin. Psychol. Sci. Pract. 26, e12277 (2019).
Brown, A. F. et al. Structural interventions to reduce and eliminate health disparities. Am. J. Public Health 109, S72–S78 (2019).
Barksdale, C. L., Pérez-Stable, E. & Gordon, J. Innovative directions to advance mental health disparities research. Am. J. Psychiatry 179, 397–401 (2022).
Harris, E. CDC report documents disparities in mental health-related ED visits. JAMA 329, 968 (2023).
Mohri, M., Rostamizadeh, A. & Talwalkar, A. Foundations of Machine Learning 2nd edn (MIT Press, 2018).
Graham, S. et al. Artificial intelligence for mental health and mental illnesses: an overview. Curr. Psychiatry Rep. 21, 1–18 (2019).
Su, C., Xu, Z., Pathak, J. & Wang, F. Deep learning in mental health outcome research: a scoping review. Transl. Psychiatry 10, 116 (2020).
Chen, I. Y., Szolovits, P. & Ghassemi, M. Can AI help reduce disparities in general medical and mental health care? AMA J. Ethics 21, 167–179 (2019).
Baird, A., Xia, Y. & Cheng, Y. Consumer perceptions of telehealth for mental health or substance abuse: a Twitter-based topic modeling analysis. JAMIA Open 5, ooac028 (2022).
Wolfswinkel, J. F., Furtmueller, E. & Wilderom, C. P. Using grounded theory as a method for rigorously reviewing literature. Eur. J. Inf. Syst. 22, 45–55 (2013).
Andersson, S., Bathula, D. R., Iliadis, S. I., Walter, M. & Skalkidou, A. Predicting women with depressive symptoms postpartum with machine learning methods. Sci. Rep. 11, 7877 (2021).
Ueda, M., Watanabe, K. & Sueki, H. Emotional distress during COVID-19 by mental health conditions and economic vulnerability: retrospective analysis of survey-linked Twitter data with a semisupervised machine learning algorithm. J. Med. Internet Res. 25, e44965 (2023).
Kung, B., Chiang, M., Perera, G., Pritchard, M. & Stewart, R. Identifying subtypes of depression in clinician-annotated text: a retrospective cohort study. Sci. Rep. 11, 22426 (2021).
Keralis, J. M. et al. Health and the built environment in United States cities: measuring associations using Google Street view-derived indicators of the built environment. BMC Public Health 20, 215 (2020).
Syed, S. et al. Identifying adverse childhood experiences with electronic health records of linked mothers and children in England: a multistage development and validation study. Lancet Digit. Health 4, e482–e496 (2022).
Zhang, X. et al. Identifying the predictors of severe psychological distress by auto-machine learning methods. Inform. Med. Unlocked 39, 101258 (2023).
Herd, T. et al. Individual and social risk and protective factors as predictors of trajectories of post-traumatic stress symptoms in adolescents. Res. Child Adolesc. Psychopathol. 12, 1739–1751 (2022).
Hueniken, K. et al. Machine learning-based predictive modeling of anxiety and depressive symptoms during 8 months of the COVID-19 global pandemic: repeated cross-sectional survey study. JMIR Mental Health 8, e32876 (2021).
Müller, S. R., Chen, X., Peters, H., Chaintreau, A. & Matz, S. C. Depression predictions from GPS-based mobility do not generalize well to large demographically heterogeneous samples. Sci. Rep. 11, 14007 (2021).
Shiba, K. et al. Uncovering heterogeneous associations of disaster‐related traumatic experiences with subsequent mental health problems: a machine learning approach. Psychiatry Clin. Neurosci. 76, 97–105 (2022).
Jha, I. P., Awasthi, R., Kumar, A., Kumar, V. & Sethi, T. Learning the mental health impact of COVID-19 in the United States with explainable artificial intelligence: observational study. JMIR Mental Health 8, e25097 (2021).
Kourou, K. et al. A machine learning-based pipeline for modeling medical, socio-demographic, lifestyle and self-reported psychological traits as predictors of mental health outcomes after breast cancer diagnosis: an initial effort to define resilience effects. Comput. Biol. Med. 131, 104266 (2021).
Coley, R. Y., Johnson, E., Simon, G. E., Cruz, M. & Shortreed, S. M. Racial/ethnic disparities in the performance of prediction models for death by suicide after mental health visits. JAMA Psychiatry 78, 726–734 (2021).
Simon, G. E. et al. What health records data are required for accurate prediction of suicidal behavior? J. Am. Med. Inform. Assoc. 26, 1458–1465 (2019).
Goldman-Mellor, S. J., Bhat, H. S., Allen, M. H. & Schoenbaum, M. Suicide risk among hospitalized versus discharged deliberate self-harm patients: generalized random forest analysis using a large claims data set. Am. J. Prevent. Med. 62, 558–566 (2022).
Park, H. & Lee, K. Using boosted machine learning to predict suicidal ideation by socioeconomic status among adolescents. J. Person. Med. 12, 1357 (2022).
Jordan, J. T. & McNiel, D. E. Homicide–suicide in the United States: moving toward an empirically derived typology. J. Clin. Psychiatry 82, 28068 (2021).
Santamaría-García, H. et al. Uncovering social-contextual and individual mental health factors associated with violence via computational inference. Patterns 2, 100176 (2021).
Qasrawi, R. et al. Machine learning techniques for identifying mental health risk factor associated with schoolchildren cognitive ability living in politically violent environments. Front. Psychiatry 14, 1071622 (2023).
Marcon, G. et al. Patterns of high-risk drinking among medical students: a web-based survey with machine learning. Comput. Biol. Med. 136, 104747 (2021).
Cox, J. W. et al. Identifying factors associated with opioid cessation in a biracial sample using machine learning. Explor. Med. 1, 27 (2020).
Subbaraman, M. S., Lui, C. K., Karriker-Jaffe, K. J., Greenfield, T. K. & Mulia, N. Predictors of alcohol screening quality in a US general population sample and subgroups of heavy drinkers. Prevent. Med. Rep. 29, 101932 (2022).
Kong, Y., Zhou, J., Zheng, Z., Amaro, H. & Guerrero, E. G. Using machine learning to advance disparities research: subgroup analyses of access to opioid treatment. Health Serv. Res. 57, 411–421 (2022).
Baird, A., Cheng, Y. & Xia, Y. Use of machine learning to examine disparities in completion of substance use disorder treatment. PLoS ONE 17, e0275054 (2022).
Foster, J. C., Taylor, J. M. & Ruberg, S. J. Subgroup identification from randomized clinical trial data. Stat. Med. 30, 2867–2880 (2011).
Deng, C. et al. Practical guidance on modeling choices for the virtual twins method. J. Pharm. Stat. 33, 653–676 (2023).
Rajkomar, A., Hardt, M., Howell, M. D., Corrado, G. & Chin, M. H. Ensuring fairness in machine learning to advance health equity. Ann. Intern. Med. 169, 866–872 (2018).
Wang, S. et al. Unpacking the inter- and intra-urban differences of the association between health and exposure to heat and air quality in Australia using global and local machine learning models. Sci. Total Environ. 871, 162005 (2023).
Haimson, O. L. Mapping gender transition sentiment patterns via social media data: toward decreasing transgender mental health disparities. J. Am. Med. Inform. Assoc. 26, 749–758 (2019).
Ramezani, N. et al. The relationship between community public health, behavioral health service accessibility, and mass incarceration. BMC Health Serv. Res. 22, 966 (2022).
Castelnovo, A. et al. A clarification of the nuances in the fairness metrics landscape. Sci. Rep. 12, 4209 (2022).
Thompson, H. M. et al. Bias and fairness assessment of a natural language processing opioid misuse classifier: detection and mitigation of electronic health record data disadvantages across racial subgroups. J. Am. Med. Inform. Assoc. 28, 2393–2403 (2021).
Park, Y. et al. Comparison of methods to reduce bias from clinical prediction models of postpartum depression. JAMA Netw. Open 4, e213909 (2021).
Park, J., Arunachalam, R., Silenzio, V. & Singh, V. K. Fairness in mobile phone-based mental health assessment algorithms: exploratory study. JMIR Form. Res. 6, e34366 (2022).
Solans Noguero, D., Ramírez-Cifuentes, D., Ríssola, E. A. & Freire, A. Gender bias when using artificial intelligence to assess anorexia nervosa on social media: data-driven study. J. Med. Internet Res. 25, e45184 (2023).
Deferio, J. J., Breitinger, S., Khullar, D., Sheth, A. & Pathak, J. Social determinants of health in mental health care and research: a case for greater inclusion. J. Am. Med. Inform. Assoc. 26, 895–899 (2019).
Cirillo, D. et al. Sex and gender differences and biases in artificial intelligence for biomedicine and healthcare. NPJ Digit. Med. 3, 81 (2020).
Perna, G., Alciati, A., Daccò, S., Grassi, M. & Caldirola, D. Personalized psychiatry and depression: the role of sociodemographic and clinical variables. Psychiatry Investig. 17, 193 (2020).
Walsh, C. G. et al. Stigma, biomarkers, and algorithmic bias: recommendations for precision behavioral health with artificial intelligence. JAMIA Open 3, 9–15 (2020).
Straw, I. & Callison-Burch, C. Artificial intelligence in mental health and the biases of language based models. PLoS ONE 15, e0240376 (2020).
Timmons, A. C. et al. A call to action on assessing and mitigating bias in artificial intelligence applications for mental health. Perspect. Psychol. Sci. 18, 1062–1096 (2023).
Gillett, G., Tomlinson, A., Efthimiou, O. & Cipriani, A. Predicting treatment effects in unipolar depression: a meta-review. Pharmacol. Ther. 212, 107557 (2020).
Wager, S. & Athey, S. Estimation and inference of heterogeneous treatment effects using random forests. J. Am. Stat. Assoc. 113, 1228–1242 (2018).
Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K. & Galstyan, A. A survey on bias and fairness in machine learning. ACM Comput. Surv. 54, 1–35 (2021).
Domino, M. E. et al. Do timely mental health services reduce re‐incarceration among prison releasees with severe mental illness? Health Serv. Res. 54, 592–602 (2019).
Kaiser, T. et al. Heterogeneity of treatment effects in trials on psychotherapy of depression. Clin. Psychol. Sci. Pract. 3, 294–303 (2022).
Baird, A., Cheng, Y. & Xia, Y. Determinants of outpatient substance use disorder treatment length-of-stay and completion: the case of a treatment program in the southeast US. Sci. Rep. 13, 13961 (2023).
Bouvier, F. et al. Do machine learning methods lead to similar individualized treatment rules? A comparison study on real data. Stat. Med. 11, 2043–2061 (2024).
Brand, J. E., Xu, J., Koch, B. & Geraldo, P. Uncovering sociological effect heterogeneity using tree-based machine learning. Sociol. Methodol. 51, 189–223 (2021).
Athey, S. & Imbens, G. W. Machine learning methods for estimating heterogeneous causal effects. Stat 1050, 1–26 (2015).
Cheng, L. et al. Evaluation methods and measures for causal learning algorithms. IEEE Trans. Artif. Intell. 3, 924–943 (2022).
Cintron, D. W. et al. Heterogeneous treatment effects in social policy studies: an assessment of contemporary articles in the health and social sciences. Ann. Epidemiol. 70, 79–88 (2022).
Schuetze, B. A. & von Hippel, P. How not to fool ourselves about heterogeneity of treatment effects. Preprint at OSF https://osf.io/preprints/psyarxiv/zg8hv (2023).
Gonzales, A., Guruswamy, G. & Smith, S. R. Synthetic data in health care: a narrative review. PLoS Digit. Health 2, e0000082 (2023).
Pataranutaporn, P. et al. AI-generated characters for supporting personalized learning and well-being. Nat. Mach. Intell. 3, 1013–1022 (2021).
Walonoski, J. et al. Synthea: an approach, method and software mechanism for generating synthetic patients and the syntheticelectronic health care record. J. Am. Med. Inform. Assoc. 25, 230–238 (2018).
Begoli, E., Brown, K., Srinivas, S., & Tamang, S. SynthNotes: a generator framework for high-volume, high fidelity synthetic mental health notes. In Proc. 2018 International Conference on Management of Data (IEEE, 2018).
Baird, A. & Xia, Y. Precision digital health: a proposed framework and future research opportunities. Bus. Inform. Syst. Eng. 66, 261–271 (2024).
Sai, S. et al. Generative AI for transformative healthcare: a comprehensive study of emerging models, applications, case studies and limitations. IEEE Access 12, 31078–31106 (2024).
Abd-Alrazaq, A. A., Rababeh, A., Alajlani, M., Bewick, B. M. & Househ, M. Effectiveness and safety of using chatbots to improve mental health: systematic review and meta-analysis. J. Med. Internet Res. 22, e16021 (2020).
Baowaly, M. K., Lin, C.-C., Liu, C.-L. & Chen, K.-T. Synthesizing electronic health records using improved generative adversarial networks. J. Am. Med. Inform. Assoc. 26, 228–241 (2019).
Wang, H. et al. Scientific discovery in the age of artificial intelligence. Nature 620, 47–60 (2023).
Stokel-Walker, C. & van Noorden, R. The promise and peril of generative AI. Nature 614, 214–216 (2023).
Coss, N. A. et al. Does clinical research account for diversity in deploying digital health technologies? NPJ Digit. Med. 6, 187 (2023).
Koinis, L., Mobbs, R. J., Fonseka, R. D. & Natarajan, P. A commentary on the potential of smartphones and other wearable devices to be used in the identification and monitoring of mental illness. Ann. Transl. Med. 10, 1420 (2022).
Muhammad Ali, N., Muhammad Ali, M. & Brohi, P. in The Smart Cyber Ecosystem for Sustainable Development (eds Kumar, P. et al.) 235–250 (Wiley, 2021).
Cheng, S. W. et al. The now and future of ChatGPT and GPT in psychiatry. Psychiatry Clin. Neurosci. 77, 592–596 (2023).
Koutsouleris, N., Hauser, T. U., Skvortsova, V. & De Choudhury, M. From promise to practice: towards the realisation of AI-informed mental health care. Lancet Digit. Health 4, e829–e840 (2022).
Hauser, T. U., Skvortsova, V., De Choudhury, M. & Koutsouleris, N. The promise of a model-based psychiatry: building computational models of mental ill health. Lancet Digit. Health 4, e816–e828 (2022).
Tordoff, D. M. et al. Mental health outcomes in transgender and nonbinary youths receiving gender-affirming care. JAMA Netw. Open 5, e220978 (2022).
Munn, Z. et al. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med. Res. Method 18, 143 (2018).
Tricco, A. C. et al. PRISMA extension for scoping reviews (PRISMA-ScR): checklist and explanation. Ann. Intern. Med. 169, 467–473 (2018).
Acknowledgements
We appreciate the assistance of L. Davordzi in this project, a graduate research assistant completing her Master of Science in Analytics (MSA).
Author information
Authors and Affiliations
Contributions
A.B. and Y.X. contributed to the design and implementation of the research, to the analysis of the data, to the development of final results and to the writing of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Mental Health thanks Debarshi Datta, Gayatri Kawlra and Emma Stanley for their contribution to the peer review of this work.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Baird, A., Xia, Y. Applying analytics to sociodemographic disparities in mental health. Nat. Mental Health 3, 124–138 (2025). https://doi.org/10.1038/s44220-024-00359-2
Received:
Accepted:
Published:
Version of record:
Issue date:
DOI: https://doi.org/10.1038/s44220-024-00359-2


