Abstract
The integration of large language models (LLMs) into mental healthcare and research heralds a potentially transformative shift, one offering enhanced access to care, efficient data collection, and innovative therapeutic tools. This paper reviews the development, function, and burgeoning use of LLMs in psychiatry, highlighting their potential to enhance mental healthcare through improved diagnostic accuracy, personalized care, and streamlined administrative processes. It is also acknowledged that LLMs introduce challenges related to computational demands, potential for misinterpretation, and ethical concerns, necessitating the development of pragmatic frameworks to ensure their safe deployment. We explore both the promise of LLMs in enriching psychiatric care and research through examples such as predictive analytics and therapy chatbots and risks including labor substitution, privacy concerns, and the necessity for responsible AI practices. We conclude by advocating for processes to develop responsible guardrails, including red-teaming, multi-stakeholder-oriented safety, and ethical guidelines/frameworks, to mitigate risks and harness the full potential of LLMs for advancing mental health.
Lay Summary
Imagine a conversation with an advanced computer program that could help doctors understand and treat mental health better. In this paper, we discuss using such advanced programs, called Large Language Models, in psychiatry. While they might offer quicker, tailored help and improve how mental health services work, there are many big questions left to answer about how to use this technology safely and ethically, to benefit patients without causing unintended harm.
Similar content being viewed by others
Introduction
New technologies can profoundly affect mental health research and psychiatric treatment [1]. For example, mobile computing devices and telehealth software offered unprecedented opportunities to improve access to care (e.g., interacting with a psychiatrist or psychotherapist via a smartphone), monitor clinical status (e.g., using a tablet to track personal outcomes), and collect clinically relevant data (e.g., allowing a smartwatch to record blood pressure to gauge stress levels). These tools offer advantages including convenience, anonymity, affordability, and the ability to reach underserved populations. These advantages also reduce the costs [2] and logistical challenges of traditional research methods. Currently, marked advances in the sophistication of large language models (LLMs) [3] may be poised to alter population mental health, mental health research, and mental healthcare practices. To understand the implications of this technological advance, we provide an overview of LLMs and describe some prominent applications for mental health. Additionally, we discuss the opportunities and risks LLMs might pose to the public’s mental well-being and to healthcare professionals’ efforts.
Function and use of LLMs
The term artificial intelligence (AI) denotes both (i) a scientific discipline studying how to build and understand computer systems that can complete endeavors deemed difficult because of the intelligence they seem to involve and (ii) those computer systems themselves [4]. As a scientific discipline, AI contains subfields that focus on the development of mechanical devices that can accomplish tasks for humans via movement, computing systems that can extract information from images, and software that can construe information in natural language for the purpose of devising further information for a user. As computer systems exhibit intelligence, AI possesses diverse abilities to solve problems, devise complex plans, integrate diverse technical information, and generate content (e.g., text, images, and audio). AI tools that generate content such as text and images have received attention over the past year due to the release of systems that make these tools easily accessible online (e.g., by February 2023, OpenAI’s suite of AI tools reached 100 million active users in 2 months) [5].
While the remarkable performance of the most advanced AI tools is quite novel, they rely on technical infrastructure and training processes—namely, artificial neural networks and machine learning—that have existed for decades [6,7,8,9] but rapidly progressed recently with the expansion of computing power that allows them to operate at substantially larger scale [10]. Artificial neural networks [8] are systems of weighted mathematical functions that slightly resemble neurons in the brain, ‘firing’—that is, producing an output—when the value of their inputs reaches some threshold value [8, 11, 12]. Machine learning constitutes a wide range of statistical methods [13], including those for estimating parameters and specifying connection structures in artificial neural networks [12]. Progress in these methods has occurred in recent decades [6,7,8,9,10,11,12,13,14], particularly with the development of simplified network architecture (i.e., transformers) capable of rapidly and efficiently conducting parallel processing of sequential data such as text [15].
From these methods, models with impressive language capacities have emerged (e.g. ChatGPT-4 from OpenAI [16]). Humans can hold advanced conversations with LLM partners via a standard interchange: a user supplies a ‘prompt’ to the system in the form of a written statement and the LLM generates an output in the form of the most probable completion of the prompt. Prompts typically consist of instructions, context, input data, and output indicators, although these categories can be arbitrary (e.g. as in few-shot learning). Once a user supplies a prompt as input, the system breaks the prompt’s words into fragments that are passed through the model’s neural network to produce an output that it returns to the user [12]. This output is the model’s estimate of the most probable continuation of the input that the user supplied.
Because of their myriad cross-domain applications and public accessibility, LLMs have rapidly entered widespread use. Their abilities now figure in tools ranging from online search assistants to social bots, from essay-writing tools to clinical therapeutic assistants.
Potential implications of LLMs for mental health
The application of LLMs as general-use technological tools has already yielded anticipated and unanticipated consequences [17]. LLMs notably excel in efficiently acquiring information, condensing content, and tackling reasoning-intensive problems [3]. In the context of mental healthcare, LLMs could conceivably quickly assimilate patient data, summarize therapy sessions, and aid in complex diagnostic problem-solving, thus potentially saving users and health systems both significant time and effort. Furthermore, when integrated into roles as personal assistants, LLMs can help individuals keep better track of their priorities and tasks, thus helping them achieve objectives in their daily lives. In this section, we discuss implications that include equitable access to the tools, manners in which LLMs may reshape mental healthcare systems, certain population mental health risks posed by the tools, and a framework for considering and evaluating such risks.
Equitable access to tailored LLMs
While ‘out-of-the-box’ LLMs might perform adroitly at mental healthcare-related tasks, they can also be further aligned to such tasks via approaches such as fine-tuning (i.e., trained specifically using additional data after its initial training), in-context-learning, or retrieval augmented generation, among other techniques to enhance task-specific performance (i.e., allowing longer prompts). Such models when fine-tuned for mental healthcare could, for example, offer a detailed list of therapeutic resources when asked how to manage a particular phobia. However, while the capacity to process information is increasing, processing such information at scale can be quite costly in terms of computing power and development resources, and therefore not all providers or institutions may benefit equally from the ability to fine-tune LLMs. This could become particularly problematic in mental health contexts, where accuracy in understanding the nuanced expressions of symptoms, emotions, and self-reported experiences is paramount.
Managerial and process-related considerations
LLMs have the potential to reshape health system managerial activity and work processes [18]. In a recent example of this possibility, Jiang et al. [19] demonstrated how an LLM trained on clinical notes could successfully predict patients’ readmission, in-hospital mortality, comorbidity index, length of stay, and insurance denial in the NYU Langone Health System. On its face, the findings relate solely to prognosis, but viewed more broadly the results provide grounds for “systems solutions”—redesigns of the hospital’s management and processes [20] that enhance decision-making via acting on the implications of AI predictions [18]. Operational decisions about personnel schedules and equipment needs could draw on the readmission and length of stay predictions reported in the paper—that is, a prediction that a wave of patients might require readmission or longer stays could activate increases in staffing levels and hospital material orders. Likewise, financial and accounting decisions could follow from bulk predictions of insurance denial: predictions of denial could galvanize steps to redirect funds from other parts of the organization or gather charitable resources to cover care a patient cannot afford. From simple predictions made by those five outcomes and additional training relating to the administrative implications of those outcomes, the LLM could automate managerial decisions within the hospital system, as research has discussed in other contexts [18]. Such possibilities portend a future in which not only clerical tasks become automated but also certain administrative ones.
Population mental health risks of LLMs
One significant risk associated with the general capacities of LLMs is that they not only make human labor more efficient, they can substitute for that labor [21]. If such labor substitution occurs rapidly and en masse, population mental health could experience declines mirroring those from prior economic downturns; clinical communities should prepare for such downturns. Another potential impact of LLMs on population mental health relates to the ease of developing conversational applications via the models. Due, perhaps, to the development of these tools for chat applications that require a warm tone, such as customer service, production LLMs typically exhibit a distinctive attribute: they provide pleasant interactions. On its face, this pleasant tone appears beneficial. Having an approachable AI with which to converse might seem to add a friend, albeit simulated, to one’s life and it might make day-to-day tasks more pleasing. But a fluid conversation partner that recalls detailed aspects of your life, coupled with a warm tone, runs the risk of sparking a feedback loop whereby users substitute human-to-human socialization with human-to-machine interaction. Innovative research studying the iterative rollout of Facebook across U.S. college campuses now provides reason to believe that social media may have produced unintended and costly consequences for individual well-being [22], and LLMs raise similar—if not greater—potential for such outcomes [23]. Moreover, a subset of adults can experience the onset or worsening of depression with greater social media use [24]. This risk grows because, much like social media, some AI chat applications may be optimized for user engagement, thus encouraging prolonged interaction. However, unlike social media, LLMs interact in ways that involve instantaneous conversation, the provision of information users sought for its value in professional or leisure purposes, and extensive opportunities for customization that can further enhance engagement.
It is important to recognize that the impact of LLMs on individual well-being can differ from person to person or may depend on how they are used, much like the effects of social media. Previous researchers not only have produced AI models with the explicit goal of enhancing users’ peer-to-peer empathy, but they have also demonstrated that the use of such models via human-AI collaboration actually can increase user empathy over time [25]. Consequentially, a substantial outcome involves the possibility of augmenting real-world or online human interactions with LLM interactions due to feelings of isolation or loneliness, which are currently at an all-time high [26]. For example, LLMs can be helpful for people looking to increase their social activities locally (e.g. via aiding internet searching for such options). On the converse, those suffering from social anxiety, social disconnection, or loneliness might be more inclined to use LLMs as a substitute for human-to-human socialization. These risks and benefits are ultimately empirical questions that require further study and consideration of (1) the situations in which LLMs are used, (2) the alternatives available to any given individual, (3) the principles guiding LLMs’ responses to queries, and (4) the degree of monitoring of the consequences of interactions with LLMs.
A framework for risk assessment
Scholars [27] have proposed a framework to better examine the broad social impact of LLMs, which includes grievous problems such as embedding biases, reinforcing stereotypes, violating privacy, and exacerbating inequalities, among other issues. LLMs – similar to other artificial intelligence and machine learning tools [28] – are subject to bias [29], both stemming from the content on which they are initially trained and the subsequent reinforcement input designed to shape their behavior. As such, practitioners should be cognizant of potential biases in the models that may manifest in unexpected ways in practice. Added to this are concerns about the potential for nonsensical or fanciful responses by LLMs to produce misinformation or confusion among users, particularly those users with pre-existing thought disorders. Given the complex determinants of population mental health, such broad-scale social implications [17] require holistic evaluation.
In the medical realm, there are many existing and proposed frameworks/guidelines for developing AI systems that include LLMs. However, none to our knowledge provide a sufficiently clear understanding of what safety parameters should be assessed and whose safety should be considered when determining whether an AI medical application, tool, or device is safe to launch [30,31,32,33]. Moreover, current frameworks/guidelines often lack a surveillance component to monitor and identify the long-term risks and benefits of AI medical systems. One approach that mitigates these concerns is the Biological-Psychological, Economic, and Social (BPES) framework [34]. This innovative framework applies aspects of the biopsychosocial model used in psychiatry while providing a descriptive approach that guides developers and regulatory agencies on how to assess whether an AI medical system is safe to launch (Fig. 1). More specifically, it asks developers to demonstrate the safety of their AI technology for multiple stakeholders (patients, healthcare professionals, industry, and government healthcare agencies) while using the BPES domains as parameters [34]. Another contribution of the BPES framework is its compatibility with and ability to provide context for all existing and future AI medical regulatory and non-regulatory frameworks/guidelines, including red-teaming approaches, that are concerned with human safety and AI product/system efficacy [34]. Although potentially useful, one aspect this model does not comprehensively address is the need for human-centered, iterative, participatory design prior to randomized trial conductions; such considerations could be incorporated in future iterations.
Adapted from the pharmaceutical industry’s multi-phase clinical drug trial model, the BPES Framework is a translational scientific approach that can be applied to any regulatory or non-regulatory guideline that assesses an AI-based medical system’s safety and efficacy, before it is launched to the public. Figure created with Biorender.com.
The BPES framework, or others like it, could also be nested within broader efforts to evaluate “Software as a Medical Device” (SaMD). For example, the International Medical Device Regulators Forum (IMDRF [35]) is a global group of medical device regulators have begun expanding their harmonizing efforts to include SaMD and has established a working group on artificial intelligence/machine learning [36] focused on promoting the development of safe and effective practices. However, in current practice, effective regulation remains an aspiration rather than a reality.
Opportunities of LLMs for psychiatric care and research
LLMs [37] hold marked potential for mental healthcare and research [38], primarily due to the central role language plays in the description, manifestation of, and treatment for mental health disorders. [39] A straightforward application of LLMs in mental healthcare might focus on the status assessment of an individual’s mental health or the use of verbal/language-based interventions to change this state. LLMs could thus potentially help to assess mental illness severity, suggest possible diagnoses, generate treatment plans, monitor the effect of an intervention, provide risk assessment indicators (e.g., for recurrence or relapse of a condition), and offer evidence-based suggestions for when an intervention is no longer needed. For instance, Galatzer-Levy et al. [40] examined whether an LLM without finetuning, yet “explicitly trained on large corpuses of medical knowledge,” could project an individual’s psychiatric functioning accurately based on patient and clinical documentation. The team found that, indeed, the LLM succeeded in this task: they could not reject the null hypothesis that the LLM performed the same as human clinical assessors [40]. Such findings underscore the fact that LLMs have demonstrated potential in interpreting verbal information to infer underlying affect, cognition, and diagnosis, providing a novel approach to assessing an individual’s mental health [41].
In another investigation of the clinical utility of therapy chatbots, a recent study described the development of LUMEN, a problem-solving treatment therapy coach that utilized the natural language understanding and natural language processing features of the Amazon Alexa platform. Participants were randomized to 8 sessions with LUMEN or a waitlist control arm with functional neuroimaging, and clinical and behavioral assessments conducted before and after treatment. This pilot study found that compared to the waitlist control group, participants who received LUMEN showed increases in task-evoked dorsolateral prefrontal cortex (DLFPC) activity that correlated with improvements in problem-solving skills and reductions in the avoidant coping styles. Furthermore, there were modest effects for reductions in depression and anxiety symptoms in the LUMEN group compared to the control group [42]. Although this study demonstrated evidence that the use of therapy chatbots might reduce symptoms of internalizing disorders, it was conducted in a small sample of participants with mild-to-moderate depression and/or anxiety. Further work is needed to determine whether such digital mental health interventions might be appropriate or effective for patients with greater clinical symptom severity, or at population-level scales.
These results highlight just how quickly the technology surrounding language models is evolving. For example, with minimal instruction and data from a problem-solving treatment manual, a version of LUMEN was created with ChatGPT-4 in under an hour (by O.A. and colleagues on November 28, 2023). Beyond the advantage of speed, another benefit of using an LLM is that users can interact with the chatbot more naturally and can personalize the style and tone of the conversational interface, to accomplish tasks that include identifying potential mental health conditions, determining the causes of those conditions, recognizing emotions in conversations, and understanding the relationship between environmental events and emotional responses [43]. However, a major disadvantage – aside from the limitation that the chatbot is derived from a specific region and culture – is the lack of control and constraints on how the LLM-derived chatbot can and will respond. Notably, the performance of ChatGPT has been found to be sensitive to different prompting strategies and emotional cues [44]. Thus, current models have limitations [45], including unstable predictions and sensitivity to minor alterations in prompts [44]. Future development in this domain should consider approaches to allow for more freeform user input while restricting chatbot output to ensure treatment fidelity.
These applications of LLMs in clinical settings are a direct consequence of the models’ powerful capacity for natural language processing (NLP). For example, LLM-based NLP models can detect medication information from clinical notes, generate and process psychiatric clinical narratives, and extract clinical concepts and associated semantic relationships from text [46, 47]. Such abilities to produce thematically accurate texts from automated transcriptions of clinical interactions may be particularly useful to the psychotherapeutic setting, potentially enabling clinicians to focus more attention on patients and less on real-time or post-hoc clinical note-taking. Further, LLM-driven evaluations of textual information are increasingly being applied to healthcare scenarios like perinatal self-harm, suicide attempts, HIV risk assessment, and delirium identification [43]. For example, LLMs have generated suicide risk predictions using health records data [48]. Others have demonstrated the cost-effectiveness and improved recovery rates associated with using a conversational AI solution for “referral, triage, and clinical assessment of mild-to-moderate adult mental illness” [49].
LLMs also have the potential to enhance psychotherapy by powering chatbots that provide evidence-based interventions. Such chatbots have demonstrated significant potential in the realm of mental health applications, providing a range of psychological interventions such as cognitive behavioral therapy [50], acceptance and commitment therapy [51], and positive psychology skills [52]. Clinicians have used them to assist patients with various conditions, including panic disorder [50], hypertension [53], and cancer [52]. They have even proven beneficial in postoperative recovery scenarios [51], and have also enhanced knowledge and skills related to health management [54]. Chatbots can deliver interventions at the user’s convenience, track progress, send reminders, and provide real-time feedback, perhaps enhancing adherence to treatment and self-management practices in the process. Moreover, they can potentially reduce healthcare costs by addressing minor health concerns that do not necessitate a doctor’s visit. However, research has raised concerns that these chatbots need better domain-specific training data, fine-tuning by expert clinicians, and targeting toward more “skilled” end-users [55]. Such chatbots also should be subject to efficacy trials to determine whether or not they actually improve clinical outcomes. Further, the degree to which both users and clinicians support the use of chatbots in clinical settings is worth additional scrutiny.
Another emerging application of LLM is clinical decision support, sometimes described more hyperbolically as “precision medicine”. Even LLMs trained on broad corpora (i.e., not specific to medicine) encode large amounts of medical knowledge, recognizing medications, indications, and adverse effects, and they can even demonstrate physician-level performance on board examinations, with the highest scores for psychiatry [56]. For example, one initial investigation applied ChatGPT-4 without fine-tuning or augmentation to suggest potential next-step antidepressant treatments [57] using a set of previously validated clinical vignettes. That study found that an optimal medication was identified ~3/4 of the time. On the other hand, in nearly half of the vignettes, the model also identified a potentially contraindicated or suboptimal treatment option.
More recently, a follow-up study examined the application of an LLM augmented with excerpts from clinical treatment guidelines for bipolar depression. Rather than applying a standard retrieval-augmented generation approach, this model simply incorporated key guideline principles in the prompt itself. This approach resulted in the selection of an expert-identified next-step treatment option twice as often as the unaugmented model and outperformed a small sample of community clinicians [58]. Notably, while the rate of selection of potentially contraindicated or less preferred options was diminished in the augmented model, such options appeared in approximately 1 in 10 vignettes, highlighting the need to apply such models with care. Prior work using clinical dashboards also underscored the capacity for psychiatric decision-support tools to lead clinicians astray [59].
LLMs may also prove to be valuable in formulating psychiatric research questions. Currently, researchers must analyze a growing volume of information to formulate hypotheses and discern the constellation of factors contributing to an individual’s mental health disorder for effective treatment research planning. In comparison, LLMs can rapidly and inexpensively generate—or provide feedback on—themes, questions, and proposed mechanisms at the level of disorders, while identifying possible factors underlying a particular individual’s mental health condition and formulating an individualized treatment plan [60]. Although this rapidly developing field is currently in its infancy, lessons learned from other areas of clinical science may be instructive. Thus, a phased approach for how to develop AI systems for use in mental health (e.g., one akin to the Phase I – IV process for evaluating new pharmacological agents) is likely to be of considerable value.
Risks of LLMs for psychiatric care and research
The application of LLMs in mental health research and care, while promising, is not without potential challenges [61]. First, the unpredictability and non-deterministic nature of some LLMs’ output causes concern in clinical settings. Like humans, these models can occasionally produce factually incorrect answers, make errors in simple arithmetic or logical problems, or generate responses that are seemingly obvious yet erroneous. And unlike humans, the models sometimes produce outputs that are outlandish and/or realistic yet fake (e.g., properly formatted citations to plausible-sounding journal articles that do not actually exist). The current state of research does not provide a clear methodology for guiding LLMs to minimize these inappropriate, false, or misleading outputs. Further compounding this challenge is the fact that LLMs by default do not have the capability to accurately attribute the sources underlying the information that they provide, and it depends on which layer of the foundation model [62] supply chain the clinical applications are built on top of. Such lack of attribution compounds the practical challenge of double-checking the information provided by any given LLM.
Second, scientific understanding of the mechanisms by which LLMs generate seemingly intelligent responses is still in its infancy [63]. For instance, if a user shares feelings of intense despair or mentions self-harm, an LLM that generates responses based on learned patterns rather than a real understanding of context or emotion might fail to accurately identify the urgency or respond appropriately. This lack of genuine empathy and understanding may perpetuate negative user experiences or intensify feelings of distress. Thus, while LLMs can aid mental health services, they must be used with robust monitoring systems and cannot replace professional intervention. Such monitoring systems may in turn require the oversight of clinical professionals, potentially providing additional roles and responsibilities for mental health clinicians.
Third, LLMs inherently express values in their communication, and those values might not comport with those of users. As a field, mental healthcare clinicians and researchers typically follow the principle of beneficence, which aims to promote human welfare and reduce suffering through knowledge and interventions. However, whether any particular LLM exhibits behaviors consistent with such values remains an empirical question [64].
Fourth, the potential for clinicians and researchers to become dependent on such systems is worth considering [65] as LLMs become widely adopted in the workplace. Even if such dependence does not materialize, it is likely that the requisite skills and practice of clinicians will change in the future as LLMs are integrated into clinical care settings. The manner and degree to which such change occurs is ultimately a question that will best be gauged via longitudinal study. As with any new technology or intervention, clinical trials will be invaluable in understanding where and how LLMs can be best deployed. The key elements of such trials include careful consideration of comparators, potential adverse effects, and sources of bias [66].
Finally, a suite of additional challenges – e.g., medical data imbalance and data shortage, potential data leakage and privacy implications, the need for carefully designed and evaluated prompts, the insufficient availability of open models trained across a variety of languages, the details of ethical implementation, and the difficulty of obtaining large, annotated corpora in specific medical fields – will need to be addressed in the process of incorporating LLMs into psychiatric care and research. Of specific concern are two additional facts. First, many of the health data that might be used to adapt LLMs to psychiatric contexts may fall under formal privacy protections (e.g. the Health Insurance Portability and Accountability Act, HIPAA), and ensuring such data are not subsequently revealed by the production LLM – a form of data leakage [67] – is an open problem. Second, those firms who already possess large stores of medical data such as large hospital systems and data brokerage firms that buy out data may have an inherent advantage in the use of these tools. Efforts to ensure equitable access are needed. Such challenges underscore the need for all parties engaged in mental healthcare to contemplate carefully how best to deploy LLMs to augment individual well-being.
Establishing responsible guardrails for the use LLMs in psychiatric care and research
In the context of AI deployment, the field of Responsible AI (RAI) has examined the ethical use of AI systems in various contexts, with many industry and academic researchers suggesting design principles and guidelines for others to follow. In practice, following these principles requires a significant amount of human oversight and investment to adhere to these principles. For instance, this includes impact assessments, a deep analysis of how AI affects stakeholders, and considering societal, cultural, and ethical implications, and it requires implementing human checks and controls to guide AI behavior, ensuring alignment with ethical standards and societal norms [68].
Perhaps the most critical activity of RAI human oversight is called ‘red-teaming’ [69]. The term, borrowed military strategy, describes the practice of adopting an adversarial approach to challenge and improve strategies, plans, and systems by identifying vulnerabilities and mitigating potential harms an opponent might exploit. This concept has been subsequently applied to technology, especially in cybersecurity. With the recent boom in generative AI technologies, AI red-teaming has gained considerable traction and attention as a method of ensuring AI safety, with AI companies jumping to hire professional and volunteer red-teamers (e.g., the most recent example being OpenAI’s announcement in February 2024 of Sora, a text-to-video engine that is currently available only to red-teamers).
The range of practices for AI red-teaming varies from open-ended to targeted and from manual to automated. For example, GPT-4 went through iterative rounds of red-teaming with the help of domain experts [70]. In another example, a Multi-round Automatic Red-Teaming (MART) method was used to enhance the scalability and safety of LLMs by iteratively fine-tuning a target LLM against automatically generated adversarial prompts, reducing safety violations by up to 85% without compromising performance on non-adversarial tasks [71].
Despite AI red-teaming efforts focused on computing and ethics, the omission of psychiatric-specific concerns represents a significant gap, hindering the responsible use of LLMs in psychiatric care and research. For psychiatric applications of red-teaming, key considerations include scoping intended application scenarios, selecting a diverse group for testing across various demographics and experiences, defining the technical scope of what will be tested (e.g., public vs. proprietary models, safety layers, or user interfaces), and deciding on the testing methodology, whether it be organized teams or scattered volunteers, and open or targeted tasks. In developing guardrails and testing the manners in which psychiatric applications of LLMs may present challenges, the role of red-teaming for LLM-based applications in psychiatry should not be overlooked. AI in mental health must ethically align with therapeutic goals while balancing efficiency and sensitivity, respecting patient autonomy and confidentiality, and avoiding the reinforcement of stigmas. AI testing in mental health should also evaluate its impact on psychological well-being and therapeutic outcomes, focusing on data sensitivity, patient privacy, cultural adaptability, and the potential for emotional harm. There are emerging examples of efforts to establish AI safety benchmarks [72, 73], but it is worth emphasizing that implementing AI in mental health demands precision and human accountability and requires systems that adapt to individual needs and include robust mechanisms for clinician oversight to support, not replace, professional judgment.
The challenges we outline above thus necessitate the active engagement of domain experts in psychiatry to make the red-teaming of AI systems more robust, clinically sound, and safe for mental health. Such experts should include not only clinicians, but individuals with lived experience, family members, and policymakers – all of whom will be impacted by the success or failure of these models.
Conclusion
The use of LLMs in healthcare [19], mental health, and psychiatric research remains in its infancy. These models have shown promising potential for early detection, treatment, prediction, and scientific evaluation of mental health conditions. Increasing the reliability and predictability of LLMs can enhance clinician-researcher trust and understanding, thus facilitating the successful integration of risk models into clinical care and psychiatric research. The advent of LLMs will alter the landscape of public mental health, via direct social effects likely to stem from the technology as well as via the integration of LLMs into mental healthcare and research [60]. LLMs hold significant promise for mental healthcare and research, presenting opportunities to expedite diagnostic processes, enhance patient care, and contribute to research efforts in psychiatric science. Simultaneously, the use of LLMs for personal assistance can have far-reaching implications for population mental health, both positive and negative.
Nevertheless, the deployment of these advanced AI models will bring challenges, including ethical concerns [74]. Their limitations, such as the unpredictability of outputs in some models, lack of understanding of response generation mechanisms (i.e., lack of transparency), potential for inappropriate or false outputs, and possible mismatch of values and introduction of bias, present significant hurdles. Going forward, the next steps should involve rigorous interdisciplinary dialogue and research into understanding and overcoming these challenges via the establishment of responsible guardrails. Additionally, exploring ways to fine-tune LLMs with the guidance of clinicians and to expand their accessibility to a wider range of users is essential. Finally, the profound changes LLMs might bring justify efforts that study the system-wide effects of LLMs on population mental health, including socioeconomic consequences that might subsequently alter people’s well-being. It is through these concerted efforts that we can maximize the potential benefits of LLMs in mental healthcare and research while minimizing their associated risks.
Citation diversity statement
The authors have attested that they made efforts to be mindful of diversity in selecting the citations used in this article.
References
Haidt J, Allen N. Scrutinizing the effects of digital technology on mental health. Nature. 2020;578:167–9. https://doi.org/10.1038/d41586-020-00296-x
Gega L, Jankovic D, Saramago P, Marshall D, Dawson S, Brabyn S, et al. Digital interventions in mental health: evidence syntheses and economic modelling. Health Technol Assess. 2022;26:1–182. https://doi.org/10.3310/RCTI6942
Bubeck, S, Chandrasekaran, V, Eldan, R, Gehrke, J, Horvitz, E, Kamar, E, et al. Sparks of Artificial General Intelligence: Early experiments with GPT-4. 2023. https://doi.org/10.48550/arXiv.2303.12712, https://arxiv.org/abs/2303.12712
Russell SJ, Norvig P. Artificial intelligence: a modern approach. 4th Edition ed. Pearson Series in Artificial Intelligence. Pearson; 2021.
https://www.reuters.com/technology/chatgpt-sets-record-fastest-growing-user-base-analyst-note-2023-02-01/ Accessed 7/11/2023
Schmidhuber J. Annotated History of Modern AI and Deep Learning. 2022:75. Technical Report IDSIA-22-22 (v2). https://arxiv.org/ftp/arxiv/papers/2212/2212.11279.pdf
Taroni A. 90 years of the Ising model. Nat Phys. 2015;11:997–997. https://doi.org/10.1038/nphys3595
McCulloch WS, Pitts W. A logical calculus of the ideas immanent in nervous activity. The bulletin of mathematical biophysics. 1943/12/01 1943;5:115-33. https://doi.org/10.1007/BF02478259
Amari SI. Learning patterns and pattern sequences by self-organizing nets of threshold elements. IEEE Trans Comput. 1972;C-21:1197–206. https://doi.org/10.1109/T-C.1972.223477
LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436–44. https://doi.org/10.1038/nature14539
Krogh A. What are artificial neural networks? Nat Biotechnol. 2008;26:195–7. https://doi.org/10.1038/nbt1386
Wolfram S. What Is ChatGPT Doing… and Why Does It Work? Stephen Wolfram; 2023.
Greener JG, Kandathil SM, Moffat L, Jones DT. A guide to machine learning for biologists. Nat Rev Mol Cell Biol. 2022;23:40–55. https://doi.org/10.1038/s41580-021-00407-0
Kaplan, J, McCandlish, S, Henighan, T, Brown, TB, Chess, B, Child, R, et al. Scaling laws for neural language models. arXiv. 2020:cs.LG. 1/23/ 2020. https://doi.org/10.48550/arXiv.2001.08361, https://arxiv.org/abs/2001.08361
Vaswani, A, Shazeer, N, Parmar, N, Uszkoreit, J, Jones, L, Gomez, AN, et al. Attention is all you need. arXiv. 2017. https://doi.org/10.48550/arXiv.1706.03762
OpenAi. GPT-4 Technical Report. arXiv. 2023. 3/27/2023. https://doi.org/10.48550/arXiv.2303.08774
Rahwan I, Cebrian M, Obradovich N, Bongard J, Bonnefon JF, Breazeal C, et al. Machine behaviour. Nature. 2019;568:477–86. https://doi.org/10.1038/s41586-019-1138-y
Agrawal A, Gans J, Goldfarb A. Power and prediction: the disruptive economics of artificial intelligence. Harvard Business Review Press; 2022.
Jiang LY, Liu XC, Nejatian NP, Nasir-Moin M, Wang D, Abidin A, et al. Health system-scale language models are all-purpose prediction engines. Nature. 2023;619:357–62. https://doi.org/10.1038/s41586-023-06160-y
Obradovich N, Johnson T, Paulus MP. Managerial and Organizational Challenges in the Age of AI. JAMA Psychiatry. 2024;81:219–20. https://doi.org/10.1001/jamapsychiatry.2023.5247
Eloundou T, Manning S, Mishkin P, Rock D GPTs are GPTs: An early look at the labor market impact potential of large language models. eprint. 2023. 3/23/2023. https://doi.org/10.48550/arXiv.2303.10130, http://arxiv.org/abs/2303.10130
Huang C. A meta-analysis of the problematic social media use and mental health. Int J Soc Psychiatry. 2022;68:12–33. https://doi.org/10.1177/0020764020978434
Braghieri L, Levy RE, Makarin A. Social media and mental health. Am Econ. Rev. 2022;112:3660–93. https://doi.org/10.1257/aer.20211218
Perlis RH, Green J, Simonson M, Ognyanova K, Santillana M, Lin J, et al. Association Between Social Media Use and Self-reported Symptoms of Depression in US Adults. JAMA Netw Open. 2021;4:e2136113. https://doi.org/10.1001/jamanetworkopen.2021.36113
Sharma A, Lin IW, Miner AS, Atkins DC, Althoff T. Human–AI collaboration enables more empathic conversations in text-based peer-to-peer mental health support. Nat Mach Intell. 2023;5:45–57. https://doi.org/10.1038/s42256-022-00593-2
Our Epidemic of Loneliness and Isolation: The US Surgeon General’s Advisory on the Healing Effects of Social Connection and Community. 2023. Publications and Reports of the Surgeon General.
Solaiman, I, Talat, Z, Agnew, W, Ahmad, L, Baker, D, Blodgett, SL, et al. Evaluating the Social Impact of Generative AI Systems in Systems and Society. arXiv. 2023:cs.CY. https://doi.org/10.48550/arXiv.2306.05949
Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A. A survey on bias and fairness in machine learning. ACM Comput Surv (CSUR). 2021;54:1–35.
Parikh RB, Teeple S, Navathe AS. Addressing bias in artificial intelligence in health care. JAMA. 2019;322:2377–8.
Mesko B, Topol EJ. The imperative for regulatory oversight of large language models (or generative AI) in healthcare. NPJ Digit Med. 2023;6:120. https://doi.org/10.1038/s41746-023-00873-0
Administration TUSFaD. Artificial Intelligence and Machine Learning (AI/ML)-Enabled Medical Devices. The United States Food and Drug Administration. Accessed December 16, 2023. https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices
Union TE. EU AI Act: first regulation on artificial intelligence. The European Union. Accessed December 16th, 2023. https://www.europarl.europa.eu/news/en/headlines/society/20230601STO93804/eu-ai-act-first-regulation-on-artificial-intelligence
Crossnohere NL, Elsaid M, Paskett J, Bose-Brill S, Bridges JFP. Guidelines for Artificial Intelligence in Medicine: Literature Review and Content Analysis of Frameworks. J Med Internet Res. 2022;24:e36823. https://doi.org/10.2196/36823
Khan WU, Seto E. A “Do No Harm” novel safety checklist and research approach to determine whether to launch an artificial intelligence-based medical technology: introducing the Biological-Psychological, Economic, and Social (BPES) framework. J Med Internet Res. 2023;25:e43386. https://doi.org/10.2196/43386
International Medical Device Regulators Forum. https://www.imdrf.org/
Artificial Intelligence/Machine Learning-enabled. https://www.imdrf.org/working-groups/artificial-intelligencemachine-learning-enabled
Zhao, WX, Zhou, K, Li, J, Tang, T, Wang, X, Hou, Y, et al. A Survey of Large Language Models. 2023. https://doi.org/10.48550/arXiv.2303.18223, https://arxiv.org/abs/2303.18223
Lee P, Bubeck S, Petro J. Benefits, limits, and risks of GPT-4 as an AI chatbot for medicine. N. Engl J Med. 2023;388:1233–9. https://doi.org/10.1056/NEJMsr2214184
Volkow ND, Gordon JA, Koob GF. Choosing appropriate language to reduce the stigma around mental illness and substance use disorders. Neuropsychopharmacology. 2021;46:2230–2. https://doi.org/10.1038/s41386-021-01069-4
Galatzer-Levy IR, McDuff D, Natarajan V, Karthikesalingam A, Malgaroli M. The Capability of Large Language Models to Measure Psychiatric Functioning. arXiv. 2023; https://doi.org/10.48550/arXiv.2308.01834
Lamichhane B. Evaluation of ChatGPT for NLP-based Mental Health Applications. 2023. https://doi.org/10.48550/arXiv.2303.15727, https://arxiv.org/abs/2303.15727
Kannampallil T, Ajilore OA, Lv N, Smyth JM, Wittels NE, Ronneberg CR, et al. Effects of a virtual voice-based coach delivering problem-solving treatment on emotional distress and brain function: a pilot RCT in depression and anxiety. Transl Psychiatry. 2023;13:166. https://doi.org/10.1038/s41398-023-02462-x
Hossain E, Rana R, Higgins N, Soar J, Barua PD, Pisani AR, et al. Natural Language Processing in Electronic Health Records in relation to healthcare decision-making: A systematic review. Comput Biol Med. 2023;155:106649. https://doi.org/10.1016/j.compbiomed.2023.106649
Yang K, Ji S, Zhang T, Xie Q, Kuang Z, Ananiadou S. Towards Interpretable Mental Health Analysis with ChatGPT. 2023. https://doi.org/10.48550/arXiv.2304.03347, https://arxiv.org/abs/2304.03347
Amin MM, Cambria E, Schuller BW. Will Affective Computing Emerge from Foundation Models and General AI? A First Evaluation on ChatGPT. 2023. https://doi.org/10.48550/arXiv.2303.03186, https://arxiv.org/abs/2303.03186
Peng C, Yang X, Yu Z, Bian J, Hogan WR, Wu Y. Clinical concept and relation extraction using prompt-based machine reading comprehension. Journal of the American Medical Informatics Association : JAMIA. Jun 14 2023; https://doi.org/10.1093/jamia/ocad107
Chen A, Yu Z, Yang X, Guo Y, Bian J, Wu Y. Contextualized medication information extraction using Transformer-based deep learning architectures. J Biomed Inform. 2023;142:104370 https://doi.org/10.1016/j.jbi.2023.104370
Shortreed SM, Walker RL, Johnson E, Wellman R, Cruz M, Ziebell R, et al. Complex modeling with detailed temporal predictors does not improve health records-based suicide risk prediction. NPJ Digit Med. 2023;6:47. https://doi.org/10.1038/s41746-023-00772-4.23
Rollwage M, Juchems K, Habicht J, Carrington B, Hauser T, Harper R. Conversational AI facilitates mental health assessments and is associated with improved recovery rates. medRxiv. 2022:2022.11.03.22281887. https://doi.org/10.1101/2022.11.03.22281887
Oh J, Jang S, Kim H, Kim JJ. Efficacy of mobile app-based interactive cognitive behavioral therapy using a chatbot for panic disorder. Int J Med Inf. 2020;140:104171. https://doi.org/10.1016/j.ijmedinf.2020.104171
Anthony CA, Rojas EO, Keffala V, Glass NA, Shah AS, Miller BJ, et al. Acceptance and commitment therapy delivered via a mobile phone messaging robot to decrease postoperative opioid use in patients with orthopedic trauma: randomized controlled trial. J Med Internet Res. 2020;22:e17750. https://doi.org/10.2196/17750
Greer S, Ramo D, Chang YJ, Fu M, Moskowitz J, Haritatos J. Use of the Chatbot “Vivibot” to deliver positive psychology skills and promote well-being among young people after cancer treatment: randomized controlled feasibility trial. JMIR mHealth uHealth. 2019;7:e15018. https://doi.org/10.2196/15018
Echeazarra L, Pereira J, Saracho R. TensioBot: a chatbot assistant for self-managed in-house blood pressure checking. J Med Syst. 2021;45:54. https://doi.org/10.1007/s10916-021-01730-x
Maeda E, Miyata A, Boivin J, Nomura K, Kumazawa Y, Shirasawa H, et al. Promoting fertility awareness and preconception health using a chatbot: a randomized controlled trial. Reprod Biomed Online. 2020;41:1133–43. https://doi.org/10.1016/j.rbmo.2020.09.006
Au Yeung J, Kraljevic Z, Luintel A, Balston A, Idowu E, Dobson RJ, et al. AI chatbots not yet ready for clinical use. Front Digit Health. 2023;5:1161098. https://doi.org/10.3389/fdgth.2023.1161098
Katz, U, Cohen, E, Shachar, E, Somer, J, Fink, A, Morse, E, et al. GPT versus Resident Physicians — A Benchmark Based on Official Board Scores. NEJM AI. 2024; https://doi.org/10.1056/AIdbp2300192
Perlis RH. Research Letter: Application of GPT-4 to select next-step antidepressant treatment in major depression. medRxiv. 2023; https://doi.org/10.1101/2023.04.14.23288595
Perlis RH, Goldberg JF, Ostacher MJ, Schneck CD. Clinical decision support for bipolar depression using large language models. Neuropsychopharmacology 2024; https://doi.org/10.1038/s41386-024-01841-2
Jacobs M, Pradier MF, McCoy TH,Jr, Perlis RH, Doshi-Velez F, Gajos KZ. How machine-learning recommendations influence clinician treatment selections: the example of the antidepressant selection. Transl Psychiatry. 2021;11:108. https://doi.org/10.1038/s41398-021-01224-x
van Heerden AC, Pozuelo JR, Kohrt BA. Global Mental Health Services and the Impact of Artificial Intelligence-Powered Large Language Models. JAMA Psychiatry. 2023; https://doi.org/10.1001/jamapsychiatry.2023.1253
Bowman SR. Eight Things to Know about Large Language Models. 2023. https://doi.org/10.48550/arXiv.2304.00612, https://arxiv.org/abs/2304.00612
Jones E. Explainer: What is a foundation model? Accessed April 26, 2024. https://www.adalovelaceinstitute.org/resource/foundation-models-explainer/
Kaddour J, Harris J, Mozes M, Bradley H, Raileanu R, McHardy R. Challenges and Applications of Large Language Models. eprint. 2023. 7/19/2023. https://doi.org/10.48550/arXiv.2307.10169, http://arxiv.org/abs/2307.10169
Johnson T, Obradovich N. Evidence of behavior consistent with self-interest and altruism in an artificially intelligent agent. arXiv. 2023:cs.AI. 1/5/2023. https://doi.org/10.48550/arXiv.2301.02330, https://arxiv.org/abs/2301.02330
Passi S, Vorvoreanu M. Overreliance on AI: Literature review. AETHER: AI Ethics and Effects in Engineering and Research. 2022:1–23.
Perlis RH, Fihn SD. Evaluating the Application of Large Language Models in Clinical Research Contexts. JAMA Netw Open. 2023;6:e2335924. https://doi.org/10.1001/jamanetworkopen.2023.35924
Carlini, N, Tramer, F, Wallace, E, Jagielski, M, Herbert-Voss, A, Lee, K, et al. Extracting training data from large language models. 30th USENIX Security Symposium (USENIX Security 21). 2021:2633-50.
Goldberg, CB, Adams, L, Blumenthal, D, Brennan, PF, Brown, N, Butte, AJ, et al. To do no harm - and the most good - with AI in health care. Nat Med. 2024; https://doi.org/10.1038/s41591-024-02853-7
Ganguli, D, Lovitt, L, Kernion, J, Askell, A, Bai, Y, Kadavath, S, et al. Red Teaming Language Models to Reduce Harms: Methods, Scaling Behaviors, and Lessons Learned. arXiv. 2022. https://doi.org/10.48550/arXiv.2209.07858
OpenAI Achiam J, Adler S, Agarwal S, Ahmad L, Akkaya I, Leoni Aleman F, et al GPT-4 Technical Report. arXiv. 2023. https://doi.org/10.48550/arXiv.2303.08774
Ge S, Zhou C, Hou R, Khabsa M, Wang YC, Wang Q, et al. MART: Improving LLM Safety with Multi-round Automatic Red-Teaming. arXiv. 2023. http://arxiv.org/abs/2311.07689
Vidgen, B, Agrawal, A, Ahmed, AM, Akinwande, V, Al-Nuaimi, N, Alfaraj, N, et al. Introducing v0.5 of the AI Safety Benchmark from MLCommons. arXiv. April 18, 2024. https://doi.org/10.48550/arXiv.2404.12241
Gabriel, I, Manzini, A, Keeling, G, Hendricks, LA, Rieser, V, Iqbal, H, et al The Ethics of Advanced AI Assistants. Accessed April 22, 2024. https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/ethics-of-advanced-ai-assistants/the-ethics-of-advanced-ai-assistants-2024-i.pdf
Birhane A, Kasirzadeh A, Leslie D, Wachter S. Science in the age of large language models. Nat Rev Phys. 2023;5:277–80. https://doi.org/10.1038/s42254-023-00581-4
Acknowledgements
N.O., M.P., and S.K. thank Tim Johnson for his extensive contributions to the initial draft of the manuscript and thank members of the Laureate Institute for Brain Research’s Large Language Models and Mental Health Discussion Group for useful conversations.
Funding
This work was partly funded by The William K. Warren Foundation, the National Institute of General Medical Sciences Center (Grant 2 P20 GM121312), the National Institute on Drug Abuse (U01DA050989), and the National Institute for Mental Health (R01MH127225; R01MH123804).
Author information
Authors and Affiliations
Contributions
NO, SK, MP: Substantial contributions to the conception of the work. NO, SK, WK, JS, RP, OA, MP: (1) Drafting and revising the work critically for important intellectual content, (2) Final approval of the version to be published, and (3) Agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.
Corresponding author
Ethics declarations
Competing interests
Dr. Obradovich is part of OpenAI’s Researcher Access Program and has received resources from OpenAI to support research activities outside of this work. Dr. Khalsa provides uncompensated service on executive boards of the International Society for Contemplative Research and the Float Research Collective, serves on the editorial boards of Biological Psychology and JMIR Mental Health, and has served on a scientific advisory board for Janssen Pharmaceuticals. Dr. Khan has no conflicts of interest or disclosures to make that are relevant to the content. Dr. Suh is employed by Microsoft Research. Dr. Perlis has served on scientific advisory boards for Circular Genomics, Genomind, Belle Artificial Intelligence, Swan AI Studios, and Psy Therapeutics. Dr. Ajilore is the co-founder of KeyWise AI and has served as a consultant for Otsuka Pharmaceuticals on the advisory boards for Sage Therapeutics, Embodied Labs, and Blueprint Health. Dr. Paulus advises Spring Care, Inc., receives royalties from an article on methamphetamine in UpToDate, and has a compensated consulting agreement with Boehringer Ingelheim International GmbH.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Obradovich, N., Khalsa, S.S., Khan, W.U. et al. Opportunities and risks of large language models in psychiatry. NPP—Digit Psychiatry Neurosci 2, 8 (2024). https://doi.org/10.1038/s44277-024-00010-z
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1038/s44277-024-00010-z
This article is cited by
-
Racial bias in AI-mediated psychiatric diagnosis and treatment: a qualitative comparison of four large language models
npj Digital Medicine (2025)
-
Navigating promise and perils: applying artificial intelligence to the perinatal mental health care cascade
npj Health Systems (2025)
-
Language models in digital psychiatry: challenges with simplification of healthcare materials
NPP—Digital Psychiatry and Neuroscience (2025)
-
The role of prompt, voice, and personality factors in the acceptance and evaluation of AI-generated mindfulness exercises
Scientific Reports (2025)
-
Exploring the potential of ChatGPT as a digital advisor in acute psychiatric crises: a feasibility study
Der Nervenarzt (2025)