{Reference Type}: Journal Article {Title}: Large language models in psychiatry: Opportunities and challenges. {Author}: Volkmer S;Meyer-Lindenberg A;Schwarz E; {Journal}: Psychiatry Res {Volume}: 339 {Issue}: 0 {Year}: 2024 Sep 11 {Factor}: 11.225 {DOI}: 10.1016/j.psychres.2024.116026 {Abstract}: The ability of Large Language Models (LLMs) to analyze and respond to freely written text is causing increasing excitement in the field of psychiatry; the application of such models presents unique opportunities and challenges for psychiatric applications. This review article seeks to offer a comprehensive overview of LLMs in psychiatry, their model architecture, potential use cases, and clinical considerations. LLM frameworks such as ChatGPT/GPT-4 are trained on huge amounts of text data that are sometimes fine-tuned for specific tasks. This opens up a wide range of possible psychiatric applications, such as accurately predicting individual patient risk factors for specific disorders, engaging in therapeutic intervention, and analyzing therapeutic material, to name a few. However, adoption in the psychiatric setting presents many challenges, including inherent limitations and biases in LLMs, concerns about explainability and privacy, and the potential damage resulting from produced misinformation. This review covers potential opportunities and limitations and highlights potential considerations when these models are applied in a real-world psychiatric context.