{Reference Type}: Journal Article {Title}: Transformers and large language models in healthcare: A review. {Author}: Nerella S;Bandyopadhyay S;Zhang J;Contreras M;Siegel S;Bumin A;Silva B;Sena J;Shickel B;Bihorac A;Khezeli K;Rashidi P; {Journal}: Artif Intell Med {Volume}: 154 {Issue}: 0 {Year}: 2024 Jun 5 {Factor}: 7.011 {DOI}: 10.1016/j.artmed.2024.102900 {Abstract}: With Artificial Intelligence (AI) increasingly permeating various aspects of society, including healthcare, the adoption of the Transformers neural network architecture is rapidly changing many applications. Transformer is a type of deep learning architecture initially developed to solve general-purpose Natural Language Processing (NLP) tasks and has subsequently been adapted in many fields, including healthcare. In this survey paper, we provide an overview of how this architecture has been adopted to analyze various forms of healthcare data, including clinical NLP, medical imaging, structured Electronic Health Records (EHR), social media, bio-physiological signals, biomolecular sequences. Furthermore, which have also include the articles that used the transformer architecture for generating surgical instructions and predicting adverse outcomes after surgeries under the umbrella of critical care. Under diverse settings, these models have been used for clinical diagnosis, report generation, data reconstruction, and drug/protein synthesis. Finally, we also discuss the benefits and limitations of using transformers in healthcare and examine issues such as computational cost, model interpretability, fairness, alignment with human values, ethical implications, and environmental impact.