Abstract
Clinical abbreviation disambiguation is a crucial task in the biomedical domain, as the accurate identification of the intended meanings or expansions of abbreviations in clinical texts is vital for medical information retrieval and analysis. Existing approaches have shown promising results, but challenges such as limited instances and ambiguous interpretations persist. In this paper, we propose an approach to address these challenges and enhance the performance of clinical abbreviation disambiguation. Our objective is to leverage the power of Large Language Models (LLMs) and employ a Generative Model (GM) to augment the dataset with contextually relevant instances, enabling more accurate disambiguation across diverse clinical contexts. We integrate the contextual understanding of LLMs, represented by BlueBERT and Transformers, with data augmentation using a Generative Model, called Biomedical Generative Pre-trained Transformer (BIOGPT), that is pretrained on an extensive corpus of biomedical literature to capture the intricacies of medical terminology and context. By providing the BIOGPT with relevant medical terms and sense information, we generate diverse instances of clinical text that accurately represent the intended meanings of abbreviations. We evaluate our approach on the widely recognized CASI dataset, carefully partitioned into training, validation, and test sets. The incorporation of data augmentation with the GM improves the model’s performance, particularly for senses with limited instances, effectively addressing dataset imbalance and challenges posed by similar concepts. The results demonstrate the efficacy of our proposed method, showcasing the significance of LLMs and generative techniques in clinical abbreviation disambiguation. Our model achieves a good accuracy on the test set, outperforming previous methods.
Similar content being viewed by others
Data Availability
The datasets analyzed during the current study are available in the Clinical Abbreviation Sense Inventory repository, https://conservancy.umn.edu/handle/11299/137703.
References
B. Duganová, “Medical language – a unique linguistic phenomenon,” JAHR, 2019.
C. DesRoches, “Healthcare in the new age of transparency,” Semin Dial, vol. 33, Jul. 2020, doi: https://doi.org/10.1111/sdi.12934
L. Brunetti, J. Santell, and R. Hicks, “The Impact of Abbreviations on Patient Safety,” Joint Commission journal on quality and patient safety / Joint Commission Resources, vol. 33, pp. 576–583, Jul. 2007, doi: https://doi.org/10.1016/S1553-7250(07)33062-6
“Shared Learning-Reported Incidents Involving Hydromorphone,” 2006. [Online]. Available: http://www.ismp-canada.org/download/CMIRPS%20Core%20Data%20Set%20for%20Individu
A. Coghlan, S. Turner, and S. Coverdale, “Danger in discharge summaries: Abbreviations create confusion for both author and recipient,” Intern Med J, vol. 53, Jul. 2021, doi: https://doi.org/10.1111/imj.15582
K. E. Walsh and J. H. Gurwitz, “Medical abbreviations: writing little and communicating less.,” Arch Dis Child, vol. 93, no. 10, pp. 816–817, Oct. 2008, doi: https://doi.org/10.1136/adc.2008.141473
H. Xu, P. D. Stetson, and C. Friedman, “A study of abbreviations in clinical notes.,” AMIA Annu Symp Proc, vol. 2007, pp. 821–825, Oct. 2007.
B. Collard and A. Royal, “The use of abbreviations in surgical note keeping,” Annals of Medicine and Surgery, vol. 4, no. 2, pp. 100–102, 2015, doi: https://doi.org/10.1016/j.amsu.2015.03.008
I. F. Kuhn, “Abbreviations and acronyms in healthcare: when shorter isn’t sweeter.,” Pediatr Nurs, vol. 33, no. 5, pp. 392–398, 2007.
T. I. Amosa, L. I. bt Izhar, P. Sebastian, I. B. Ismail, O. Ibrahim, and S. L. Ayinla, “Clinical Errors from Acronym Use in Electronic Health Record: A Review of NLP-based Disambiguation Techniques,” IEEE Access, 2023, doi: https://doi.org/10.1109/ACCESS.2023.3284682
Y. Peng, S. Yan, and Z. lu, “Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets,” Jul. 2019, pp. 58–65. doi: https://doi.org/10.18653/v1/W19-5006
R. Luo et al, “BioGPT: generative pre-trained transformer for biomedical text generation and mining,” Brief Bioinform, vol. 23, no. 6, p. bbac409, Nov. 2022, doi: https://doi.org/10.1093/bib/bbac409
A. Schwartz and M. Hearst, “A Simple Algorithm For Identifying Abbreviation Definitions in Biomedical Text,” Pac Symp Biocomput, vol. 4, pp. 451–462, Feb. 2003, doi: https://doi.org/10.1142/9789812776303_0042
J. Toole, “A Hybrid Approach to the Identification and Expansion of Abbreviations,” May 2000.
V. García, J. Sánchez, and R. Mollineda, “Exploring the Performance of Resampling Strategies for the Class Imbalance Problem,” 2010, pp. 541–549. doi: https://doi.org/10.1007/978-3-642-13022-9_54
R. Navigli, S. Faralli, A. Soroa, O. de Lacalle, and E. Agirre, “Two Birds with One Stone: Learning Semantic Models for Text Categorization and Word Sense Disambiguation,” in Proceedings of the 20th ACM International Conference on Information and Knowledge Management, in CIKM ’11. New York, NY, USA: Association for Computing Machinery, 2011, pp. 2317–2320. doi: https://doi.org/10.1145/2063576.2063955
S. Pakhomov, T. Pedersen, and C. G. Chute, “Abbreviation and acronym disambiguation in clinical discourse.,” AMIA Annu Symp Proc, vol. 2005, pp. 589–593, 2005.
S. Moon, S. Pakhomov, and G. B. Melton, “Automated disambiguation of acronyms and abbreviations in clinical texts: window and training size considerations.,” AMIA Annu Symp Proc, vol. 2012, pp. 1310–1319, 2012.
M. Joshi, S. Pakhomov, T. Pedersen, and C. G. Chute, “A comparative study of supervised learning as applied to acronym expansion in clinical reports.,” AMIA Annu Symp Proc, vol. 2006, pp. 399–403, 2006.
G. P. Finley, S. V. S. Pakhomov, R. McEwan, and G. B. Melton, “Towards Comprehensive Clinical Abbreviation Disambiguation Using Machine-Labeled Training Data.,” AMIA Annu Symp Proc, vol. 2016, pp. 560–569, 2016.
Y. Wu, J. Xu, Y. Zhang, and W. Qi, Clinical Abbreviation Disambiguation Using Neural Word Embeddings. 2015. doi: https://doi.org/10.18653/v1/W15-3822
Y. Wu et al, “A long journey to short abbreviations: developing an open-source framework for clinical abbreviation recognition and disambiguation (CARD).,” J Am Med Inform Assoc, vol. 24, no. e1, pp. e79–e86, Apr. 2017, doi: https://doi.org/10.1093/jamia/ocw109
A. Jaber and P. Mart\’\inez, “Disambiguating Clinical Abbreviations using Pre-trained Word Embeddings,” in Proceedings of the 14th International Joint Conference on Biomedical Engineering Systems and Technologies, {SCITEPRESS} - Science and Technology Publications, 2021. doi: https://doi.org/10.5220/0010256105010508
R. Socher and C. Manning, “Deep Learning for NLP (without Magic).”
V. Kumar, D. Reforgiato Recupero, D. Riboni, and R. Helaoui, “Ensembling Classical Machine Learning and Deep Learning Approaches for Morbidity Identification From Clinical Notes,” IEEE Access, vol. PP, p. 1, Dec. 2020, doi: https://doi.org/10.1109/ACCESS.2020.3043221
V. Joopudi, B. Dandala, and M. Devarakonda, “A convolutional route to abbreviation disambiguation in clinical text,” J Biomed Inform, vol. 86, pp. 71–78, 2018, doi: https://doi.org/10.1016/j.jbi.2018.07.025
M. Skreta et al, “Automatically disambiguating medical acronyms with ontology-aware deep learning,” Nat Commun, vol. 12, no. 1, p. 5319, 2021, doi: https://doi.org/10.1038/s41467-021-25578-4
C.-X. Zhang, M.-L. Wang, and X.-Y. Gao, “Biomedical Word Sense Disambiguation Based on Graph Attention Networks,” IEEE Access, vol. 10, pp. 123328–123336, 2022, doi: https://doi.org/10.1109/ACCESS.2022.3224802
N. B. Link et al, “Binary acronym disambiguation in clinical notes from electronic health records with an application in computational phenotyping,” Int J Med Inform, vol. 162, p. 104753, 2022, doi: https://doi.org/10.1016/j.ijmedinf.2022.104753
A. Rajkomar et al, “Deciphering clinical abbreviations with a privacy protecting machine learning system,” Nat Commun, vol. 13, no. 1, Dec. 2022, doi: https://doi.org/10.1038/s41467-022-35007-9
J. Lee et al, “BioBERT: a pre-trained biomedical language representation model for biomedical text mining,” Bioinformatics, vol. 36, no. 4, pp. 1234–1240, Feb. 2020, doi: https://doi.org/10.1093/bioinformatics/btz682
J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,” in NAACL, 2019.
S. Kula, R. Kozik, and M. Choraś, “Implementation of the BERT-derived architectures to tackle disinformation challenges,” Neural Comput Appl, vol. 34, no. 23, pp. 20449–20461, 2022, doi: https://doi.org/10.1007/s00521-021-06276-0
S. Hao, P. Zhang, S. Liu, and Y. Wang, “Sentiment recognition and analysis method of official document text based on BERT–SVM model,” Neural Comput Appl, 2023, doi: https://doi.org/10.1007/s00521-023-08226-4
A. Kashyap, H. Burris, C. Callison-Burch, and M. R. Boland, “The CLASSE GATOR (CLinical Acronym SenSE disambiGuATOR): A Method for predicting acronym sense from neonatal clinical notes,” Int J Med Inform, vol. 137, p. 104101, 2020, doi: https://doi.org/10.1016/j.ijmedinf.2020.104101
A. Jaber and P. Martínez, “Disambiguating Clinical Abbreviations Using a One-Fits-All Classifier Based on Deep Learning Techniques,” Methods Inf Med, vol. 61, pp. E28–E34, Jun. 2022, doi: https://doi.org/10.1055/s-0042-1742388
P. Kacker, A. Cupallari, A. G. Subramanian, and N. Jain, “ABB-BERT: A BERT model for disambiguating abbreviations and contractions,” Jul. 2022, [Online]. Available: http://arxiv.org/abs/2207.04008
S. Seneviratne, E. Daskalaki, A. Lenskiy, and H. Suominen, “m-Networks: Adapting the Triplet Networks for Acronym Disambiguation,” 2022. [Online]. Available: https://github.com/sandaruSen/m_networks
A. Wagh and M. Khanna, “Clinical Abbreviation Disambiguation Using Clinical Variants of BERT,” in Multi-disciplinary Trends in Artificial Intelligence, R. Morusupalli, T. S. Dandibhotla, V. V. Atluri, D. Windridge, P. Lingras, and V. R. Komati, Eds., Cham: Springer Nature Switzerland, 2023, pp. 214–224.
M. Cevik, S. Mohammad Jafari, M. Myers, and S. Yildirim, “Sequence Labeling for Disambiguating Medical Abbreviations,” J Healthc Inform Res, vol. 7, no. 4, pp. 501–526, 2023, doi: https://doi.org/10.1007/s41666-023-00146-1
A. Vaswani et al, “Attention Is All You Need,” Jun. 2017.
S. Moon, S. Pakhomov, N. Liu, J. Ryan, and G. Melton, “A sense inventory for clinical abbreviations and acronyms created using clinical notes and medical dictionary resources,” J Am Med Inform Assoc, vol. 21, Jun. 2013, doi: https://doi.org/10.1136/amiajnl-2012-001506
S. P. S. M. G. Moon, “Clinical Abbreviation Sense Inventory. Retrieved from the University of Minnesota Digital Conservancy,” 2012.
J. Kaur and P. Buttar, “A Systematic Review on Stopword Removal Algorithms,” vol. 4, pp. 207–210, Apr. 2018.
Y. Peng, S. Yan, and Z. lu, Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets. 2019. doi: https://doi.org/10.18653/v1/W19-5006
A. E. W. Johnson et al, “MIMIC-III, a freely accessible critical care database,” Sci Data, vol. 3, no. 1, p. 160035, 2016, doi: https://doi.org/10.1038/sdata.2016.35
C. Huang, Y. Li, C. C. Loy, and X. Tang, “Learning Deep Representation for Imbalanced Classification,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jun. 2016, pp. 5375–5384. doi: https://doi.org/10.1109/CVPR.2016.580
M. Khushi et al, “A Comparative Performance Analysis of Data Resampling Methods on Imbalance Medical Data,” IEEE Access, vol. 9, pp. 109960–109975, 2021, doi: https://doi.org/10.1109/ACCESS.2021.3102399
N. Chawla, “Data Mining for Imbalanced Datasets: An Overview,” in Data Mining and Knowledge Discovery Handbook, ISBN 978-0-387-09822-7. Springer Science + Business Media, LLC, 2010, p. 875, vol. 5, 2005, pp. 853–867. doi: https://doi.org/10.1007/0-387-25465-X_40
I. Li et al, “A Neural Topic-Attention Model for Medical Term Abbreviation Disambiguation,” CoRR, vol. abs/1910.1, 2019, [Online]. Available: http://arxiv.org/abs/1910.14076
G. Adams, M. Ketenci, S. Bhave, A. Perotte, and N. Elhadad, “Zero-Shot Clinical Acronym Expansion via Latent Meaning Cells.,” Proc Mach Learn Res, vol. 136, pp. 12–40, Dec. 2020.
M. Agrawal, S. Hegselmann, H. Lang, Y. Kim, and D. Sontag, “Large Language Models are Few-Shot Clinical Information Extractors,” in Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022.
J. Kim et al, “Improved Clinical Abbreviation Expansion via Non-Sense-Based Approaches,” 2020. [Online]. Available: https://github.com/
Funding
No funds, grants, or other support was received.
Author information
Authors and Affiliations
Contributions
all authors contributed equally to this work.
Corresponding author
Ethics declarations
Conflict of Interest
The authors declare that they have no conflict of interest.
Protection of Human and Animal Subjects
Ethical approval was not required for this study.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic Supplementary Material
Below is the link to the electronic supplementary material.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Hosseini, M., Hosseini, M. & Javidan, R. Leveraging Large Language Models for Clinical Abbreviation Disambiguation. J Med Syst 48, 27 (2024). https://doi.org/10.1007/s10916-024-02049-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10916-024-02049-z