From pre-training to fine-tuning: : An in-depth analysis of Large Language Models in the biomedical domain
References
Index Terms
- From pre-training to fine-tuning: An in-depth analysis of Large Language Models in the biomedical domain
Recommendations
A Broad Study of Pre-training for Domain Generalization and Adaptation
Computer Vision – ECCV 2022AbstractDeep models must learn robust and transferable representations in order to perform well on new domains. While domain transfer methods (e.g., domain adaptation, domain generalization) have been proposed to learn transferable representations across ...
Pre-trained Language Models in Biomedical Domain: A Systematic Survey
Pre-trained language models (PLMs) have been the de facto paradigm for most natural language processing tasks. This also benefits the biomedical domain: researchers from informatics, medicine, and computer science communities propose various PLMs trained ...
Exploiting semantic annotations for open information extraction: an experience in the biomedical domain
AbstractThe increasing amount of unstructured text published on the Web is demanding new tools and methods to automatically process and extract relevant information. Traditional information extraction has focused on harvesting domain-specific, pre-...
Comments
Please enable JavaScript to view thecomments powered by Disqus.Information & Contributors
Information
Published In
Publisher
Elsevier Science Publishers Ltd.
United Kingdom
Publication History
Author Tags
Qualifiers
- Research-article
Contributors
Other Metrics
Bibliometrics & Citations
Bibliometrics
Article Metrics
- 0Total Citations
- 0Total Downloads
- Downloads (Last 12 months)0
- Downloads (Last 6 weeks)0