Domain-specific or Uncertainty-aware models: Does it really make a difference for biomedical text classification?

Aman Sinha, Timothee Mickus, Marianne Clausel, Mathieu Constant, Xavier Coubez

Research output: Chapter in Book/Report/Conference proceedingConference contributionScientificpeer-review

Abstract

The success of pretrained language models (PLMs) across a spate of use-cases has led to significant investment from the NLP community towards building domain-specific foundational models. On the other hand, in mission critical settings such as biomedical applications, other aspects also factor in---chief of which is a model's ability to produce reasonable estimates of its own uncertainty. In the present study, we discuss these two desiderata through the lens of how they shape the entropy of a model's output probability distribution. We find that domain specificity and uncertainty awareness can often be successfully combined, but the exact task at hand weighs in much more strongly.
Original languageEnglish
Title of host publicationProceedings of the 23rd Workshop on Biomedical Natural Language Processing
EditorsDina Demner-Fushman, Sophia Ananiadou, Makoto Miwa, Kirk Roberts, Junichi Tsujii
Number of pages10
Place of PublicationStroudsburg
PublisherThe Association for Computational Linguistics
Publication date1 Aug 2024
Pages202-211
ISBN (Electronic)979-8-89176-130-8
Publication statusPublished - 1 Aug 2024
MoE publication typeA4 Article in conference proceedings
EventWorkshop on Biomedical Natural Language Processing - Bangkok, Thailand
Duration: 16 Aug 202416 Aug 2024
Conference number: 23

Fields of Science

  • 6121 Languages
  • 113 Computer and information sciences

Cite this