Skip to main navigation Skip to search Skip to main content

Extended Results for Enhancing Abstract Screening Classification in Evidence-Based Medicine: Incorporating Domain Knowledge into Pre-trained Models

  • Regina Ofori-Boateng*
  • , Magaly Aceves-Martins
  • , Nirmalie Wirantuga
  • , Carlos Francisco Moreno-García
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

2 Downloads (Pure)

Abstract

Evidence-based medicine (EBM) is a foundational element in medical research, playing a crucial role in shaping healthcare policies and clinical decision-making. However, the rigorous processes required for EBM, particularly during the abstract screening phase, pose substantial challenges to researchers. While many have sought to automate this stage using Pre-trained Language Models (PLMs), these efforts often face obstacles due to the specificity of the domain, especially when dealing with EBM studies related to both human and animal subjects. To address this, our initial research presented a state-of-the-art (SOTA) transfer learning approach that enhanced four abstract screening by embedding domain-specific knowledge into PLMs without modifying their base weights utilising the concepts of adapters. Extending the previous work, in this study, we evaluate the same methodology on four animal and human EBM datasets. Our evaluation, conducted on the further four EBM abstract screening datasets, demonstrates that the proposed method significantly improves the screening process and outperforms strong baseline PLMs.

Original languageEnglish
Title of host publicationProceedings of the SICSA Workshop on Reasoning, Evaluation and Application of Large Language Models (SICSA-REALLM 2024)
Subtitle of host publicationAberdeen, United Kingdom, October 17th, 2024
EditorsKyle Martin, Pedram Salimi, Vihanga Wijayasekara
PublisherCEUR-WS
Pages11-18
Number of pages8
Volume3822
Publication statusPublished - 17 Oct 2024
Event2024 SICSA Workshop on Reasoning, Evaluation and Application of Large Language Models, SICSA-REALLM 2024 - Aberdeen, United Kingdom
Duration: 17 Oct 202417 Oct 2024

Publication series

NameCEUR Workshop Proceedings
PublisherCEUR-WS
Volume3822
ISSN (Print)1613-0073

Conference

Conference2024 SICSA Workshop on Reasoning, Evaluation and Application of Large Language Models, SICSA-REALLM 2024
Country/TerritoryUnited Kingdom
CityAberdeen
Period17/10/2417/10/24

Funding

The authors thank members of the Childhood Obesity in Mexico (COMO)4 project for supporting this research.

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 3 - Good Health and Well-being
    SDG 3 Good Health and Well-being

Keywords

  • Domain Integration
  • Evidence-Based Medicine
  • Large/Pre-trained Language Models
  • Transfer Learning

Fingerprint

Dive into the research topics of 'Extended Results for Enhancing Abstract Screening Classification in Evidence-Based Medicine: Incorporating Domain Knowledge into Pre-trained Models'. Together they form a unique fingerprint.

Cite this