• LOGIN
    Login with username and password
Repository logo

BORIS Portal

Bern Open Repository and Information System

  • Publications
  • Projects
  • Research Data
  • Organizations
  • Researchers
  • Statistics
  • More
  • LOGIN
    Login with username and password
Repository logo
Unibern.ch
  1. Home
  2. Publications
  3. Title and abstract screening for literature reviews using large language models: an exploratory study in the biomedical domain.
 

Title and abstract screening for literature reviews using large language models: an exploratory study in the biomedical domain.

Options
  • Details
  • Files
BORIS DOI
10.48350/197864
Publisher DOI
10.1186/s13643-024-02575-4
PubMed ID
38879534
Description
BACKGROUND

Systematically screening published literature to determine the relevant publications to synthesize in a review is a time-consuming and difficult task. Large language models (LLMs) are an emerging technology with promising capabilities for the automation of language-related tasks that may be useful for such a purpose.

METHODS

LLMs were used as part of an automated system to evaluate the relevance of publications to a certain topic based on defined criteria and based on the title and abstract of each publication. A Python script was created to generate structured prompts consisting of text strings for instruction, title, abstract, and relevant criteria to be provided to an LLM. The relevance of a publication was evaluated by the LLM on a Likert scale (low relevance to high relevance). By specifying a threshold, different classifiers for inclusion/exclusion of publications could then be defined. The approach was used with four different openly available LLMs on ten published data sets of biomedical literature reviews and on a newly human-created data set for a hypothetical new systematic literature review.

RESULTS

The performance of the classifiers varied depending on the LLM being used and on the data set analyzed. Regarding sensitivity/specificity, the classifiers yielded 94.48%/31.78% for the FlanT5 model, 97.58%/19.12% for the OpenHermes-NeuralChat model, 81.93%/75.19% for the Mixtral model and 97.58%/38.34% for the Platypus 2 model on the ten published data sets. The same classifiers yielded 100% sensitivity at a specificity of 12.58%, 4.54%, 62.47%, and 24.74% on the newly created data set. Changing the standard settings of the approach (minor adaption of instruction prompt and/or changing the range of the Likert scale from 1-5 to 1-10) had a considerable impact on the performance.

CONCLUSIONS

LLMs can be used to evaluate the relevance of scientific publications to a certain review topic and classifiers based on such an approach show some promising results. To date, little is known about how well such systems would perform if used prospectively when conducting systematic literature reviews and what further implications this might have. However, it is likely that in the future researchers will increasingly use LLMs for evaluating and classifying scientific publications.
Date of Publication
2024-06-15
Publication Type
Article
Subject(s)
600 - Technology::610 - Medicine & health
Keyword(s)
Biomedicine Large language models Natural language processing Systematic literature review Title and abstract screening
Language(s)
en
Contributor(s)
Dennstädt, Fabio
Universitätsklinik für Radio-Onkologie
Zink, Johannes
Putora, Paul Martin
Universitätsklinik für Radio-Onkologie
Hastings, Janna
Cihoric, Nikola
Universitätsklinik für Radio-Onkologie
Additional Credits
Universitätsklinik für Radio-Onkologie
Series
Systematic Reviews
Publisher
BioMed Central
ISSN
2046-4053
Access(Rights)
open.access
Show full item
BORIS Portal
Bern Open Repository and Information System
Build: ae9592 [15.12. 16:43]
Explore
  • Projects
  • Funding
  • Publications
  • Research Data
  • Organizations
  • Researchers
  • Audiovisual Material
  • Software & other digital items
More
  • About BORIS Portal
  • Send Feedback
  • Cookie settings
  • Service Policy
Follow us on
  • Mastodon
  • YouTube
  • LinkedIn
UniBe logo