A deep neural network model for classifying pharmacy practice publications into research domains

Res Social Adm Pharm. 2024 Nov 5:S1551-7411(24)00387-5. doi: 10.1016/j.sapharm.2024.10.009. Online ahead of print.

Abstract

Background: Pharmacy practice faculty research profiles extend beyond the clinical and social domains, which are core elements of pharmacy practice. But as highlighted by journal editors in the Granada Statements, there is no consensus on these terms. Four domains (clinical, education, social & administrative, and basic & translational) of pharmacy practice faculty research are proposed.

Objectives: To develop a classifier for categorizing pharmacy practice faculty publications into four proposed domains, and to compare the model with zero-shot performances of state-of-the-art, general purpose large language models (gpLLMs).

Methods: One thousand abstracts from 2018 to 2021 documents published by pharmacy practice faculty were reviewed, labelled and used to screen and finetune several Bidirectional Encoders Representations from Transformers (BERT) models. The selected model was compared with zero-shot performances of 7 state-of-the-art gpLLMs including ChatGPT-4o, Gemini-1.5-Pro, Claude-3.5, LLAMA-3.1 and Mistral Large, using 80 randomly selected abstracts from 2023 publications labelled with ≥80% consensus by all authors. Classification metrics included F1, recall, precision and accuracy, and reproducibility was measured with Cohen's kappa. A use case was demonstrated by testing the null hypothesis that the research domain distribution of faculty publications was independent of the pandemic.

Result: The model - Pharmacy Practice Research Domain Classifier (PPRDC) produced a 5-fold stratified cross-validation metrics of 89.4 ± 1.7, 90.2 ± 2.2, 89.0 ± 1.7, and 95.5 ± 0.6, for F1, recall, precision and accuracy, respectively. PPRDC produced perfectly reproducible classifications (Cohen's kappa = 1.0) and outperformed zero-shot performances of all gpLLMs. F1 scores were 96.2 ± 1.6, 92.7 ± 1.2, 85.8 ± 3.2, and 83.1 ± 9.8 for education, clinical, social, and translational domains, respectively.

Conclusions: PPRDC (https://sadeosun-pprdc.streamlit.app) performed better than gpLLMs in this abstract classification task. Among several other impacts, PPRDC opens a new frontier in bibliometric studies; it will also advance the goals of the Grenada Statements by aiding authors and journal editors in journal selection and article prioritization decisions, respectively.

Keywords: Artificial intelligence; BERT; Classifier; Granada statements; Large language model; Pharmacy practice; Research domain.