sBERT: Parameter-Efficient Transformer-Based Deep Learning Model for Scientific Literature Classification

Mohammad Munzir Ahanger, Mohd Arif Wani, Vasile Palade

Research output: Contribution to journalArticlepeer-review

10 Downloads (Pure)

Abstract

This paper introduces a parameter-efficient transformer-based model designed for scientific literature classification. By optimizing the transformer architecture, the proposed model significantly reduces memory usage, training time, inference time, and the carbon footprint associated with large language models. The proposed approach is evaluated against various deep learning models and demonstrates superior performance in classifying scientific literature. Comprehensive experiments conducted on datasets from Web of Science, ArXiv, Nature, Springer, and Wiley reveal that the proposed model’s multi-headed attention mechanism and enhanced embeddings contribute to its high accuracy and efficiency, making it a robust solution for text classification tasks.
Original languageEnglish
Article number4030022
Pages (from-to)397-421
Number of pages25
JournalKnowledge
Volume4
Issue number3
Early online date18 Jul 2024
DOIs
Publication statusE-pub ahead of print - 18 Jul 2024

Bibliographical note

© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).

Keywords

  • machine learning
  • deep learning
  • NLP
  • text classification
  • scientific literature classification

Fingerprint

Dive into the research topics of 'sBERT: Parameter-Efficient Transformer-Based Deep Learning Model for Scientific Literature Classification'. Together they form a unique fingerprint.

Cite this