Abstract
This paper introduces a parameter-efficient transformer-based model designed for scientific literature classification. By optimizing the transformer architecture, the proposed model significantly reduces memory usage, training time, inference time, and the carbon footprint associated with large language models. The proposed approach is evaluated against various deep learning models and demonstrates superior performance in classifying scientific literature. Comprehensive experiments conducted on datasets from Web of Science, ArXiv, Nature, Springer, and Wiley reveal that the proposed model’s multi-headed attention mechanism and enhanced embeddings contribute to its high accuracy and efficiency, making it a robust solution for text classification tasks.
Original language | English |
---|---|
Article number | 4030022 |
Pages (from-to) | 397-421 |
Number of pages | 25 |
Journal | Knowledge |
Volume | 4 |
Issue number | 3 |
Early online date | 18 Jul 2024 |
DOIs | |
Publication status | E-pub ahead of print - 18 Jul 2024 |
Bibliographical note
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).Keywords
- machine learning
- deep learning
- NLP
- text classification
- scientific literature classification