Specific aspects of using large language models for text abstracting
https://doi.org/10.33186/1027-3689-2025-11-203-214
Abstract
In the context of spike of science publications, the automatic abstracting based on AI technologies has become a relevant task. The existing abstracting models use the trained large language models which deployment requires significant hardware resources. Meanwhile, specialized models based on the same transformer architecture do not require such big resources and therefore, can be used both on local servers and in the cloud environment at a much lower cost. The authors discuss the results of the ROUGE assessment of the abstracts generated in the LLM MBart (specialized model) and T-lite (universal model). The original large scale prompt was formed of the articles published in “Scientific and technical libraries” journal in 2025. The analysis findings evidences that MBart model demonstrates the better ROUGE metric value. However, the obtained data do not evidence on the quality of abstracts generated by the compared models, as the ROUGE metric shows just the match value for the words and phrases in the abstract and the reference text. The authors conclude that the “lightish” models, like MBart, may be deployed just locally in the libraries and without graphic processor, which would be more preferable for their practical common use.
About the Authors
M. V. GoncharovRussian Federation
Mikhail V. Goncharov – Cand. Sc. (Engineering), Associate Professor, Leading Researcher, Head, Group for Perspective Research and Analytic Forecasting
Moscow
K. A. Kolosov
Russian Federation
Kirill A. Kolosov – Cand. Sc. (Engineering), Leading Researcher
Moscow
References
1. Achkar P., Gollub T., Potthast M. Ask, Retrieve, Summarize: A Modular Pipeline for Scientific Literature Summarization // arXiv preprint arXiv:2505.16349. 2025.
2. Uckan T. A hybrid model for extractive summarization: Leveraging graph entropy to improve large language model performance // Ain Shams Engineering Journal. 2025. Vol. 16. № 5. (103348).
3. Shrai`berg Ia. L., Volkova K. Iu. Voprosy` avtorskogo prava v otnoshenii proizvedenii`, sozdanny`kh pri pomoshchi generativnogo iskusstvennogo intellekta // Nauchny`e i tekhnicheskie biblioteki. 2025. № 2. S. 115–130.
4. Lin T. et al. A survey of transformers // AI open. 2022. Vol. 3. P. 111–132.
5. Gusev I. Dataset for automatic summarization of Russian news // Conference on Artificial Intelligence and Natural Language. Cham : Springer International Publishing, 2020. P. 122–134.
6. By`chkova E. F., Kolosov K. A. Analiz vozmozhnostei` avtomaticheskogo referirovaniia statei` na primere istochnikov bazy` danny`kh «E`kologiia: nauka i tekhnologii» GPNTB Rossii // Nauchny`e i tekhnicheskie biblioteki. 2023. № 10. S. 99–120.
7. Karci A. Fractional order entropy: New perspectives // Optik. 2016. Vol. 127. № 20. P. 9172–9177.
8. Lin C. Y. Rouge: A package for automatic evaluation of summaries // Text summarization branches out. 2004. P. 74–81.
Review
For citations:
Goncharov M.V., Kolosov K.A. Specific aspects of using large language models for text abstracting. Scientific and Technical Libraries. 2025;1(11):203-214. (In Russ.) https://doi.org/10.33186/1027-3689-2025-11-203-214































