Preview

Scientific and Technical Libraries

Advanced search

Specific aspects of using large language models for text abstracting

https://doi.org/10.33186/1027-3689-2025-11-203-214

Abstract

In the context of spike of science publications, the automatic abstracting based on AI technologies has become a relevant task. The existing abstracting models use the trained large language models which deployment requires significant hardware resources. Meanwhile, specialized models based on the same transformer architecture do not require such big resources and therefore, can be used both on local servers and in the cloud environment at a much lower cost. The authors discuss the results of the ROUGE assessment of the abstracts generated in the LLM MBart (specialized model) and T-lite (universal model). The original large scale prompt was formed of the articles published in “Scientific and technical libraries” journal in 2025. The analysis findings evidences that MBart model demonstrates the better ROUGE metric value. However, the obtained data do not evidence on the quality of abstracts generated by the compared models, as the ROUGE metric shows just the match value for the words and phrases in the abstract and the reference text. The authors conclude that the “lightish” models, like MBart, may be deployed just locally in the libraries and without graphic processor, which would be more preferable for their practical common use.

About the Authors

M. V. Goncharov
Russian National Public Library for Science and Technology
Russian Federation

Mikhail V. Goncharov – Cand. Sc. (Engineering), Associate Professor, Leading Researcher, Head, Group for Perspective Research and Analytic Forecasting

Moscow 



K. A. Kolosov
Russian National Public Library for Science and Technology
Russian Federation

Kirill A. Kolosov – Cand. Sc. (Engineering), Leading Researcher

Moscow 



References

1. Achkar P., Gollub T., Potthast M. Ask, Retrieve, Summarize: A Modular Pipeline for Scientific Literature Summarization // arXiv preprint arXiv:2505.16349. 2025.

2. Uckan T. A hybrid model for extractive summarization: Leveraging graph entropy to improve large language model performance // Ain Shams Engineering Journal. 2025. Vol. 16. № 5. (103348).

3. Shrai`berg Ia. L., Volkova K. Iu. Voprosy` avtorskogo prava v otnoshenii proizvedenii`, sozdanny`kh pri pomoshchi generativnogo iskusstvennogo intellekta // Nauchny`e i tekhnicheskie biblioteki. 2025. № 2. S. 115–130.

4. Lin T. et al. A survey of transformers // AI open. 2022. Vol. 3. P. 111–132.

5. Gusev I. Dataset for automatic summarization of Russian news // Conference on Artificial Intelligence and Natural Language. Cham : Springer International Publishing, 2020. P. 122–134.

6. By`chkova E. F., Kolosov K. A. Analiz vozmozhnostei` avtomaticheskogo referirovaniia statei` na primere istochnikov bazy` danny`kh «E`kologiia: nauka i tekhnologii» GPNTB Rossii // Nauchny`e i tekhnicheskie biblioteki. 2023. № 10. S. 99–120.

7. Karci A. Fractional order entropy: New perspectives // Optik. 2016. Vol. 127. № 20. P. 9172–9177.

8. Lin C. Y. Rouge: A package for automatic evaluation of summaries // Text summarization branches out. 2004. P. 74–81.


Review

For citations:


Goncharov M.V., Kolosov K.A. Specific aspects of using large language models for text abstracting. Scientific and Technical Libraries. 2025;1(11):203-214. (In Russ.) https://doi.org/10.33186/1027-3689-2025-11-203-214

Views: 24


Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.


ISSN 1027-3689 (Print)
ISSN 2686-8601 (Online)