Distillation of Large Language Models for Text Simplification

Автор(и)

  • Олександр Скуржанський КНУ ім. Тараса Шевченка

DOI:

https://doi.org/10.31713/MCIT.2023.071

Анотація

This work presents a comprehensive methodology for harnessing the capabilities of Large Language Models to address specific Natural Language Processing tasks, with a focus on Text Simplification. While LLMs have demonstrated their prowess in tackling a wide range of NLP challenges, their demanding computational requirements can render them impractical for real-time online inference. In response to this limitation, we suggest the concept of text distillation, a technique aimed at effectively transferring the knowledge stored within LLMs to more compact and computationally efficient neural networks.

##submission.downloads##

Опубліковано

2023-11-22

Як цитувати

Скуржанський, О. (2023). Distillation of Large Language Models for Text Simplification. Моделювання, керування та інформаційні технології, (6), 230–231. https://doi.org/10.31713/MCIT.2023.071