Transformer-Based Composite Language Models for Text Evaluation and Classification
Mihailo Škorić,
Miloš Utvić and
Ranka Stanković ()
Additional contact information
Mihailo Škorić: Faculty of Mining and Geology, University of Belgrade, Djusina 7, 11120 Belgrade, Serbia
Miloš Utvić: Faculty of Philology, University of Belgrade, Studentski Trg 3, 11000 Belgrade, Serbia
Ranka Stanković: Faculty of Mining and Geology, University of Belgrade, Djusina 7, 11120 Belgrade, Serbia
Mathematics, 2023, vol. 11, issue 22, 1-25
Abstract:
Parallel natural language processing systems were previously successfully tested on the tasks of part-of-speech tagging and authorship attribution through mini-language modeling, for which they achieved significantly better results than independent methods in the cases of seven European languages. The aim of this paper is to present the advantages of using composite language models in the processing and evaluation of texts written in arbitrary highly inflective and morphology-rich natural language, particularly Serbian. A perplexity-based dataset, the main asset for the methodology assessment, was created using a series of generative pre-trained transformers trained on different representations of the Serbian language corpus and a set of sentences classified into three groups (expert translations, corrupted translations, and machine translations). The paper describes a comparative analysis of calculated perplexities in order to measure the classification capability of different models on two binary classification tasks. In the course of the experiment, we tested three standalone language models (baseline) and two composite language models (which are based on perplexities outputted by all three standalone models). The presented results single out a complex stacked classifier using a multitude of features extracted from perplexity vectors as the optimal architecture of composite language models for both tasks.
Keywords: language modeling; language models; composite structures; machine learning; Serbian language; text classification (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2023
References: View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/11/22/4660/pdf (application/pdf)
https://www.mdpi.com/2227-7390/11/22/4660/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:11:y:2023:i:22:p:4660-:d:1281438
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().