Text data augmentation and pre-trained Language Model for enhancing text classification of low-resource languages.

Atabay Ziyaden, Amir Yelenov, Fuad Hajiyev, Samir Rustamov, Alexandr Pak
Author Information
  1. Atabay Ziyaden: Kazakh-British Technical University, Almaty, Kazakhstan.
  2. Amir Yelenov: Institute of Information and Computational Technologies, Almaty, Kazakhstan. ORCID
  3. Fuad Hajiyev: School of Information Technologies and Engineering, ADA University, Baku, Azerbaijan.
  4. Samir Rustamov: School of Information Technologies and Engineering, ADA University, Baku, Azerbaijan.
  5. Alexandr Pak: Kazakh-British Technical University, Almaty, Kazakhstan.

Abstract

Background: In the domain of natural language processing (NLP), the development and success of advanced language models are predominantly anchored in the richness of available linguistic resources. Languages such as Azerbaijani, which is classified as a low-resource, often face challenges arising from limited labeled datasets, consequently hindering effective model training.
Methodology: The primary objective of this study was to enhance the effectiveness and generalization capabilities of news text classification models using text augmentation techniques. In this study, we solve the problem of working with low-resource languages using translations using the Facebook mBart50 model, as well as the Google Translate API and a combination of mBart50 and Google Translate thus expanding the capabilities when working with text.
Results: The experimental outcomes reveal a promising uptick in classification performance when models are trained on the augmented dataset compared with their counterparts using the original data. This investigation underscores the immense potential of combined data augmentation strategies to bolster the NLP capabilities of underrepresented languages. As a result of our research, we have published our labeled text classification dataset and pre-trained RoBERTa model for the Azerbaijani language.

Keywords

References

  1. PeerJ Comput Sci. 2023 Feb 8;9:e1224 [PMID: 37346576]
  2. PeerJ Comput Sci. 2021 Nov 16;7:e681 [PMID: 34901419]
  3. PeerJ Comput Sci. 2022 Dec 15;8:e1128 [PMID: 37346317]
  4. PeerJ Comput Sci. 2023 May 23;9:e1353 [PMID: 37346628]
  5. PeerJ Comput Sci. 2023 Jan 13;9:e1176 [PMID: 37346684]

Word Cloud

Created with Highcharts 10.0.0languagetextclassificationusingaugmentationmodelsAzerbaijanilow-resourcemodelcapabilitieslanguagesdataTextprocessingNLPlabeledstudyworkingmBart50GoogleTranslatedatasetpre-trainedlearningBackground:domainnaturaldevelopmentsuccessadvancedpredominantlyanchoredrichnessavailablelinguisticresourcesLanguagesclassifiedoftenfacechallengesarisinglimiteddatasetsconsequentlyhinderingeffectivetrainingMethodology:primaryobjectiveenhanceeffectivenessgeneralizationnewstechniquessolveproblemtranslationsFacebookwellAPIcombinationthusexpandingResults:experimentaloutcomesrevealpromisinguptickperformancetrainedaugmentedcomparedcounterpartsoriginalinvestigationunderscoresimmensepotentialcombinedstrategiesbolsterunderrepresentedresultresearchpublishedRoBERTaLanguageModelenhancingDeepLow-resourceMachineNatural

Similar Articles

Cited By