Text clustering based on pre-trained models and autoencoders.

Qiang Xu, Hao Gu, ShengWei Ji
Author Information
  1. Qiang Xu: School of Artificial Intelligence and Big Data, Hefei University, Hefei, Anhui, China.
  2. Hao Gu: School of Artificial Intelligence and Big Data, Hefei University, Hefei, Anhui, China.
  3. ShengWei Ji: School of Artificial Intelligence and Big Data, Hefei University, Hefei, Anhui, China.

Abstract

Text clustering is the task of grouping text data based on similarity, and it holds particular importance in the medical field. sIn healthcare, medical data clustering is a highly active and effective research area. It not only provides strong support for making correct medical decisions from medical datasets but also aids in patient record management and medical information retrieval. With the development of the healthcare industry, a large amount of medical data is being generated, and traditional medical data clustering faces significant challenges. Many existing text clustering algorithms are primarily based on the bag-of-words model, which has issues such as high dimensionality, sparsity, and the neglect of word positions and context. Pre-trained models are a deep learning-based approach that treats text as a sequence to accurately capture word positions and context information. Moreover, compared to traditional K-means and fuzzy C-means clustering models, deep learning-based clustering algorithms are better at handling high-dimensional, complex, and nonlinear data. In particular, clustering algorithms based on autoencoders can learn data representations and clustering information, effectively reducing noise interference and errors during the clustering process. This paper combines pre-trained language models with deep embedding clustering models. Experimental results demonstrate that our model performs exceptionally well on four public datasets, outperforming most existing text clustering algorithms, and can be applied to medical data clustering.

Keywords

References

  1. Science. 2015 Jul 17;349(6245):261-6 [PMID: 26185244]
  2. J Big Data. 2022;9(1):15 [PMID: 35194542]
  3. Neural Comput. 2019 Jul;31(7):1235-1270 [PMID: 31113301]
  4. Bioinformatics. 2001 Sep;17(9):763-74 [PMID: 11590094]
  5. Neural Netw. 2017 Apr;88:22-31 [PMID: 28157556]
  6. J Med Internet Res. 2019 Jan 08;21(1):e10013 [PMID: 30622098]

Word Cloud

Created with Highcharts 10.0.0clusteringmedicaldatamodelstextdeepbasedalgorithmsinformationmodelpre-trainedTextparticularhealthcaredatasetstraditionalexistingwordpositionscontextlearning-basedautoencoderscantaskgroupingsimilarityholdsimportancefieldsInhighlyactiveeffectiveresearchareaprovidesstrongsupportmakingcorrectdecisionsalsoaidspatientrecordmanagementretrievaldevelopmentindustrylargeamountgeneratedfacessignificantchallengesManyprimarilybag-of-wordsissueshighdimensionalitysparsityneglectPre-trainedapproachtreatssequenceaccuratelycaptureMoreovercomparedK-meansfuzzyC-meansbetterhandlinghigh-dimensionalcomplexnonlinearlearnrepresentationseffectivelyreducingnoiseinterferenceerrorsprocesspapercombineslanguageembeddingExperimentalresultsdemonstrateperformsexceptionallywellfourpublicoutperformingappliedautoencoderembeddedlearning

Similar Articles

Cited By

No available data.