Revolutionary text clustering: Investigating transfer learning capacity of SBERT models through pooling techniques
dc.authorid | ORTAKCI, Yasin/0000-0002-0683-2049 | |
dc.contributor.author | Ortakci, Yasin | |
dc.date.accessioned | 2024-09-29T15:57:40Z | |
dc.date.available | 2024-09-29T15:57:40Z | |
dc.date.issued | 2024 | |
dc.department | Karabük Üniversitesi | en_US |
dc.description.abstract | Large Language Models (LLMs), one of the most advanced representatives of neural networks, have revolutionized the field of natural language processing. Among the many applications of these models, text clustering is gaining increasing interest. In particular, the fact that LLMs digitize text more semantically and contextually than existing methods in the literature has led LLMs to produce more successful results with clustering algorithms. However, since these models are not specifically designed for text clustering, they can lead to processing times that exceed acceptable runtime thresholds. To address this challenge, the Sentence-BERT (SBERT) model has been proposed as a solution, offering the ability to accurately measure text similarity by transforming entire texts into dense, fixed-size vectors. SBERT has been integrated into various LLMs, resulting in the creation of diverse SBERT model variants. This study aims to assess the transfer learning capabilities of SBERT models in the context of text clustering. Furthermore, it investigates the influence of CLS (classification token), mean, and max pooling techniques on the performance of these models. In this direction, we applied these pooling techniques to DistilBERT, DistilRoBERTa, ALBERT, and MPNET based SBERT models and compared their performance on different corpora. The results show that there is no clear superiority among the SBERT models. However, the mean pooling emerged as the most effective method in 13 out of 16 text clustering tasks. This finding underscores the high compatibility of the mean pooling technique with SBERT models. | en_US |
dc.identifier.doi | 10.1016/j.jestch.2024.101730 | |
dc.identifier.issn | 2215-0986 | |
dc.identifier.scopus | 2-s2.0-85195462434 | en_US |
dc.identifier.scopusquality | Q1 | en_US |
dc.identifier.uri | https://doi.org/10.1016/j.jestch.2024.101730 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14619/4923 | |
dc.identifier.volume | 55 | en_US |
dc.identifier.wos | WOS:001252306300001 | en_US |
dc.identifier.wosquality | N/A | en_US |
dc.indekslendigikaynak | Web of Science | en_US |
dc.indekslendigikaynak | Scopus | en_US |
dc.language.iso | en | en_US |
dc.publisher | Elsevier - Division Reed Elsevier India Pvt Ltd | en_US |
dc.relation.ispartof | Engineering Science and Technology-An International Journal-Jestech | en_US |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
dc.rights | info:eu-repo/semantics/openAccess | en_US |
dc.subject | SBERT | en_US |
dc.subject | Large language models | en_US |
dc.subject | Sentence embeddings | en_US |
dc.subject | Text clustering | en_US |
dc.subject | Pooling techniques | en_US |
dc.title | Revolutionary text clustering: Investigating transfer learning capacity of SBERT models through pooling techniques | en_US |
dc.type | Article | en_US |