Telecom Fraud Recognition Based on Large Language Model Neuron Selection
Lanlan Jiang,
Cheng Zhang,
Xingguo Qin,
Ya Zhou,
Guanglun Huang,
Hui Li () and
Jun Li ()
Additional contact information
Lanlan Jiang: School of Business, Guilin University of Electronic Technology, Guilin 541004, China
Cheng Zhang: School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin 541004, China
Xingguo Qin: School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin 541004, China
Ya Zhou: School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin 541004, China
Guanglun Huang: School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin 541004, China
Hui Li: School of Informatics, Xiamen University, Xiamen 361005, China
Jun Li: School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin 541004, China
Mathematics, 2025, vol. 13, issue 11, 1-17
Abstract:
In the realm of natural language processing (NLP), text classification constitutes a task of paramount significance for large language models (LLMs). Nevertheless, extant methodologies predominantly depend on the output generated by the final layer of LLMs, thereby neglecting the wealth of information encapsulated within neurons residing in intermediate layers. To surmount this shortcoming, we introduce LENS (Linear Exploration and Neuron Selection), an innovative technique designed to identify and sparsely integrate salient neurons from intermediate layers via a process of linear exploration. Subsequently, these neurons are transmitted to downstream modules dedicated to text classification. This strategy effectively mitigates noise originating from non-pertinent neurons, thereby enhancing both the accuracy and computational efficiency of the model. The detection of telecommunication fraud text represents a formidable challenge within NLP, primarily attributed to its increasingly covert nature and the inherent limitations of current detection algorithms. In an effort to tackle the challenges of data scarcity and suboptimal classification accuracy, we have developed the LENS-RMHR (Linear Exploration and Neuron Selection with RoBERTa, Multi-head Mechanism, and Residual Connections) model, which extends the LENS framework. By incorporating RoBERTa, a multi-head attention mechanism, and residual connections, the LENS-RMHR model augments the feature representation capabilities and improves training efficiency. Utilizing the CCL2023 telecommunications fraud dataset as a foundation, we have constructed an expanded dataset encompassing eight distinct categories that encapsulate a diverse array of fraud types. Furthermore, a dual-loss function has been employed to bolster the model’s performance in multi-class classification scenarios. Experimental results reveal that LENS-RMHR demonstrates superior performance across multiple benchmark datasets, underscoring its extensive potential for application in the domains of text classification and telecommunications fraud detection.
Keywords: neuron selection; linear exploration; large language model; telecom fraud recognition (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/13/11/1784/pdf (application/pdf)
https://www.mdpi.com/2227-7390/13/11/1784/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:13:y:2025:i:11:p:1784-:d:1665628
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().