Blin: A Multi-Task Sequence Recommendation Based on Bidirectional KL-Divergence and Linear Attention
Yanfeng Bai,
Haitao Wang () and
Jianfeng He
Additional contact information
Yanfeng Bai: Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
Haitao Wang: Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
Jianfeng He: Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
Mathematics, 2024, vol. 12, issue 15, 1-16
Abstract:
Sequence recommendation is a prominent research area within recommender systems, focused on predicting items that users may be interested in by modeling their historical interaction sequences. However, due to data sparsity, user interaction sequences in sequence recommendation are typically short. A common approach to address this issue is filling sequences with zero values, significantly reducing the effective utilization of input space. Furthermore, traditional sequence recommendation methods based on self-attention mechanisms exhibit quadratic complexity with respect to sequence length. These issues affect the performance of recommendation algorithms. To tackle these challenges, we propose a multi-task sequence recommendation model, Blin, which integrates bidirectional KL divergence and linear attention. Blin abandons the conventional zero-padding strategy, opting instead for random repeat padding to enhance sequence data. Additionally, bidirectional KL divergence loss is introduced as an auxiliary task to regularize the probability distributions obtained from different sequence representations. To improve the computational efficiency compared to traditional attention mechanisms, a linear attention mechanism is employed during sequence encoding, significantly reducing the computational complexity while preserving the learning capacity of traditional attention. Experimental results on multiple public datasets demonstrate the effectiveness of the proposed model.
Keywords: sequence recommendation; self-attention mechanism; consistency training; data augmentation (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2024
References: View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/12/15/2391/pdf (application/pdf)
https://www.mdpi.com/2227-7390/12/15/2391/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:12:y:2024:i:15:p:2391-:d:1447298
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().