Predictive Prompts with Joint Training of Large Language Models for Explainable Recommendation
Ching-Sheng Lin (),
Chung-Nan Tsai,
Shao-Tang Su,
Jung-Sing Jwo,
Cheng-Hsiung Lee and
Xin Wang
Additional contact information
Ching-Sheng Lin: Master Program of Digital Innovation, Tunghai University, Taichung 40704, Taiwan
Chung-Nan Tsai: Lam Research Japan GK, Kanagawa 222-0033, Japan
Shao-Tang Su: Master Program of Digital Innovation, Tunghai University, Taichung 40704, Taiwan
Jung-Sing Jwo: Master Program of Digital Innovation, Tunghai University, Taichung 40704, Taiwan
Cheng-Hsiung Lee: Master Program of Digital Innovation, Tunghai University, Taichung 40704, Taiwan
Xin Wang: Department of Epidemiology and Biostatistics, University at Albany School of Public Health, State University of New York, Rensselaer, NY 12144, USA
Mathematics, 2023, vol. 11, issue 20, 1-12
Abstract:
Large language models have recently gained popularity in various applications due to their ability to generate natural text for complex tasks. Recommendation systems, one of the frequently studied research topics, can be further improved using the capabilities of large language models to track and understand user behaviors and preferences. In this research, we aim to build reliable and transparent recommendation system by generating human-readable explanations to help users obtain better insights into the recommended items and gain more trust. We propose a learning scheme to jointly train the rating prediction task and explanation generation task. The rating prediction task learns the predictive representation from the input of user and item vectors. Subsequently, inspired by the recent success of prompt engineering, these predictive representations are served as predictive prompts, which are soft embeddings, to elicit and steer any knowledge behind language models for the explanation generation task. Empirical studies show that the proposed approach achieves competitive results compared with other existing baselines on the public English TripAdvisor dataset of explainable recommendations.
Keywords: large language models; recommendation systems; human-readable explanations; rating prediction task; explanation generation task; prompt engineering; predictive prompt (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2023
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (1)
Downloads: (external link)
https://www.mdpi.com/2227-7390/11/20/4230/pdf (application/pdf)
https://www.mdpi.com/2227-7390/11/20/4230/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:11:y:2023:i:20:p:4230-:d:1256691
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().