Unified Training for Cross-Lingual Abstractive Summarization by Aligning Parallel Machine Translation Pairs
Shaohuan Cheng,
Wenyu Chen,
Yujia Tang,
Mingsheng Fu and
Hong Qu ()
Additional contact information
Shaohuan Cheng: School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
Wenyu Chen: School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
Yujia Tang: School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
Mingsheng Fu: School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
Hong Qu: School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
Mathematics, 2024, vol. 12, issue 13, 1-16
Abstract:
Cross-lingual summarization (CLS) is essential for enhancing global communication by facilitating efficient information exchange across different languages. However, owing to the scarcity of CLS data, recent studies have employed multi-task frameworks to combine parallel monolingual summaries. These methods often use independent decoders or models with non-shared parameters because of the mismatch in output languages, which limits the transfer of knowledge between CLS and its parallel data. To address this issue, we propose a unified training method for CLS that combines parallel machine translation (MT) pairs with CLS pairs, jointly training them within a single model. This design ensures consistent input and output languages and promotes knowledge sharing between the two tasks. To further enhance the model’s capability to focus on key information, we introduce two additional loss terms to align the hidden representations and probability distributions between the parallel MT and CLS pairs. Experimental results demonstrate that our method outperforms competitive methods in both full-dataset and low-resource scenarios on two benchmark datasets, Zh2EnSum and En2ZhSum.
Keywords: cross-lingual summarization; multi-task learning; machine translation; low-resource scenario (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2024
References: View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/12/13/2107/pdf (application/pdf)
https://www.mdpi.com/2227-7390/12/13/2107/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:12:y:2024:i:13:p:2107-:d:1429140
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().