Improving Neural Cross-lingual Abstractive Summarization via Employing Optimal Transport Distance for Knowledge Distillation
This repository contains the implementation of the paper Improving Neural Cross-lingual Abstractive Summarization via Employing Optimal Transport Distance for Knowledge Distillation.
Thong Nguyen, Luu Anh Tuan (AAAI 2022)
In this paper, we propose a novel Knowledge Distillation framework to tackle Neural Cross-Lingual Summarization for morphologically or structurally distant languages. In our framework, we propose a novel Knowledge Distillation framework to tackle Neural Cross-Lingual Summarization for morphologically or structurally distant languages. Extensive experiments in both high and low-resourced settings on multiple Cross-Lingual Summarization datasets that belong to pairs of morphologically and structurally distant languages demonstrate that extensive experiments in both high and low-resourced settings on multiple Cross-Lingual Summarization datasets that belong to pairs of morphologically and structurally distant languages.
@article{nguyen2021improving,
title={Improving Neural Cross-Lingual Summarization via Employing Optimal Transport Distance for Knowledge Distillation},
author={Nguyen, Thong and Tuan, Luu Anh},
journal={arXiv preprint arXiv:2112.03473},
year={2021}
}
- python3
- transformers
- pyrouge
- numpy
- pytorch 1.7.0
- Download and put the dataset in the
data
folder: https://drive.google.com/file/d/1m6Dhab5J74v9gAtgolVoorlF_2G-HlqR/view?usp=sharing - Train the monolingual teacher model by running
./run/continual_NCLS/ncls_train_<l1>2<l2>-<l1>2<l1>.sh
- Train the cross-lingual student model through executing
./run/continual_NCLS/ncls_train_<l1>2<l2>-<l1>2<l2>-ot_loss.sh
- Evaluate the cross-lingual student model via executing
./run/continual_NCLS/ncls_test_<l1>2<l2>-<l1>2<l2>-ot_loss.sh
Our implementation is based on the official code of MCLAS.