Skip to content

nguyentthong/CrossSummOptimalTransport

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Improving Neural Cross-lingual Abstractive Summarization via Employing Optimal Transport Distance for Knowledge Distillation

This repository contains the implementation of the paper Improving Neural Cross-lingual Abstractive Summarization via Employing Optimal Transport Distance for Knowledge Distillation.

Thong Nguyen, Luu Anh Tuan (AAAI 2022)

Teaser image In this paper, we propose a novel Knowledge Distillation framework to tackle Neural Cross-Lingual Summarization for morphologically or structurally distant languages. In our framework, we propose a novel Knowledge Distillation framework to tackle Neural Cross-Lingual Summarization for morphologically or structurally distant languages. Extensive experiments in both high and low-resourced settings on multiple Cross-Lingual Summarization datasets that belong to pairs of morphologically and structurally distant languages demonstrate that extensive experiments in both high and low-resourced settings on multiple Cross-Lingual Summarization datasets that belong to pairs of morphologically and structurally distant languages.

@article{nguyen2021improving,
  title={Improving Neural Cross-Lingual Summarization via Employing Optimal Transport Distance for Knowledge Distillation},
  author={Nguyen, Thong and Tuan, Luu Anh},
  journal={arXiv preprint arXiv:2112.03473},
  year={2021}
}

Requirements

  • python3
  • transformers
  • pyrouge
  • numpy
  • pytorch 1.7.0

How to Run

  1. Download and put the dataset in the data folder: https://drive.google.com/file/d/1m6Dhab5J74v9gAtgolVoorlF_2G-HlqR/view?usp=sharing
  2. Train the monolingual teacher model by running ./run/continual_NCLS/ncls_train_<l1>2<l2>-<l1>2<l1>.sh
  3. Train the cross-lingual student model through executing ./run/continual_NCLS/ncls_train_<l1>2<l2>-<l1>2<l2>-ot_loss.sh
  4. Evaluate the cross-lingual student model via executing ./run/continual_NCLS/ncls_test_<l1>2<l2>-<l1>2<l2>-ot_loss.sh

Acknowledgement

Our implementation is based on the official code of MCLAS.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published