Skip to content

Latest commit

 

History

History
30 lines (17 loc) · 2.21 KB

File metadata and controls

30 lines (17 loc) · 2.21 KB

Decoder head

Exploration of unsupervised translation using fastai and pytorch

Imagine you have trained a language model (LM) to perform some task, say predict the next word or perform sentiment analysis. You freeze the model with removing one piece of information from it - the mapping from embeddings to words.

Can you create a setup in which the mapping from words to embeddings will be learnable? Something that the model could learn utilizing the information encoded in the LM? Surprisingly, the answer is yes. This repository is based on original work by Aza Raskin exploring this idea.

Decoder Head

Please read this for a more full description of a potential Decoder Head architecture.

How well will the model be able to learn this mapping? How will it handle synonyms? Is there a way to present this task to the model to make the learning more efficient? Iterating on the original idea and answering these questions is what the experiments in this repository will center on.

Special thanks

Special thanks to the authors of the MultiFiT: Efficient Multi-lingual Language Model Fine-tuning that is Julian Eisenschlos, Sebastian Ruder, Piotr Czapla, Marcin Kardas, Sylvain Gugger and Jeremy Howard! We trained our models on the wikipedia dumps that you were so kind to provide!

Another round of thanks goes out to authors of MUSE framework @ facebookresearch, that is Guillaume Lample and Alexis Conneau. We used the ground-truth bilingual dictionaries to evaluate the performance of our models.