Skip to content

We investigate the extent to which popular recurrent and attention-based neural models trained with an auto-regressive language modeling objective can represent the hierarchical nature of language by probing linguistic and structural properties using diagnostic classifiers.

Notifications You must be signed in to change notification settings

michaeljneely/probing-ar-language-models

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Uni-Directional Context is Insufficient for Hierarchical Language Understanding

Abstract

We investigate the extent to which popular recurrent and attention-based neural models trained with an auto-regressive language modeling objective can represent the hierarchical nature of language by probing linguistic and structural properties using diagnostic classifiers. We tune our classifiers to be selective, for which we introduce a novel tree distance control task and the generalized selectivity metric. Our results suggest that access to bidirectional context improves the quality of extracted parse tree geometry.

Authors

  • Michael J. Neely
  • Vanessa Botha

Notes

This is a student project from the University of Amsterdam's Natural Language Processing 2, Spring 2020 course. It requires access to some code and data located here.

About

We investigate the extent to which popular recurrent and attention-based neural models trained with an auto-regressive language modeling objective can represent the hierarchical nature of language by probing linguistic and structural properties using diagnostic classifiers.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published