Probing for Bridging Inference in Transformer Language Models - ANR - Agence nationale de la recherche Accéder directement au contenu
Communication Dans Un Congrès Année : 2021

Probing for Bridging Inference in Transformer Language Models

Onkar Pandit
  • Fonction : Auteur
  • PersonId : 1052276
Yufang Hou
  • Fonction : Auteur
  • PersonId : 1104994

Résumé

We probe pre-trained transformer language models for bridging inference. We first investigate individual attention heads in BERT and observe that attention heads at higher layers prominently focus on bridging relations incomparison with the lower and middle layers, also, few specific attention heads concentrate consistently on bridging. More importantly, we consider language models as a whole in our second approach where bridging anaphora resolution is formulated as a masked token prediction task (Of-Cloze test). Our formulation produces optimistic results without any finetuning, which indicates that pre-trained language models substantially capture bridging inference. Our further investigation shows that the distance between anaphor-antecedent and the context provided to language models play an important role in the inference.
Fichier principal
Vignette du fichier
BridgingProbingBERT.pdf (341.25 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03284110 , version 1 (12-07-2021)

Identifiants

  • HAL Id : hal-03284110 , version 1

Citer

Onkar Pandit, Yufang Hou. Probing for Bridging Inference in Transformer Language Models. NAACL 2021 - Annual Conference of the North American Chapter of the Association for Computational Linguistics, Jun 2021, Online Conference, Mexico. ⟨hal-03284110⟩
41 Consultations
61 Téléchargements

Partager

Gmail Facebook X LinkedIn More