Tracking the Traces of Passivization and Negation in Contextualized Representations

Research output: Chapter in Book/Report/Conference proceedingConference contributionScientificpeer-review

Abstract

Contextualized word representations encode rich information about syntax and semantics, alongside specificities of each context of use. While contextual variation does not always reflect actual meaning shifts, it can still reduce the similarity of embeddings for word instances having the same meaning. We explore the imprint of two specific linguistic alternations, namely passivization and negation, on the representations generated by neural models trained with two different objectives: masked language modeling and translation. Our exploration methodology is inspired by an approach previously proposed for removing societal biases from word vectors. We show that passivization and negation leave their traces on the representations, and that neutralizing this information leads to more similar embeddings for words that should preserve their meaning in the transformation. We also find clear differences in how the respective features generalize across datasets.
Original languageEnglish
Title of host publicationProceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP
EditorsAfra Alishahi, Yonatan Belinkov, Grzegorz Chrupała, Dieuwke Hupkes, Yuval Pinter, Hassan Sajjad
Number of pages13
Place of PublicationStroudsburg
PublisherThe Association for Computational Linguistics
Publication date20 Nov 2020
Pages136-148
ISBN (Electronic)978-1-952148-86-6
DOIs
Publication statusPublished - 20 Nov 2020
MoE publication typeA4 Article in conference proceedings
EventBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP - Online event
Duration: 20 Nov 202020 Nov 2020
Conference number: 3

Fields of Science

  • 6121 Languages
  • 113 Computer and information sciences

Cite this