Skip to Main Content (Press Enter)

Logo CNR
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills

UNI-FIND
Logo CNR

|

UNI-FIND

cnr.it
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills
  1. Outputs

Contextual and Non-Contextual Word Embeddings: an in-depth Linguistic Investigation

Conference Paper
Publication Date:
2020
abstract:
In this paper we present a comparison between the linguistic knowledge encoded in the internal representations of a contextual Language Model (BERT) and a contextual-independent one (Word2vec). We use a wide set of probing tasks, each of which corresponds to a distinct sentence-level feature extracted from different levels of linguistic annotation. We show that, although BERT is capable of understanding the full context of each word in an input sequence, the implicit knowledge encoded in its aggregated sentence representations is still comparable to that of a contextual-independent model. We also find that BERT is able to encode sentence-level properties even within single-word embeddings, obtaining comparable or even superior results than those obtained with sentence representations.
Iris type:
04.01 Contributo in Atti di convegno
Keywords:
nlp; interpretability; representation learning
List of contributors:
Miaschi, Alessio; Dell'Orletta, Felice
Authors of the University:
DELL'ORLETTA FELICE
MIASCHI ALESSIO
Handle:
https://iris.cnr.it/handle/20.500.14243/421763
  • Overview

Overview

URL

https://www.aclweb.org/anthology/2020.repl4nlp-1.15
  • Use of cookies

Powered by VIVO | Designed by Cineca | 26.5.0.0 | Sorgente dati: PREPROD (Ribaltamento disabilitato)