Skip to Main Content (Press Enter)

Logo CNR
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills

UNI-FIND
Logo CNR

|

UNI-FIND

cnr.it
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills
  1. Outputs

On Robustness and Sensitivity of a Neural Language Model: A Case Study on Italian L1 Learner Errors

Academic Article
Publication Date:
2022
abstract:
In this paper, we propose a comprehensive linguistic study aimed at assessing the implicit behavior of one of the most prominent Neural Language Models (NLM) based on Transformer architectures, BERT (Devlin et al., 2019), when dealing with a particular source of noisy data, namely essays written by L1 Italian learners containing a variety of errors targeting grammar, orthography and lexicon. Differently from previous works, we focus on the pre-training stage and we devise two complementary evaluation tasks aimed at assessing the impact of errors on sentence-level inner representations in terms of semantic robustness and linguistic sensitivity. While the first evaluation perspective is meant to probe the model's ability to encode the semantic similarity between sentences also in the presence of errors, the second type of probing task evaluates the influence of errors on BERT's implicit knowledge of a set of raw and morpho-syntactic properties of a sentence. Our experiments show that BERT's ability to compute sentence similarity and to correctly encode multi-leveled linguistic information of a sentence are differently modulated by the category of errors and that the error hierarchies in terms of robustness and sensitivity change across layer-wise representations.
Iris type:
01.01 Articolo in rivista
Keywords:
nlp; interpretability; transformers; learner errors
List of contributors:
Miaschi, Alessio; Dell'Orletta, Felice; Venturi, Giulia; Brunato, DOMINIQUE PIERINA
Authors of the University:
BRUNATO DOMINIQUE PIERINA
DELL'ORLETTA FELICE
MIASCHI ALESSIO
VENTURI GIULIA
Handle:
https://iris.cnr.it/handle/20.500.14243/417257
  • Overview

Overview

URL

https://doi.org/10.1109/TASLP.2022.3226333
  • Use of cookies

Powered by VIVO | Designed by Cineca | 26.5.0.0 | Sorgente dati: PREPROD (Ribaltamento disabilitato)