Skip to Main Content (Press Enter)

Logo CNR
  • ×
  • Home
  • Persone
  • Pubblicazioni
  • Strutture
  • Competenze

UNI-FIND
Logo CNR

|

UNI-FIND

cnr.it
  • ×
  • Home
  • Persone
  • Pubblicazioni
  • Strutture
  • Competenze
  1. Pubblicazioni

Evaluating Text-To-Text Framework for Topic and Style Classification of Italian texts

Contributo in Atti di convegno
Data di Pubblicazione:
2022
Abstract:
In this paper, we propose an extensive evaluation of the first text-to-text Italian Neural Language Model (NLM), IT5 [1], on a classification scenario. In particular, we test the performance of IT5 on several tasks involving both the classification of the topic and the style of a set of Italian posts. We assess the model in two different configurations, single- and multi-task classification, and we compare it with a more traditional NLM based on the Transformer architecture (i.e. BERT). Moreover, we test its performance in a few-shot learning scenario. We also perform a qualitative investigation on the impact of label representations in modeling the classification of the IT5 model. Results show that IT5 could achieve good results, although generally lower than the BERT model. Nevertheless, we observe a significant performance improvement of the Text-to-text model in a multi-task classification scenario. Finally, we found that altering the representation of the labels mainly impacts the classification of the topic.
Tipologia CRIS:
04.01 Contributo in Atti di convegno
Keywords:
bert; style classification; t5; text-to-text; topic classification; transformers
Elenco autori:
Miaschi, Alessio; Dell'Orletta, Felice
Autori di Ateneo:
DELL'ORLETTA FELICE
MIASCHI ALESSIO
Link alla scheda completa:
https://iris.cnr.it/handle/20.500.14243/415084
Pubblicato in:
CEUR WORKSHOP PROCEEDINGS
Series
  • Dati Generali

Dati Generali

URL

http://www.scopus.com/record/display.url?eid=2-s2.0-85143252156&origin=inward
  • Utilizzo dei cookie

Realizzato con VIVO | Designed by Cineca | 26.5.0.0 | Sorgente dati: PREPROD (Ribaltamento disabilitato)