Skip to Main Content (Press Enter)

Logo CNR
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills

UNI-FIND
Logo CNR

|

UNI-FIND

cnr.it
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills
  1. Outputs

Emotion Classification from Speech and Text in Videos Using a Multimodal Approach

Academic Article
Publication Date:
2022
abstract:
Emotion classification is a research area in which there has been very intensive literature production concerning natural language processing, multimedia data, semantic knowledge discovery, social network mining, and text and multimedia data mining. This paper addresses the issue of emotion classification and proposes a method for classifying the emotions expressed in multimodal data extracted from videos. The proposed method models multimodal data as a sequence of features extracted from facial expressions, speech, gestures, and text, using a linguistic approach. Each sequence of multimodal data is correctly associated with the emotion by a method that models each emotion using a hidden Markov model. The trained model is evaluated on samples of multimodal sentences associated with seven basic emotions. The experimental results demonstrate a good classification rate for emotions.
Iris type:
01.01 Articolo in rivista
Keywords:
emotion classification; multimodal intera
List of contributors:
Grifoni, Patrizia; Ferri, Fernando; Caschera, MARIA CHIARA
Authors of the University:
CASCHERA MARIA CHIARA
FERRI FERNANDO
GRIFONI PATRIZIA
Handle:
https://iris.cnr.it/handle/20.500.14243/414841
Published in:
MULTIMODAL TECHNOLOGIES AND INTERACTION
Journal
  • Overview

Overview

URL

https://www.mdpi.com/2414-4088/6/4/28
  • Use of cookies

Powered by VIVO | Designed by Cineca | 26.5.0.0 | Sorgente dati: PREPROD (Ribaltamento disabilitato)