Skip to Main Content (Press Enter)

Logo CNR
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills

UNI-FIND
Logo CNR

|

UNI-FIND

cnr.it
  • ×
  • Home
  • People
  • Outputs
  • Organizations
  • Expertise & Skills
  1. Outputs

Phonetically-Based Multi-Layered Neural Networks for Vowel Classification

Academic Article
Publication Date:
1990
abstract:
The vowel sub-component of a speaker-independent phoneme classification system will be described. The architecture of the vowel classifier is based on an ear model followed by a set of Multi-Layered Neural Networks (MLNN). MLNNs are trained to learn how to recognize articulatory features like the place of articulation and the manner of articulation related to tongue position. Experiments are performed on 10 English vowels showing a recognition rate higher than 95% on new speakers. When features are used for recognition, comparable results are obtained for vowels and diphthongs not used for training and pronounced by new speakers. This suggests that MLNNs suitably fed by the data computed by an ear model have good generalization capabilities over new speakers and new sounds.
Iris type:
01.01 Articolo in rivista
Keywords:
Classification & Recognition; Multi-Layered Neural Networks; Articulatory Features; Vowels; Ear Model
List of contributors:
Cosi, Piero
Handle:
https://iris.cnr.it/handle/20.500.14243/175542
Published in:
SPEECH COMMUNICATION
Journal
  • Overview

Overview

URL

http://www.sciencedirect.com/science/article/pii/0167639390900417
  • Use of cookies

Powered by VIVO | Designed by Cineca | 26.5.0.0 | Sorgente dati: PREPROD (Ribaltamento disabilitato)