[go: up one dir, main page]

Emiliani et al., 1993 - Google Patents

Combined Application of Neural Network and Artificial Intelligence Methods to Automatic Speech Recognition in a Continuous Utterance

Emiliani et al., 1993

Document ID
9228035532790546445
Author
Emiliani U
Podini P
Sani F
Publication year
Publication venue
Artificial Neural Nets and Genetic Algorithms: Proceedings of the International Conference in Innsbruck, Austria, 1993

External Links

Snippet

A very efficient approach to using an artificial supervised neural network in Automatic Speech Recognition in the case of speaker dependent continuous utterance is presented in this paper; it has been tested in the Italian language but in principle not limited to it. An …
Continue reading at link.springer.com (other versions)

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the analysis technique
    • G10L25/30Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the analysis technique using neural networks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters
    • G10L25/18Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters
    • G10L25/09Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters the extracted parameters being zero crossing rates
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
    • G10L25/93Discriminating between voiced and unvoiced parts of speech signals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/04Segmentation; Word boundary detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding or deleting nodes or connections, pruning

Similar Documents

Publication Publication Date Title
US5150323A (en) Adaptive network for in-band signal separation
Das et al. Urban sound classification using convolutional neural network and long short term memory based on multiple features
Waibel Modular construction of time-delay neural networks for speech recognition
CN112257521B (en) CNN underwater acoustic signal target identification method based on data enhancement and time-frequency separation
Mehyadin et al. Birds sound classification based on machine learning algorithms
AU2019101150A4 (en) Speaker Identity Recognition System Based on Deep Learning
Medhat et al. Automatic classification of music genre using masked conditional neural networks
Koops et al. Automatic segmentation and deep learning of bird sounds
Si et al. Variational information bottleneck for effective low-resource audio classification
Anderson et al. Dynamic speech categorization with recurrent networks
Bengio et al. Programmable execution of multi-layered networks for automatic speech recognition
Saritha et al. Raw waveform based speaker identification using deep neural networks
Emiliani et al. Combined Application of Neural Network and Artificial Intelligence Methods to Automatic Speech Recognition in a Continuous Utterance
Pankajakshan et al. All-cony net for bird activity detection: Significance of learned pooling
Hunt Recurrent neural networks for syllabification
Tolba et al. Detection and Discrimination of Arabic Phonemes Using Long Short-Term Memory (LSTM) Model
Buermann et al. Speech recognition using very deep neural networks: Spectrograms vs cochleagrams
Das et al. An application programming interface to recognize emotion using speech features
KR0173204B1 (en) Speech Recognition Method of Variation Unit using Korean Variation Grouping Tree
Aleksandrovsky et al. Novel speech processing mechanism derived from auditory neocortical circuit analysis
Dutta et al. Interpretable acoustic representation learning on breathing and speech signals for covid-19 detection
Kasabov et al. Phoneme Recognition with Hierarchical Self Organised Neural Networks and Fuzzy Systems-A Case Study
Yegnanarayana et al. A speaker verification system using prosodic features.
Setianingrum et al. Speech Recognition of Sundanese Dialect Using Convolutional Neural Network Method with Mel-Spectrogram Feature Extraction
Kepuska et al. Investigation of phonemic context in speech using self-organizing feature maps