Emiliani et al., 1993 - Google Patents
Combined Application of Neural Network and Artificial Intelligence Methods to Automatic Speech Recognition in a Continuous UtteranceEmiliani et al., 1993
- Document ID
- 9228035532790546445
- Author
- Emiliani U
- Podini P
- Sani F
- Publication year
- Publication venue
- Artificial Neural Nets and Genetic Algorithms: Proceedings of the International Conference in Innsbruck, Austria, 1993
External Links
Snippet
A very efficient approach to using an artificial supervised neural network in Automatic Speech Recognition in the case of speaker dependent continuous utterance is presented in this paper; it has been tested in the Italian language but in principle not limited to it. An …
- 230000001537 neural 0 title abstract description 13
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/27—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the analysis technique
- G10L25/30—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the analysis technique using neural networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters
- G10L25/18—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1822—Parsing for meaning understanding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters
- G10L25/09—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00 characterised by the type of extracted parameters the extracted parameters being zero crossing rates
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
- G10L25/93—Discriminating between voiced and unvoiced parts of speech signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/04—Segmentation; Word boundary detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
- G06N3/082—Learning methods modifying the architecture, e.g. adding or deleting nodes or connections, pruning
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5150323A (en) | Adaptive network for in-band signal separation | |
Das et al. | Urban sound classification using convolutional neural network and long short term memory based on multiple features | |
Waibel | Modular construction of time-delay neural networks for speech recognition | |
CN112257521B (en) | CNN underwater acoustic signal target identification method based on data enhancement and time-frequency separation | |
Mehyadin et al. | Birds sound classification based on machine learning algorithms | |
AU2019101150A4 (en) | Speaker Identity Recognition System Based on Deep Learning | |
Medhat et al. | Automatic classification of music genre using masked conditional neural networks | |
Koops et al. | Automatic segmentation and deep learning of bird sounds | |
Si et al. | Variational information bottleneck for effective low-resource audio classification | |
Anderson et al. | Dynamic speech categorization with recurrent networks | |
Bengio et al. | Programmable execution of multi-layered networks for automatic speech recognition | |
Saritha et al. | Raw waveform based speaker identification using deep neural networks | |
Emiliani et al. | Combined Application of Neural Network and Artificial Intelligence Methods to Automatic Speech Recognition in a Continuous Utterance | |
Pankajakshan et al. | All-cony net for bird activity detection: Significance of learned pooling | |
Hunt | Recurrent neural networks for syllabification | |
Tolba et al. | Detection and Discrimination of Arabic Phonemes Using Long Short-Term Memory (LSTM) Model | |
Buermann et al. | Speech recognition using very deep neural networks: Spectrograms vs cochleagrams | |
Das et al. | An application programming interface to recognize emotion using speech features | |
KR0173204B1 (en) | Speech Recognition Method of Variation Unit using Korean Variation Grouping Tree | |
Aleksandrovsky et al. | Novel speech processing mechanism derived from auditory neocortical circuit analysis | |
Dutta et al. | Interpretable acoustic representation learning on breathing and speech signals for covid-19 detection | |
Kasabov et al. | Phoneme Recognition with Hierarchical Self Organised Neural Networks and Fuzzy Systems-A Case Study | |
Yegnanarayana et al. | A speaker verification system using prosodic features. | |
Setianingrum et al. | Speech Recognition of Sundanese Dialect Using Convolutional Neural Network Method with Mel-Spectrogram Feature Extraction | |
Kepuska et al. | Investigation of phonemic context in speech using self-organizing feature maps |