Chen et al., 2017 - Google Patents
Deep learning for dialogue systemsChen et al., 2017
View PDF- Document ID
- 9490846896874953118
- Author
- Chen Y
- Celikyilmaz A
- Hakkani-Tur D
- Publication year
- Publication venue
- Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts
External Links
Snippet
In the past decade, goal-oriented spoken dialogue systems have been the most prominent component in today's virtual personal assistants. The classic dialogue systems have rather complex and/or modular pipelines. The advance of deep learning technologies has recently …
- 238000011160 research 0 abstract description 18
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/183—Speech classification or search using natural language modelling using context dependencies, e.g. language models
- G10L15/19—Grammatical context, e.g. disambiguation of the recognition hypotheses based on word sequence rules
- G10L15/197—Probabilistic grammars, e.g. word n-grams
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1822—Parsing for meaning understanding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/065—Adaptation
- G10L15/07—Adaptation to the speaker
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L2015/088—Word spotting
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/226—Taking into account non-speech caracteristics
- G10L2015/228—Taking into account non-speech caracteristics of application context
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/20—Handling natural language data
- G06F17/27—Automatic analysis, e.g. parsing
- G06F17/2765—Recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Chen et al. | Deep learning for dialogue systems | |
Zeyer et al. | Improved training of end-to-end attention models for speech recognition | |
Kheddar et al. | Automatic speech recognition using advanced deep learning approaches: A survey | |
US10867599B2 (en) | Determining state of automated assistant dialog | |
CN114830148A (en) | Controlled benchmarked text generation | |
Ruede et al. | Yeah, right, uh-huh: a deep learning backchannel predictor | |
CN113646835A (en) | Joint automatic speech recognition and speaker binarization | |
Chi et al. | Speaker role contextual modeling for language understanding and dialogue policy learning | |
Wang et al. | Recent advances on human-computer dialogue | |
Chen et al. | Dynamic time-aware attention to speaker roles and contexts for spoken language understanding | |
Qian et al. | Neural approaches to automated speech scoring of monologue and dialogue responses | |
Su et al. | Deep learning for conversational AI | |
Celikyilmaz et al. | Deep learning in spoken and text-based dialog systems | |
Mamyrbayev et al. | Identifying the influence of transfer learning method in developing an end-toend automatic speech recognition system with a low data level | |
Radzikowski et al. | Dual supervised learning for non-native speech recognition | |
Qian et al. | Spoken language understanding of human-machine conversations for language learning applications | |
Sahay et al. | Modeling intent, dialog policies and response adaptation for goal-oriented interactions | |
Chen et al. | Open-domain neural dialogue systems | |
Qiu | Construction of english speech recognition model by fusing cnn and random deep factorization tdnn | |
Yun et al. | End-to-end emotional speech recognition using acoustic model adaptation based on knowledge distillation | |
Huang et al. | Chatbot: design, architecture, and applications | |
Yang et al. | A closer look at reinforcement learning-based automatic speech recognition | |
Qu et al. | Combining articulatory features with end-to-end learning in speech recognition | |
Ilievski | Building advanced dialogue managers for goal-oriented dialogue systems | |
Debatin et al. | Offline Speech Recognition Development |