CN109272995A - Audio recognition method, device and electronic equipment - Google Patents
Audio recognition method, device and electronic equipment Download PDFInfo
- Publication number
- CN109272995A CN109272995A CN201811126924.0A CN201811126924A CN109272995A CN 109272995 A CN109272995 A CN 109272995A CN 201811126924 A CN201811126924 A CN 201811126924A CN 109272995 A CN109272995 A CN 109272995A
- Authority
- CN
- China
- Prior art keywords
- user
- language model
- voice
- determining
- speech recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000004891 communication Methods 0.000 claims description 2
- 238000005516 engineering process Methods 0.000 abstract description 10
- 230000008901 benefit Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 230000003993 interaction Effects 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000005034 decoration Methods 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 208000024827 Alzheimer disease Diseases 0.000 description 1
- 238000000802 evaporation-induced self-assembly Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
Landscapes
- Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Machine Translation (AREA)
Abstract
The embodiment of the invention discloses a kind of audio recognition method, device and electronic equipments.Wherein method includes: the voice messaging for obtaining collected user's input;Voice messaging is identified with the language model that user matches according at least one, obtains speech recognition result.The embodiment of the present invention identifies voice messaging with the language model that user matches using at least one, this language model by matching with user knows otherwise voice messaging, realize the purpose that voice messaging is identified by appointed language model, not only increase the accuracy to voice messaging identification, it ensure that recognition result can meet the individual demand of user, and improve the accuracy and recognition efficiency of speech recognition, solves the technical issues of even wrong identification can not be identified caused by the indiscriminate general language model used in the related technology is identified, improve user experience.
Description
Technical Field
The embodiment of the invention relates to the technical field of voice recognition, in particular to a voice recognition method, a voice recognition device and electronic equipment.
Background
With the development of voice recognition technology, the application fields of voice wakeup are very wide, such as robots, mobile terminals, wearable devices, smart home devices, vehicle-mounted devices, and the like. However, in the related art, the speech recognition technology can only recognize some conventional words, and there is a technical problem that the words with strong specialization or even rarely used are not recognized or wrongly recognized.
Disclosure of Invention
In view of this, embodiments of the present invention provide a speech recognition method, a speech recognition device and an electronic device, which can achieve the above technical problem.
In order to solve the above problems, embodiments of the present invention mainly provide the following technical solutions:
in a first aspect, an embodiment of the present invention provides a speech recognition method, where the method includes:
acquiring collected voice information input by a user;
and recognizing the voice information according to at least one language model matched with the user to obtain a voice recognition result.
In a second aspect, an embodiment of the present invention further provides a speech recognition apparatus, where the apparatus includes:
the voice acquisition module is used for acquiring the collected voice information input by the user;
and the voice recognition module is used for recognizing the voice information according to at least one language model matched with the user to obtain a voice recognition result.
In a third aspect, an embodiment of the present invention further provides an electronic device, including:
at least one processor;
and at least one memory, bus connected with the processor; wherein,
the processor and the memory complete mutual communication through the bus;
the processor is used to call program instructions in the memory to perform the speech recognition method.
In a fourth aspect, embodiments of the present invention also provide a non-transitory computer-readable storage medium storing computer instructions that cause a computer to perform a speech recognition method.
By the technical scheme, the technical scheme provided by the embodiment of the invention at least has the following advantages: the voice information is recognized by adopting at least one language model matched with the user, and the mode of recognizing the voice information by the language model matched with the user realizes the aim of recognizing the voice information by the appointed language model, thereby not only improving the accuracy of recognizing the voice information, ensuring that the recognition result can meet the individual requirements of the user, but also improving the accuracy and the recognition efficiency of the voice recognition, solving the technical problem of incapability of recognition and even wrong recognition caused by the recognition of an undifferentiated universal language model adopted in the related technology, and improving the user experience.
The foregoing description is only an overview of the technical solutions of the embodiments of the present invention, and the embodiments of the present invention can be implemented according to the content of the description in order to make the technical means of the embodiments of the present invention more clearly understood, and the detailed description of the embodiments of the present invention is provided below in order to make the foregoing and other objects, features, and advantages of the embodiments of the present invention more clearly understandable.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the preferred embodiments. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the embodiments of the invention. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
fig. 1 is a flow chart illustrating a speech recognition method according to an embodiment of the present invention;
FIG. 2 is a flow chart illustrating a process for determining a language model matching a user according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a speech recognition apparatus according to an embodiment of the present invention;
FIG. 4 is a schematic structural diagram of another speech recognition apparatus provided in an embodiment of the present invention;
fig. 5 shows a schematic structural diagram of an electronic device according to an embodiment of the present invention.
Detailed Description
Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
Example one
As shown in fig. 1, the present invention provides a speech recognition method, comprising the steps of:
and step S101, acquiring the collected voice information input by the user.
The method provided by the embodiment of the invention is executed by the electronic equipment, and when the electronic equipment is in a working state, the purpose of acquiring the collected voice information input by the user is realized.
Specifically, the electronic device may include a terminal device or a cloud server.
In specific application, the voice information input by the user can be acquired by the terminal equipment, or acquired by the voice acquisition equipment connected with the terminal equipment. When the voice information input by the user is collected by the voice collecting device connected with the terminal device, the voice collecting device may be a microphone or the like. During practical application, the connection between the voice acquisition equipment and the terminal equipment can be realized through a data line or through wireless connection modes such as Bluetooth, so that the voice information input by the user acquired by the voice acquisition equipment is transmitted to the terminal equipment, and the terminal equipment or the cloud server acquires the voice information input by the user acquired by the voice acquisition equipment.
And S102, recognizing the voice information according to at least one language model matched with the user to obtain a voice recognition result.
Specifically, the language model matched with the user may be a specific language model capable of identifying a domain-specific term to which the user belongs, such as a language model corresponding to a computer domain term, a language model corresponding to a legal domain term, a language model of a financial domain term, and the like. Performing speech recognition via a particular language model that matches the user can not only increase the speed of speech recognition, but also increase the accuracy of speech recognition.
Specifically, there may be one or more language models that match the user.
The speech recognition method provided by the embodiment of the invention utilizes at least one language model matched with the user to recognize the speech information, and the language model matched with the user can be used for pertinently recognizing the professional nouns of the specific field to which the user belongs, so that the aim of performing speech recognition through the specified language model matched with the user is fulfilled, the accuracy of speech recognition can be improved, the recognition result can be ensured to meet the aim of the user, the speed of speech recognition is improved, the technical problem that the recognition result does not meet the requirements of the user or even cannot be recognized due to the fact that the speech recognition is performed by a non-differential general language model adopted in the related technology is solved, and the user experience is improved.
In practical application, the electronic device applying the voice recognition method provided by the invention can be a terminal device and can also be a cloud server.
When the electronic equipment is a cloud server, voice information input by a user is collected by the terminal equipment or audio collection equipment connected with the terminal equipment, the collected voice information input by the user is sent to the cloud server, the cloud server identifies the voice information according to at least one language model matched with the user, an obtained voice identification result is fed back to the terminal equipment, and the terminal equipment processes the voice identification result fed back by the cloud server to finish voice information identification.
When the electronic device is a terminal device (such as a mobile phone, a sound device, a tablet computer, a notebook computer, a PC, a wearable device, etc.), the terminal device (such as a mobile phone) acquires voice information input by a user, which is acquired by a microphone on the mobile phone, in a power-on state, and the mobile phone recognizes the voice information according to at least one language model matched with the user to obtain a voice recognition result.
In some examples, prior to recognizing the speech information according to at least one language model that matches the user, the method further comprises: at least one language model matching the user is determined.
For the embodiment of the invention, after at least one language model matched with the user is determined, the purpose is to use the determined at least one language model matched with the user as a language model matched with the user for identifying the voice information, so that the language model for identifying the voice information is specified, and the accuracy for identifying the voice information is improved.
Specifically, in some embodiments, one implementation of determining at least one language model matching the user may include step S201 (not shown) or step S202.
Step S201, based on the received selection instruction of the user for the language model, at least one language model matched with the user is determined.
The embodiment of the invention utilizes the received selection instruction of the user aiming at the language model to determine the language model matched with the user. Specifically, the selection instruction of the user may be to select one language model or to select a plurality of language models.
Specifically, the selection instruction may be sent by the user through a human-computer interaction interface provided by the terminal device. For example, the terminal device provides an identification of a language model of a plurality of domain-specific terms on a human-computer interaction interface, such as the identification can be legal, medical, computer, financial, etc. Supposing that the marks selected by the user on the human-computer interaction interface are legal, computer and medical, the human-computer interaction interface generates corresponding selection instructions according to the marks selected by the user, so that the terminal equipment determines a language model matched with the user according to the received selection instructions, and the finally determined language model is a language model of a legal field term, a language model of a computer field term and a language model of a financial field term.
The following description is given by taking an example in which the electronic device determines at least one language model matching the user as the terminal device.
During practical application, the terminal device can be a mobile phone, a pad, a notebook, a wearable device and intelligent devices such as a smart watch and a smart sound box. All language models can be stored locally in the terminal equipment or stored in the cloud. Thus, there may be two ways to determine a language model that matches the user: one is that based on the received selection instruction of the user for the language model, the language model stored in the cloud is downloaded to the terminal device, and the language model downloaded to the terminal device is used as the language model matched with the user to identify the voice information; and the other is to determine a language model matched with the user in the language models stored locally in the terminal equipment based on the received selection instruction of the user for the language models.
As shown in fig. 2, step S202 includes steps S2021 and S2022.
Step S2021, obtain the user' S history input record and/or user attributes.
In the embodiment of the invention, the history input record of the user is the content input by the user in a certain period of time in the past, wherein the source of the content input by the user can be the content input by text or the content input by voice. Specifically, the content input by the user can be input in a search engine, can be input in an input method, and can also be input by using other software tools in actual application. For example, in the last month, the content input by the user in a certain input method (such as a dog search input method and a hundredth input method) mainly relates to editing codes, composing novels and papers.
In the embodiment of the invention, the user attributes can comprise occupation, profession, interests and the like of the user, wherein the user interests can be determined by the user according to preset interest classification selection. In particular, user interests may include stars, sports, movies, prose, law, technology, and so forth.
Step S2022, determining at least one field corresponding to the user according to the historical input record and/or the user attribute, and taking the language model corresponding to the at least one field as the language model matched with the user.
The embodiment of the invention determines at least one field corresponding to the user through the historical input record or the user attribute of the user, thereby determining the corresponding language model according to the determined at least one field and completing the purpose of determining at least one language model matched with the user.
In specific application, only the historical input record of the user or only the user attribute may be acquired, or both the historical input record and the user attribute of the user may be acquired, so as to achieve the purpose of determining the language model matched with the user.
In practical application, the step S201, the step S202, or both the step S201 and the step S202 may be referred to for determining at least one language model matching the user.
In some examples, after determining at least one language model matching the user with reference to step S201, step S202, or both step S201 and step S202, the method further includes: and updating the language model matched with the user based on the user-defined word stock.
According to the embodiment of the invention, at least one language model matched with the user is updated by the user-defined word bank, so that during voice recognition, the updated at least one language model matched with the user can be quickly recognized according to the user-defined voice word bank.
Specifically, the user-defined word stock may be a user-defined speech word stock, and may also be a user-defined word stock of word meaning.
For example, when the user-defined word bank is a user-defined speech word bank, as in the user-defined word bank: the result corresponding to the customized voice "A" is a "new message", and the result corresponding to the customized voice "B" is a "DNA"; the result corresponding to the customized voice 'C' is 'science and technology creation destiny'. During voice recognition, after the language model matched with the user is updated by the user-defined word bank, when the updated language model matched with the user detects that the voice input by the user is 'A', the result can be directly output as 'new information' without further recognition, thereby accelerating the recognition efficiency.
For example, when the user-defined word bank is a word bank of the meaning, it is assumed that the language model matched with the user is a language model of a professional term in the legal field, the user-defined word bank of the meaning includes a medical special word "alzheimer's disease" commonly used by the user and a financial special word "colleague", the professional term commonly used by the user in the word bank of the meaning is updated to the language model of the professional term in the legal field, so that during later speech recognition, the updated language model of the professional term in the legal field can recognize the professional term in the illegal field quickly while recognizing the professional term in the legal field.
Specifically, in some embodiments, another implementation of determining at least one language model that matches a user includes: step S301, step S302, and step S303.
In step S301 (not shown), a user-defined word library is obtained.
Step S302 (not shown in the figure), generating a personalized language model of the user based on the user-defined word stock;
step S303 (not shown), determining the personalized language model of the user as the language model matching the user.
According to the embodiment of the invention, the personalized language model suitable for the user is generated according to the user-defined word stock, and the personalized language model is determined as the language model matched with the user.
In some embodiments, after determining at least one language model matching the user with reference to step S201, step S202, or both step S201 and step S202, the personalized language model may also be determined as the language model matching the user with reference to steps S301 to S303 in this embodiment.
For example, after determining that the language model is a language model of a legal field term, a language model of a computer field term, and a language model of a financial field term with reference to step S201, step S202, or both step S201 and step S202, the personalized language model is simultaneously determined as a language model matching the user. Therefore, when the speech recognition is carried out by using the language models of the professional nouns in different fields, the personalized language model can be used for carrying out the recognition quickly, and a recognition result is obtained. In practical application, when the language models of the professional nouns in different fields and the personalized language model exist in the determined language model matched with the user at the same time, the priority of the personalized language model can be set to be higher than that of the language model matched with the user, so that the personalized language model is used for preferentially recognizing in the recognition process.
The following describes embodiments of the present invention by taking an example of applying the method provided by the embodiments of the present invention to a search engine.
When the search engine runs in the background or foreground, the mobile phone acquires the voice information input by the user, then identifies the voice information by using at least one language model which is downloaded locally and matched with the user, obtains a voice identification result and outputs the voice identification result to an input window of the mobile phone search engine, and simultaneously controls the search engine to search according to the voice identification result and displays the search result. Or the mobile phone acquires the voice information input by the user, then sends the voice information to the server, and at least one language model which is pre-stored on the server and matched with the user identifies the voice information to obtain a voice identification result, then searches based on the voice identification, and feeds back the corresponding search result to the mobile phone terminal.
In order to further explain the speech recognition method provided by the present invention, the following describes an embodiment of the present invention by taking an example of applying the method provided by the embodiment of the present invention to an input method application program.
When a user downloads and installs an input method, the input method application program can display language models of word banks in different fields to the user so that the user can select the language models, and the input method downloads the corresponding language models to the server according to the selection of the user and stores the language models locally to serve as the language models matched with the user; or the input method application program downloads all language models to the local area during installation, and during installation, the input method application program can display the identification information of the language models of the word banks in different fields to the user so as to facilitate the user to select, so that the language models corresponding to the selection of the user are preferentially identified during later speech information identification, and after an identification result is obtained, the identification result is displayed on an editing interface to complete input.
Example two
As shown in fig. 3, which is a schematic structural diagram of a speech recognition apparatus according to an embodiment of the present invention, the speech recognition apparatus 30 according to an embodiment of the present invention may include: a voice acquisition module 301 and a voice recognition module 302.
The voice acquiring module 301 is configured to acquire acquired voice information input by a user;
the speech recognition module 302 is configured to recognize the speech information according to at least one language model matched with the user, so as to obtain a speech recognition result.
The voice recognition device provided by the embodiment of the invention utilizes at least one language model matched with a user to recognize voice information, and the mode of recognizing the voice information through the language model matched with the user realizes the purpose of recognizing the voice information through the specified language model, thereby not only improving the accuracy of recognizing the voice information, ensuring that the recognition result can meet the personalized requirements of the user, but also improving the accuracy and the recognition efficiency of voice recognition, solving the technical problem that the indiscriminate universal language model adopted in the related technology can not be recognized or even be recognized wrongly, and improving the user experience.
The speech recognition apparatus of this embodiment can execute the speech recognition method provided in the first embodiment of the present invention, and the implementation principles thereof are similar, and are not described herein again.
Further, as shown in fig. 4, the apparatus 30 further includes: a first model determining module 303, a custom thesaurus obtaining module 304, a personalized model generating module 305 and a second model determining module 306.
The first model determining module 303 is configured to determine at least one language model matching the user before recognizing the speech information according to the at least one language model matching the user;
a custom thesaurus obtaining module 304, configured to obtain a user-defined thesaurus;
a personalized model generation module 305 for generating a personalized language model of the user based on the user-defined word stock,
the second model determining module 306 is configured to determine the personalized language model of the user as a language model matching the user before recognizing the speech information according to at least one language model matching the user.
In some embodiments, the first model determining module 303 comprises: a first determining unit 3031 (not shown in the figure), wherein the first determining unit 3031 is configured to determine at least one language model matching the user based on a received instruction of the user for selecting the language model.
In some embodiments, the first model determining module 303 comprises: a user data unit 3032 (not shown in the figure) and a second determination unit 3033 (not shown in the figure), wherein,
a user data unit 3031, configured to obtain a history input record and/or a user attribute of a user;
a second determining unit 3033, configured to determine at least one domain corresponding to the user according to the history input record and/or the user attribute, and use a language model corresponding to the at least one domain as a language model matched with the user.
Further, as shown in fig. 4, the apparatus 30 further includes: model re-determination module 307.
The model re-determination module 307 is configured to update the language model matched with the user based on the user-defined word bank.
The speech recognition device provided by the embodiment can update the language model matched with the user based on the user-defined word bank so as to realize the purposes of recognizing the speech information according to the user-defined word bank and improving the recognition efficiency.
In a specific application, the speech recognition device provided by the embodiment of the invention can not comprise a user-defined word bank obtaining module and a personalized model generating module, and the purpose of updating the language model matched with the user by obtaining the user-defined word bank is realized.
The speech recognition device of the embodiment of the present invention can execute the speech recognition method provided in the first embodiment, which is similar to the first embodiment in the implementation principle, and is not described herein again.
EXAMPLE III
An embodiment of the present invention provides an electronic device, as shown in fig. 5, an electronic device 600 shown in fig. 5 includes: a processor 6001 and a memory 6003. Processor 6001 and memory 6003 are coupled, such as via bus 6002. Further, the electronic device 600 may also include a transceiver 6006. It should be noted that the transceiver 6006 is not limited to one in practical application, and the structure of the electronic device 600 is not limited to the embodiment of the present invention.
The processor 6001 could be a CPU, general purpose processor, DSP, ASIC, FPGA or other programmable logic device, transistor logic device, hardware component, or any combination thereof. Which may implement or perform the various illustrative logical blocks, modules, and circuits described in connection with the disclosure. The processor 6001 might also be a combination that performs a computing function, such as a combination comprising one or more microprocessors, a combination of a DSP and a microprocessor, or the like.
The bus 6002 may include a path that conveys information between the aforementioned components. The bus 6002 may be a PCI bus, an EISA bus, or the like. The bus 6002 can be divided into an address bus, a data bus, a control bus, and so forth. For ease of illustration, only one thick line is shown in FIG. 5, but this is not intended to represent only one bus or type of bus.
Memory 6003 can be, but is not limited to, a ROM or other type of static storage device that can store static information and instructions, a RAM or other type of dynamic storage device that can store information and instructions, an EEPROM, CD-ROM or other optical disk storage, optical disk storage (including compact disk, laser disk, optical disk, digital versatile disk, blu-ray disk, etc.), magnetic disk storage media or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer.
The memory 6003 is used to store application code that implements aspects of the invention and is controlled for execution by the processor 6001. Processor 6001 is configured to execute application code stored in memory 6003 to implement the speech recognition apparatus provided by the embodiments shown in fig. 3 and 4.
The electronic equipment applying the voice recognition method provided by the embodiment of the invention utilizes at least one language model matched with the user to recognize the voice information, and the mode of recognizing the voice information through the language model matched with the user realizes the purpose of recognizing the voice information through the specified language model, thereby not only improving the accuracy of recognizing the voice information and ensuring that the recognition result can meet the personalized requirements of the user, but also improving the accuracy and the recognition efficiency of the voice recognition, solving the technical problem that the indiscriminate universal language model adopted in the related technology can not be recognized or even wrongly recognized, and improving the user experience.
Example four
Embodiments of the present invention provide a non-transitory computer-readable storage medium storing computer instructions that cause a computer to perform a speech recognition method as shown in one of the above-described method embodiments.
Compared with the prior art, the non-transitory computer readable storage medium provided by the embodiment of the invention has the advantages that the voice information is identified by utilizing at least one language model matched with the user, the voice information is identified by the language model matched with the user, the aim of identifying the voice information by specifying the language model is fulfilled, the accuracy of identifying the voice information is improved, the identification result can meet the personalized requirement of the user, the accuracy and the identification efficiency of the voice identification are improved, the technical problem that the identification cannot be carried out or even is mistakenly identified due to the identification of an undifferentiated universal language model adopted in the related technology is solved, and the user experience is improved.
It should be understood that, although the steps in the flowcharts of the figures are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and may be performed in other orders unless explicitly stated herein. Moreover, at least a portion of the steps in the flow chart of the figure may include multiple sub-steps or multiple stages, which are not necessarily performed at the same time, but may be performed at different times, which are not necessarily performed in sequence, but may be performed alternately or alternately with other steps or at least a portion of the sub-steps or stages of other steps.
The foregoing is only a partial embodiment of the present invention, and it should be noted that, for those skilled in the art, various modifications and decorations can be made without departing from the principle of the present invention, and these modifications and decorations should also be regarded as the protection scope of the present invention.
Claims (10)
1. A speech recognition method, comprising:
acquiring collected voice information input by a user;
and recognizing the voice information according to at least one language model matched with the user to obtain a voice recognition result.
2. The method of claim 1, wherein prior to recognizing the speech information according to at least one language model matching the user, further comprising:
at least one language model matching the user is determined.
3. The method of claim 2, wherein determining at least one language model that matches the user comprises:
and determining at least one language model matched with the user based on the received selection instruction of the user for the language model.
4. The method of claim 2, wherein determining at least one language model that matches the user comprises:
acquiring a historical input record and/or user attributes of a user;
and determining at least one field corresponding to the user according to the historical input record and/or the user attribute, and taking a language model corresponding to the at least one field as a language model matched with the user.
5. The method of claim 2, wherein determining at least one language model that matches the user comprises:
acquiring the word bank customized by the user;
generating a personalized language model of the user based on the user-defined word stock;
and determining the personalized language model of the user as a language model matched with the user.
6. The method of claim 1, further comprising:
and updating the language model matched with the user based on the user-defined word bank.
7. A speech recognition apparatus, comprising:
the voice acquisition module is used for acquiring the collected voice information input by the user;
and the voice recognition module is used for recognizing the voice information according to at least one language model matched with the user to obtain a voice recognition result.
8. The apparatus of claim 7, wherein the speech recognition module, prior to recognizing the speech information according to at least one language model matching the user, further comprises:
and the model determining module is used for determining at least one language model matched with the user.
9. An electronic device, comprising:
at least one processor;
and at least one memory, bus connected with the processor; wherein,
the processor and the memory complete mutual communication through the bus;
the processor is configured to invoke program instructions in the memory to perform the speech recognition method of any of claims 1 to 6.
10. A non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform the speech recognition method of any one of claims 1 to 6.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201811126924.0A CN109272995A (en) | 2018-09-26 | 2018-09-26 | Audio recognition method, device and electronic equipment |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201811126924.0A CN109272995A (en) | 2018-09-26 | 2018-09-26 | Audio recognition method, device and electronic equipment |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN109272995A true CN109272995A (en) | 2019-01-25 |
Family
ID=65197827
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201811126924.0A Pending CN109272995A (en) | 2018-09-26 | 2018-09-26 | Audio recognition method, device and electronic equipment |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN109272995A (en) |
Cited By (11)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109903750A (en) * | 2019-02-21 | 2019-06-18 | 科大讯飞股份有限公司 | A kind of audio recognition method and device |
| CN110288993A (en) * | 2019-06-26 | 2019-09-27 | 广州探迹科技有限公司 | A kind of individualized intelligent voice interactive method and device based on container technique |
| CN110428816A (en) * | 2019-02-26 | 2019-11-08 | 北京蓦然认知科技有限公司 | A kind of method and device voice cell bank training and shared |
| CN111402864A (en) * | 2020-03-19 | 2020-07-10 | 北京声智科技有限公司 | Voice processing method and electronic equipment |
| CN111986651A (en) * | 2020-09-02 | 2020-11-24 | 上海优扬新媒信息技术有限公司 | Man-machine interaction method and device and intelligent interaction terminal |
| CN112992127A (en) * | 2019-12-12 | 2021-06-18 | 杭州海康威视数字技术股份有限公司 | Voice recognition method and device |
| WO2021129439A1 (en) * | 2019-12-28 | 2021-07-01 | 科大讯飞股份有限公司 | Voice recognition method and related product |
| CN113138674A (en) * | 2020-01-19 | 2021-07-20 | 北京搜狗科技发展有限公司 | Input method and related device |
| CN113763925A (en) * | 2021-05-26 | 2021-12-07 | 腾讯科技(深圳)有限公司 | Speech recognition method, speech recognition device, computer equipment and storage medium |
| CN116318716A (en) * | 2023-02-17 | 2023-06-23 | 支付宝(杭州)信息技术有限公司 | Authentication method, device, storage medium, electronic equipment and product |
| CN117198289A (en) * | 2023-09-28 | 2023-12-08 | 阿波罗智联(北京)科技有限公司 | Voice interaction method, device, equipment, medium and product |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102541505A (en) * | 2011-01-04 | 2012-07-04 | 中国移动通信集团公司 | Voice input method and system thereof |
| CN105489221A (en) * | 2015-12-02 | 2016-04-13 | 北京云知声信息技术有限公司 | Voice recognition method and device |
| CN108288467A (en) * | 2017-06-07 | 2018-07-17 | 腾讯科技(深圳)有限公司 | A kind of audio recognition method, device and speech recognition engine |
| US10032463B1 (en) * | 2015-12-29 | 2018-07-24 | Amazon Technologies, Inc. | Speech processing with learned representation of user interaction history |
-
2018
- 2018-09-26 CN CN201811126924.0A patent/CN109272995A/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102541505A (en) * | 2011-01-04 | 2012-07-04 | 中国移动通信集团公司 | Voice input method and system thereof |
| CN105489221A (en) * | 2015-12-02 | 2016-04-13 | 北京云知声信息技术有限公司 | Voice recognition method and device |
| US10032463B1 (en) * | 2015-12-29 | 2018-07-24 | Amazon Technologies, Inc. | Speech processing with learned representation of user interaction history |
| CN108288467A (en) * | 2017-06-07 | 2018-07-17 | 腾讯科技(深圳)有限公司 | A kind of audio recognition method, device and speech recognition engine |
Cited By (18)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109903750A (en) * | 2019-02-21 | 2019-06-18 | 科大讯飞股份有限公司 | A kind of audio recognition method and device |
| CN109903750B (en) * | 2019-02-21 | 2022-01-04 | 科大讯飞股份有限公司 | Voice recognition method and device |
| CN110428816A (en) * | 2019-02-26 | 2019-11-08 | 北京蓦然认知科技有限公司 | A kind of method and device voice cell bank training and shared |
| CN110428816B (en) * | 2019-02-26 | 2022-06-03 | 杭州蓦然认知科技有限公司 | Method and device for training and sharing voice cell bank |
| CN110288993A (en) * | 2019-06-26 | 2019-09-27 | 广州探迹科技有限公司 | A kind of individualized intelligent voice interactive method and device based on container technique |
| CN112992127A (en) * | 2019-12-12 | 2021-06-18 | 杭州海康威视数字技术股份有限公司 | Voice recognition method and device |
| CN112992127B (en) * | 2019-12-12 | 2024-05-07 | 杭州海康威视数字技术股份有限公司 | A method and device for speech recognition |
| US12451128B2 (en) | 2019-12-28 | 2025-10-21 | Iflytek Co., Ltd. | Voice recognition method and related product |
| WO2021129439A1 (en) * | 2019-12-28 | 2021-07-01 | 科大讯飞股份有限公司 | Voice recognition method and related product |
| CN113138674A (en) * | 2020-01-19 | 2021-07-20 | 北京搜狗科技发展有限公司 | Input method and related device |
| CN111402864A (en) * | 2020-03-19 | 2020-07-10 | 北京声智科技有限公司 | Voice processing method and electronic equipment |
| CN111986651B (en) * | 2020-09-02 | 2023-09-29 | 度小满科技(北京)有限公司 | Man-machine interaction method and device and intelligent interaction terminal |
| CN111986651A (en) * | 2020-09-02 | 2020-11-24 | 上海优扬新媒信息技术有限公司 | Man-machine interaction method and device and intelligent interaction terminal |
| CN113763925B (en) * | 2021-05-26 | 2024-03-12 | 腾讯科技(深圳)有限公司 | Speech recognition method, device, computer equipment and storage medium |
| CN113763925A (en) * | 2021-05-26 | 2021-12-07 | 腾讯科技(深圳)有限公司 | Speech recognition method, speech recognition device, computer equipment and storage medium |
| CN116318716A (en) * | 2023-02-17 | 2023-06-23 | 支付宝(杭州)信息技术有限公司 | Authentication method, device, storage medium, electronic equipment and product |
| CN117198289A (en) * | 2023-09-28 | 2023-12-08 | 阿波罗智联(北京)科技有限公司 | Voice interaction method, device, equipment, medium and product |
| CN117198289B (en) * | 2023-09-28 | 2024-05-10 | 阿波罗智联(北京)科技有限公司 | Voice interaction method, device, equipment, medium and product |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN109272995A (en) | Audio recognition method, device and electronic equipment | |
| CN110263142B (en) | Method and apparatus for outputting information | |
| KR102733920B1 (en) | System and method for natural language processing | |
| CN106155686A (en) | Interface creating method, device and system | |
| CN108416003A (en) | A kind of picture classification method and device, terminal, storage medium | |
| CN109360571A (en) | Processing method and processing device, storage medium, the computer equipment of credit information | |
| CN107656996B (en) | Man-machine interaction method and device based on artificial intelligence | |
| KR20200080400A (en) | Method for providing sententce based on persona and electronic device for supporting the same | |
| CN103678304A (en) | Method and device for pushing specific content for predetermined webpage | |
| CN108959247B (en) | Data processing method, server and computer readable medium | |
| CN109119079B (en) | Voice input processing method and device | |
| KR102787542B1 (en) | Electronic device providing variation utterance text and operating method thereof | |
| CN107589828A (en) | The man-machine interaction method and system of knowledge based collection of illustrative plates | |
| CN104866308A (en) | Scenario image generation method and apparatus | |
| CN110890088A (en) | Voice information feedback method and device, computer equipment and storage medium | |
| CN110473543B (en) | A kind of speech recognition method and device | |
| CN111401044A (en) | Title generation method and device, terminal equipment and storage medium | |
| CN106898352A (en) | Sound control method and electronic equipment | |
| US12118985B2 (en) | Electronic device and method for providing on-device artificial intelligence service | |
| CN118428336A (en) | Method, device, electronic device and storage medium for generating low-code applications | |
| US20060155546A1 (en) | Method and system for controlling input modalities in a multimodal dialog system | |
| CN111444321A (en) | Question answering method, device, electronic equipment and storage medium | |
| CN113923517B (en) | Background music generation method and device and electronic equipment | |
| CN117352132A (en) | Psychological coaching method, device, equipment and storage medium | |
| CN111680514B (en) | Information processing and model training method, device, equipment and storage medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| RJ01 | Rejection of invention patent application after publication | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190125 |