[go: up one dir, main page]

CN109688271A - The method, apparatus and terminal device of contact information input - Google Patents

The method, apparatus and terminal device of contact information input Download PDF

Info

Publication number
CN109688271A
CN109688271A CN201910039493.2A CN201910039493A CN109688271A CN 109688271 A CN109688271 A CN 109688271A CN 201910039493 A CN201910039493 A CN 201910039493A CN 109688271 A CN109688271 A CN 109688271A
Authority
CN
China
Prior art keywords
contact
pinyin
contact person
person
name
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910039493.2A
Other languages
Chinese (zh)
Inventor
黄世光
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
OneConnect Smart Technology Co Ltd
Original Assignee
OneConnect Smart Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by OneConnect Smart Technology Co Ltd filed Critical OneConnect Smart Technology Co Ltd
Priority to CN201910039493.2A priority Critical patent/CN109688271A/en
Publication of CN109688271A publication Critical patent/CN109688271A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/14Speech classification or search using statistical models, e.g. Hidden Markov Models [HMMs]
    • G10L15/142Hidden Markov Models [HMMs]
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/12Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being prediction coefficients
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/24Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being the cepstrum
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/78Detection of presence or absence of voice signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Probability & Statistics with Applications (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The present invention is suitable for terminal device technical field, provides the method and terminal device of a kind of contact information input, which comprises the acquisition contact person's voice messaging instruction for listening to user's triggering receives contact person's voice messaging of user speech input;Contact person's voice messaging is converted into contact person's phonetic, contact person's phonetic is compared with the name pinyin prestored;If contact person's phonetic only fills the associated contact information of the name pinyin with successful match to corresponding contact information input frame with a name pinyin successful match.The present invention realizes contact information input by speech recognition, reduces user's operation, convenient and efficient.

Description

Method and device for inputting contact information and terminal equipment
Technical Field
The invention belongs to the technical field of terminal equipment, and particularly relates to a method and a device for inputting contact person information and terminal equipment.
Background
With the continuous popularization of terminal devices, more and more things in people's lives depend on the terminal devices to realize.
In the prior art, contact information is input at a terminal device, usually a user triggers an address book control at a position corresponding to a contact information input box, the terminal device loads an address book contact list, then the terminal device is used for selecting a target contact in the loaded address book contact list, and finally target information is selected in the contact information of the target contact, so that the target information of the target contact is filled in the contact information input box.
Therefore, the existing mode is complex in user operation and not convenient enough.
Disclosure of Invention
In view of this, embodiments of the present invention provide a method and an apparatus for inputting contact information, and a terminal device, so as to solve the technical problems of cumbersome user operation, inconvenience and rapidness when the terminal device inputs contact information in the prior art.
A first aspect of an embodiment of the present invention provides a method for inputting contact information, including:
monitoring a command for acquiring the voice information of the contact person triggered by a user, and receiving the voice information of the contact person input by the voice of the user;
converting the contact person voice information into contact person pinyin, and comparing the contact person pinyin with prestored person name pinyin;
and if the contact person pinyin is successfully matched with one contact person pinyin, filling the contact person information associated with the successfully matched contact person pinyin into a corresponding contact person information input box.
A second aspect of an embodiment of the present invention provides an apparatus for inputting contact information, including:
the monitoring and receiving module is used for monitoring a command for acquiring the voice information of the contact person triggered by the user and receiving the voice information of the contact person input by the voice of the user;
the conversion and comparison module is used for converting the contact person voice information into contact person pinyin and comparing the contact person pinyin with prestored person name pinyin;
and the first filling module is used for filling the contact information associated with the successfully matched name pinyin into the corresponding contact information input box if the contact pinyin is successfully matched with only one name pinyin.
A third aspect of the embodiments of the present invention provides a terminal device, including a memory and a processor, where the memory stores a computer program that is executable on the processor, and when the processor executes the computer program, the following steps are implemented:
monitoring a command for acquiring the voice information of the contact person triggered by a user, and receiving the voice information of the contact person input by the voice of the user;
converting the contact person voice information into contact person pinyin, and comparing the contact person pinyin with prestored person name pinyin;
and if the contact person pinyin is successfully matched with one contact person pinyin, filling the contact person information associated with the successfully matched contact person pinyin into a corresponding contact person information input box.
A fourth aspect of embodiments of the present invention provides a computer-readable storage medium storing a computer program which, when executed by a processor, implements the steps of:
monitoring a command for acquiring the voice information of the contact person triggered by a user, and receiving the voice information of the contact person input by the voice of the user;
converting the contact person voice information into contact person pinyin, and comparing the contact person pinyin with prestored person name pinyin;
and if the contact person pinyin is successfully matched with one contact person pinyin, filling the contact person information associated with the successfully matched contact person pinyin into a corresponding contact person information input box.
In the embodiment of the invention, the contact person pinyin is obtained by identifying the contact person voice information input by the user, the contact person pinyin is compared with the prestored person name pinyin, and when the contact person pinyin is successfully matched with only one person name pinyin, the successfully matched contact person information is filled into the corresponding contact person information input frame, so that the contact person information is input by voice, the user operation is reduced, and the method and the device are convenient and quick.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the embodiments or the prior art descriptions will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without inventive exercise.
Fig. 1 is a flowchart of an implementation of a method for inputting contact information according to an embodiment of the present invention;
FIG. 2 is an exemplary diagram of a contact information input interface provided by an embodiment of the invention;
FIG. 3 is a flowchart illustrating an implementation of step 102 in a method for inputting contact information according to an embodiment of the present invention;
FIG. 4 is a diagram illustrating populating contact information in a contact information input interface according to an embodiment of the invention;
FIG. 5 is a flowchart of an implementation of another method for inputting contact information according to an embodiment of the present invention;
fig. 6 is a flowchart of an implementation of step 501 in another method for inputting contact information according to an embodiment of the present invention;
FIG. 7 is a diagram illustrating a target contact selection interface loaded on a contact information input interface according to an embodiment of the present invention;
fig. 8 is a block diagram illustrating a structure of an apparatus for inputting contact information according to an embodiment of the present invention;
fig. 9 is a schematic diagram of a terminal device according to an embodiment of the present invention.
Detailed Description
In the following description, for purposes of explanation and not limitation, specific details are set forth, such as particular system structures, techniques, etc. in order to provide a thorough understanding of the embodiments of the invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced in other embodiments that depart from these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present invention with unnecessary detail.
In order to explain the technical means of the present invention, the following description will be given by way of specific examples.
Fig. 1 shows an implementation process of a method for inputting contact information provided by an embodiment of the present invention, where the method includes steps S101 to S103. The method is suitable for the situation that the contact information input is needed. The method is executed by a contact information input device, and the contact information input device is configured on a terminal device and can be realized by software and/or hardware. The terminal equipment comprises an intelligent mobile terminal and the like. The specific implementation principle of each step is as follows.
S101, a contact person voice information acquisition instruction triggered by a user is monitored, and contact person voice information input by the user is received.
In the embodiment of the invention, the user is a user of the terminal equipment. The terminal device includes, but is not limited to, a portable terminal device. The terminal equipment comprises a microphone and other sound receiving equipment, and the terminal equipment receives the voice of the user through the sound receiving equipment.
In any application which can be installed on the terminal equipment, the application comprises but is not limited to a browser, an electronic mail, a loan application, a shopping application or an instant messaging application and the like, when a user fills in contact information, a contact voice information acquisition control appears at a position corresponding to a contact information input box in a contact information input interface, and a contact voice information acquisition instruction triggered by clicking the control by the user is monitored, so that a radio device of the terminal equipment is started to receive contact voice information input by the user according to the contact voice information acquisition instruction.
For example, as shown in fig. 2, in a contact information input interface 20 of a certain loan application of a smart phone, to input contact information in a contact information input box 21, a user may click an acquire contact voice information control 22 to trigger an acquire contact voice information instruction, so that after receiving contact voice information input by the user, the smart phone matches with contacts in an address book according to the contact voice information, and fills the successfully matched contact information in the contact information input box 21. It is to be understood that this is done by way of example only and is not to be construed as limiting.
S102, converting the contact person voice information into contact person pinyin, and comparing the contact person pinyin with pre-stored person name pinyin.
In the embodiment of the invention, after the voice information of the contact is acquired through the step 101, the voice information of the contact is converted into the pinyin of the contact, and the pinyin of the contact is compared with the pre-stored pinyin of the name of the contact. The name pinyin is the name pinyin corresponding to the contact in the contact list of the address book, and is prestored in a memory of the terminal device, so that when matching is needed, a comparison result is obtained by traversing all the prestored name pinyins.
It should be noted that the contact list includes contacts and contact information associated with each contact. The contact information includes, but is not limited to, a name, a cell phone number, a work landline number, a home landline number, a mailbox number, an instant messaging account number, or the like.
Optionally, in step 102, the contact voice information is converted into contact pinyin, as shown in fig. 3, including steps 301 to 302.
S301, extracting acoustic feature vectors of the contact person voice information.
The acoustic feature vector is a feature vector at least comprising two acoustic features of a linear prediction cepstrum coefficient and a Mel frequency cepstrum coefficient.
The more acoustic features are selected, the larger the quantity processing amount is, the calculation data amount is controlled while the result accuracy is ensured, and 2 acoustic features are selected to form an acoustic feature vector. Illustratively, 2 acoustic features that extract contact voice information constitute an acoustic feature vector: Mel-Frequency Cepstral Coefficients (MFCC) and Linear Prediction Cepstral Coefficients (LPCC). Wherein the MFCC can be in multiple stages, e.g. 7 stages, CoTo C6
And S302, analyzing the acoustic feature vector by using the trained acoustic model to obtain the contact person pinyin.
The acoustic Model may be a Hidden Markov Model (HMM), among others. And analyzing the acoustic feature vector by using the trained acoustic model to obtain the contact person pinyin.
Further, before S301, the method further includes: and preprocessing the voice information of the contact. Wherein the preprocessing includes silence removal and framing. By preprocessing the voice information of the contact person, on one hand, the data volume processing amount is reduced, and on the other hand, the accuracy of the voice recognition result is improved.
In the embodiment of the present invention, Voice Activity Detection (VAD) technology is adopted to implement endpoint Detection so as to eliminate silence.
The framing is performed using a moving window function, with an overlap between frames. Illustratively, the speech signal is processed in frames with a frame length of 25ms and a frame shift of 10ms to obtain N-frame sub-signals, i.e. each frame has a length of 25ms, and each two frames have an overlap of 25-10 ═ 15 ms. Or processing the voice signal by frames with the frame length of 20ms and the frame shift of 10ms to obtain N frame sub-signals, namely, the length of each frame is 20ms, and every two frames are overlapped by 20-10 ms.
And if the contact person voice information is preprocessed, recognizing the preprocessed contact person voice information to obtain contact person pinyin. At this time, correspondingly, step 301 includes: and extracting sub-acoustic feature vectors of each frame of contact voice information, and combining the sub-acoustic feature vectors of all frames into an acoustic feature matrix.
Accordingly, step 302 includes: and analyzing the acoustic characteristic matrix by using the trained acoustic model to obtain the contact person pinyin.
S103, if the contact person pinyin is successfully matched with only one contact person pinyin, filling the contact person information associated with the successfully matched contact person pinyin into a corresponding contact person information input box.
In the embodiment of the invention, if the contact person pinyin is successfully matched with only one person name pinyin, at the moment, the contact person associated with the successfully matched person name pinyin is the target contact person, the target contact person is determined, and the contact person information associated with the target contact person is filled into the corresponding contact person information input box.
The keywords of each item of contact information related to the target contact are matched with the keywords of the contact information input box, so that the contact information successfully matched can be filled into the corresponding contact information input box without the need of a user to execute operation to determine whether a name, a telephone, a mailbox number and the like are filled into the contact information filling box, the intelligent level is further provided, the user operation is reduced, and the user experience is improved.
Exemplarily, continuing with the example shown in fig. 2, after receiving the contact voice information input by the user, the smartphone converts the contact voice information into contact pinyin "zhangsan", and compares the contact pinyin "zhangsan" with the multiple person name pinyins of the contacts in the address book, where the contact pinyin is only successfully matched with the person name pinyin "zhangsan". And (3) the name of the contact person information associated with the unique successfully matched name pinyin zhangsan: zhang III; the mobile phone number is as follows: 1 xxxxxxxxxxxx, identification number: XXXXXXXXXXXXXXXXXX "is respectively filled into the corresponding contact information input boxes 21, as shown in fig. 4, the name" zhang san "in the contact information is filled into the" contact name "input box, and" 1 xxxxxxxxxxxxxx "in the contact information is filled into the" contact phone number "input box; the identification number "XXXXXXXXXXXXXXXXX" in the contact information is filled into the "contact identification number" input box.
In the embodiment of the invention, the target contact is obtained by acquiring the contact voice information input by the user and identifying the contact pinyin to compare with the prestored name pinyin, so that the contact information of the target contact is filled in the contact information input frame.
On the basis of the embodiment shown in fig. 1, optionally, after comparing the contact pinyin with the pre-stored person name pinyin in step 102, as shown in fig. 5, steps 501 to 504 are further included.
S501, if the contact person pinyin is successfully matched with the multiple person name pinyins, the contact person voice information is converted into the contact person pinyin with tones.
In step 102, the contact person pinyin is compared with the pre-stored person name pinyin, and the pre-stored person name pinyin is traversed to obtain a comparison result. Because the Chinese has many characters with the same pinyin, the pinyin of the name corresponding to different contact persons may have the same condition, and at the moment, the pinyin of the name successfully matched with the pinyin of the contact person has a plurality of pinyin. In order to further obtain a prepared matching result, the contact person voice information is further converted into the contact person pinyin with the tone.
Optionally, the step 501 of converting the contact voice information into a contact pinyin with tones, as shown in fig. 6, includes steps 601 to 605. A more accurate speech recognition is achieved by steps 601 to 605.
S601, extracting the acoustic feature vector of the voice information of the contact person.
And S602, analyzing the acoustic feature vector by using the trained acoustic model to obtain the contact person pinyin.
Steps 601 to 602 are the same as steps 301 and 302, and are not described herein again.
It should be noted that in other embodiments of the present invention, the contact pinyin results obtained in steps 301 and 302 may be directly used, and at this time, the steps 601 and 602 need not be repeated.
S603, intercepting a final sound voice segment corresponding to a final sound in the pinyin of the contact in the voice information of the contact.
Because the tone exists only in the vowel, in the embodiment of the invention, after the contact person pinyin corresponding to the contact person voice information is identified, the vowel voice segment of the vowel in the contact person voice information is intercepted, and the tone of the contact person pinyin is obtained based on the tone.
S604, extracting the fundamental frequency of the final voice segment.
And S605, determining the tone corresponding to the final voice segment according to the track of the fundamental frequency.
The fundamental frequency of the final speech segment is extracted through a preset sampling mode, so that the track of the fundamental frequency is obtained through the set of a plurality of sampling points, and the tone of the final is determined through the track. The present invention does not specifically limit the sampling method.
Specifically, step 605 includes: and obtaining a track curve according to the track of the fundamental frequency, respectively calculating the similarity value of the track curve and each preset track curve, and determining the tone corresponding to the preset track curve with the minimum similarity value as the tone corresponding to the final voice segment.
In the embodiment of the invention, a track curve is obtained by fitting the fundamental frequencies of a plurality of sampling points. The preset track curves are five different tones, including a first tone (sound level), a second tone (yang level), a third tone (sound up), a fourth tone (sound down) and a fifth tone (light sound), which are respectively corresponding to the base frequency track curves, so that the following formula is used:
and respectively calculating the similarity value C of the track curve Y ═ f (x) and each preset track curve Y ═ F (X).
Wherein,
the trace curve y ═ f (x) is (y)1,y2,y3,......,yi)=f(x1,x2,x3,......,xi) (ii) a The predetermined trajectory curve Y ═ F (X) is(Y1,Y2,Y3,......,Yi)=f(X1,X2,X3,......,Xi) (ii) a i takes the value from 1 to n; i and n are both positive integers.
In other embodiments of the present invention, the preset trajectory curves may also be four different tones, including a first tone (sound level), a second tone (yang level), a third tone (sound up), and a fourth tone (sound down), which correspond to the fundamental frequency trajectory curves respectively.
S502, comparing the pinyin of the contact person with the tone with the pre-stored pinyin of the name of the contact person with the tone.
In the embodiment of the invention, after the voice of the contact person with tone is obtained in step 501, the pinyin of the contact person with tone is compared with the pre-stored pinyin of the name of the contact person with tone. The name pinyin with tones is the name pinyin with tones corresponding to the contacts in the contact list of the address book, and the name pinyin with tones is prestored in a memory of the terminal device, so that when matching is needed, comparison results are obtained by traversing all the prestored name pinyins with tones.
It should be noted that the contact list includes contacts and contact information associated with each contact. The contact information includes, but is not limited to, a name, a cell phone number, a work landline number, a home landline number, a mailbox number, an instant messaging account number, or the like.
S503, if the pinyin of the contact person with the tone is successfully matched with only one pinyin of the name of the contact person with the tone, filling the information of the contact person associated with the pinyin of the name of the contact person with the tone which is successfully matched into a corresponding information input box of the contact person.
S504, if the pinyin of the contact person with the tone is successfully matched with the pinyin of the names of the multiple persons with the tone, displaying the names of the contact persons associated with the pinyin of the names of the multiple persons with the tone which is successfully matched; and if an event that the user triggers and selects the contact name is monitored, taking the contact name selected by the user as a target contact, and filling the contact information associated with the target contact into a corresponding contact information input box.
In the embodiment of the invention, if the pinyin of the contact person with tone is successfully matched with the pinyin of the names of the multiple persons with tone, the names of the contact persons associated with the pinyin of the names of the multiple persons with tone successfully matched are displayed for the user to select. And when an event that the user triggers and selects the contact name is monitored, and at the moment, the user selects a target contact from the contact names which are successfully matched, and filling the contact information associated with the target contact into the corresponding contact information input box.
The event of selecting a contact name is an event of selecting a target contact for a user, and is usually triggered by a preset operation executed by the user. The preset operation comprises mouse operation or touch operation, for example, a selection control is clicked on the touch screen through a mouse or a finger, or a target contact is selected by drawing a closed graph on the touch screen through the mouse or the finger, and the like.
Exemplarily, when the contact person pinyin "zhangsan" is successfully matched with the multiple person name pinyin "zhangsan" in the address list, the contact person voice information is converted into the contact person pinyin with toneAnd matching the contact person pinyin with the tone with the name pinyin with the tone.
Contact person pinyin with toneOnly with one name phonetic with toneAnd if the matching is successful, the contact person information 'name' corresponding to the successfully matched name pinyin with the tone is: zhang III; telephone: 1 xxxxxxxxxxxx, identification number: xxxxxxxxxxxxxxxx "fills in the corresponding contact information entry box, as shown with continued reference to fig. 4.
Contact person pinyin with tonePhonetic with multiple names with tonesIf the matching is successful, as shown in fig. 7, the target contact person selection interface 71 is loaded, and the target contact person selection interface 71 displays the contact person names "zhang san" and "zhang san" corresponding to the successfully matched name pinyin with the tone for the user to select.
If the user selects the target contact person "zhang san" on the target contact person selection interface 71, the name of the target contact person information "corresponding to the target contact person" zhang san "is: zhang III; telephone: 1 xxxxxxxxxxxx, identification number: xxxxxxxxxxxxxxxx "fills in the corresponding contact information entry box, as shown with continued reference to fig. 4.
It should be noted that, if the pinyin of the contact person is not successfully matched with the pinyin of any person name, the user is prompted to re-input the voice information of the contact person by voice, and the step flow of the method for inputting the contact person in the embodiment of the present invention is continuously executed. In addition, if the pinyin of the contact person with the tone is not successfully matched with the pinyin of the name of any contact person with the tone, prompting the user to input the voice information of the contact person again by voice, and continuing to execute the step flow of the contact person input method in the embodiment of the invention.
It should be understood that, the sequence numbers of the steps in the foregoing embodiments do not imply an execution sequence, and the execution sequence of each process should be determined by its function and inherent logic, and should not constitute any limitation to the implementation process of the embodiments of the present invention.
Fig. 8 shows a block diagram of a device for inputting contact information according to an embodiment of the present invention, which corresponds to the method for inputting contact information according to the above embodiment, and only shows the relevant parts according to the embodiment of the present invention for convenience of description.
Referring to fig. 8, the contact information input apparatus includes:
the monitoring and receiving module 81 is used for monitoring a command for acquiring the voice information of the contact person triggered by the user and receiving the voice information of the contact person input by the voice of the user;
a conversion and comparison module 82, configured to convert the contact voice information into contact pinyin, and compare the contact pinyin with a pre-stored name pinyin;
and the first filling module 83 is configured to, if the contact person pinyin is successfully matched with only one contact person pinyin, fill the contact person information associated with the successfully matched contact person pinyin into the corresponding contact person information input box.
Optionally, the apparatus for inputting contact information further includes:
the second conversion module is used for converting the contact person voice information into contact person pinyin with tones if the contact person pinyin is successfully matched with the multiple person name pinyins;
the second comparison module is used for comparing the contact person pinyin with the tone with the prestored name pinyin with the tone;
the second filling module is used for filling the contact person information associated with the successfully matched name pinyin with the tone into the corresponding contact person information input box if the pinyin of the contact person with the tone is successfully matched with the name pinyin with the tone;
the third filling module is used for displaying the contact names related to the successfully matched multiple tone name pinyins if the tone-bearing contact person pinyins are successfully matched with the multiple tone name pinyins; and if an event that the user triggers and selects the contact name is monitored, taking the contact name selected by the user as a target contact, and filling the contact information associated with the target contact into a corresponding contact information input box.
Optionally, the conversion and alignment module 82 is specifically configured to:
extracting acoustic feature vectors of the contact person voice information;
and analyzing the acoustic feature vector by using the trained acoustic model to obtain the contact person pinyin.
Optionally, the second conversion module is specifically configured to:
extracting acoustic feature vectors of the contact voice information,
analyzing the acoustic feature vector by using a trained acoustic model to obtain contact person pinyin;
intercepting a final sound voice segment corresponding to a final sound in the pinyin of the contact person in the contact person voice information;
extracting the fundamental frequency of the vowel voice segment;
and determining the tone corresponding to the final voice segment according to the track of the fundamental frequency.
Optionally, the determining the tone corresponding to the final speech segment according to the track of the fundamental frequency includes:
and obtaining a track curve according to the track of the fundamental frequency, respectively calculating the similarity value of the track curve and each preset track curve, and determining the tone corresponding to the preset track curve with the minimum similarity value as the tone corresponding to the final voice segment.
Optionally, the separately calculating the similarity value between the trajectory curve and each preset trajectory curve includes:
by the formulaRespectively calculating similarity values C of the track curve Y ═ f (x) and each preset track curve Y ═ F (X);
wherein,
the trace curve y ═ f (x) is (y)1,y2,y3,......,yi)=f(x1,x2,x3,......,xi) (ii) a The preset track curve Y is (Y) F (X)1,Y2,Y3,......,Yi)=f(X1,X2,X3,......,Xi) (ii) a i takes the value from 1 to n; i and n are both positive integers.
Fig. 9 is a schematic diagram of a terminal device according to an embodiment of the present invention. As shown in fig. 9, the terminal device 9 of this embodiment includes: a processor 90, a memory 91 and a computer program 92, such as a program for contact information entry, stored in said memory 91 and operable on said processor 90. The processor 90, when executing the computer program 92, implements the steps in the above-described method embodiment of inputting contact information, such as steps S101 to S103 shown in fig. 1. Alternatively, the processor 90 implements the functions of the modules in the device embodiments described above, such as the functions of the modules 81 to 83 shown in fig. 8, when executing the computer program 92.
Illustratively, the computer program 92 may be partitioned into one or more modules/units that are stored in the memory 91 and executed by the processor 90 to implement the present invention. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, which are used to describe the execution process of the computer program 92 in the terminal device 9.
The terminal device 9 may be a personal computer, a portable intelligent device, or the like. The terminal device 9 may include, but is not limited to, a processor 90 and a memory 91. Those skilled in the art will appreciate that fig. 9 is only an example of a terminal device 9, and does not constitute a limitation to the terminal device 9, and may include more or less components than those shown, or combine some components, or different components, for example, the terminal device may also include an input-output device, a network access device, a bus, etc.
The Processor 90 may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field-Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The memory 91 may be an internal storage unit of the terminal device 9, such as a hard disk or a memory of the terminal device 9. The memory 91 may also be an external storage device of the terminal device 9, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), and the like, which are provided on the terminal device 9. Further, the memory 91 may also include both an internal storage unit and an external storage device of the terminal device 5. The memory 91 is used for storing the computer program and other programs and data required by the terminal device. The memory 91 may also be used to temporarily store data that has been output or is to be output.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-mentioned division of the functional units and modules is illustrated, and in practical applications, the above-mentioned function distribution may be performed by different functional units and modules according to needs, that is, the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-mentioned functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present application. The specific working processes of the units and modules in the system may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated modules/units, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. Based on such understanding, all or part of the flow of the method according to the embodiments of the present invention may also be implemented by a computer program, which may be stored in a computer-readable storage medium, and when the computer program is executed by a processor, the steps of the method embodiments may be implemented.
The above-mentioned embodiments are only used for illustrating the technical solutions of the present invention, and not for limiting the same; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; such modifications and substitutions do not substantially depart from the spirit and scope of the embodiments of the present invention, and are intended to be included within the scope of the present invention.

Claims (10)

1. A method of contact information entry, comprising:
monitoring a command for acquiring the voice information of the contact person triggered by a user, and receiving the voice information of the contact person input by the voice of the user;
converting the contact person voice information into contact person pinyin, and comparing the contact person pinyin with prestored person name pinyin;
and if the contact person pinyin is successfully matched with one contact person pinyin, filling the contact person information associated with the successfully matched contact person pinyin into a corresponding contact person information input box.
2. The method of claim 1, wherein after comparing the contact pinyin to pre-stored person name pinyin, further comprising:
if the contact person pinyin is successfully matched with the multiple person name pinyins, converting the contact person voice information into the contact person pinyin with tones;
comparing the contact person pinyin with tones with the prestored name pinyin with tones;
if the pinyin of the contact person with the tone is successfully matched with only one pinyin of the name of the contact person with the tone, filling the information of the contact person associated with the successfully matched pinyin of the name of the contact person with the tone into a corresponding information input box of the contact person;
if the pinyin of the contact person with the tone is successfully matched with the pinyin of the names of the multiple persons with the tone, displaying the names of the contact persons associated with the pinyin of the names of the multiple persons with the tone which is successfully matched; and if an event that the user triggers and selects the contact name is monitored, taking the contact name selected by the user as a target contact, and filling the contact information associated with the target contact into a corresponding contact information input box.
3. The method of claim 1, wherein the converting the contact voice information to contact pinyin comprises:
extracting acoustic feature vectors of the contact person voice information;
and analyzing the acoustic feature vector by using the trained acoustic model to obtain the contact person pinyin.
4. The method of claim 2, wherein converting the contact voice information into a toned contact pinyin comprises:
extracting acoustic feature vectors of the contact voice information,
analyzing the acoustic feature vector by using a trained acoustic model to obtain contact person pinyin;
intercepting a final sound voice segment corresponding to a final sound in the pinyin of the contact person in the contact person voice information;
extracting the fundamental frequency of the vowel voice segment;
and determining the tone corresponding to the final voice segment according to the track of the fundamental frequency.
5. The method of claim 4, wherein the determining the tone corresponding to the final speech segment according to the trace of the fundamental frequency comprises:
and obtaining a track curve according to the track of the fundamental frequency, respectively calculating the similarity value of the track curve and each preset track curve, and determining the tone corresponding to the preset track curve with the minimum similarity value as the tone corresponding to the final voice segment.
6. The method of claim 5, wherein said separately calculating a similarity value of said trajectory profile to each of the predetermined trajectory profiles comprises:
by the formulaRespectively calculating similarity values C of the track curve Y ═ f (x) and each preset track curve Y ═ F (X);
wherein,
the trace curve y ═ f (x) is (y)1,y2,y3,......,yi)=f(x1,x2,x3,......,xi) (ii) a The preset track curve Y is (Y) F (X)1,Y2,Y3,......,Yi)=f(X1,X2,X3,......,Xi) (ii) a i takes the value from 1 to n; i and n are both positive integers.
7. The method according to claim 3 or 4, wherein the acoustic feature vector is a feature vector including at least two acoustic features of linear prediction cepstral coefficients and mel-frequency cepstral coefficients.
8. An apparatus for contact information input, comprising:
the monitoring and receiving module is used for monitoring a command for acquiring the voice information of the contact person triggered by the user and receiving the voice information of the contact person input by the voice of the user;
the conversion and comparison module is used for converting the contact person voice information into contact person pinyin and comparing the contact person pinyin with prestored person name pinyin;
and the first filling module is used for filling the contact information associated with the successfully matched name pinyin into the corresponding contact information input box if the contact pinyin is successfully matched with only one name pinyin.
9. A terminal device comprising a memory and a processor, the memory having stored thereon a computer program operable on the processor, wherein the processor, when executing the computer program, performs the steps of the method according to any one of claims 1 to 7.
10. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 7.
CN201910039493.2A 2019-01-16 2019-01-16 The method, apparatus and terminal device of contact information input Pending CN109688271A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910039493.2A CN109688271A (en) 2019-01-16 2019-01-16 The method, apparatus and terminal device of contact information input

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910039493.2A CN109688271A (en) 2019-01-16 2019-01-16 The method, apparatus and terminal device of contact information input

Publications (1)

Publication Number Publication Date
CN109688271A true CN109688271A (en) 2019-04-26

Family

ID=66192458

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910039493.2A Pending CN109688271A (en) 2019-01-16 2019-01-16 The method, apparatus and terminal device of contact information input

Country Status (1)

Country Link
CN (1) CN109688271A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112364212A (en) * 2020-11-04 2021-02-12 北京致远互联软件股份有限公司 Voice name recognition method based on approximate voice recognition
CN113241078A (en) * 2021-05-10 2021-08-10 杭州魔点科技有限公司 Attendance machine-based voice recognition method and system
CN113593569A (en) * 2021-07-27 2021-11-02 德邦物流股份有限公司 Electronic bill generation method and device, electronic equipment and storage medium
CN116010562A (en) * 2023-03-28 2023-04-25 之江实验室 A name matching method, device, equipment and medium based on multiple data sources

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102164205A (en) * 2011-03-02 2011-08-24 宇龙计算机通信科技(深圳)有限公司 Quick input method of contact person information and mobile terminal
CN102263855A (en) * 2011-06-03 2011-11-30 宇龙计算机通信科技(深圳)有限公司 Method, system, mobile terminal and browser for inputting contact information
CN105607799A (en) * 2015-12-15 2016-05-25 深圳市永兴元科技有限公司 Data processing method and apparatus
CN107016994A (en) * 2016-01-27 2017-08-04 阿里巴巴集团控股有限公司 The method and device of speech recognition
CN107357772A (en) * 2017-07-04 2017-11-17 贵州小爱机器人科技有限公司 List filling method, device and computer equipment
US20180109675A1 (en) * 2016-09-15 2018-04-19 International Business Machines Corporation Multi-dimensional auditory notification composition based on sender categorizations
CN108595141A (en) * 2018-03-22 2018-09-28 珠海市魅族科技有限公司 Pronunciation inputting method and device, computer installation and computer readable storage medium
CN108874872A (en) * 2018-04-26 2018-11-23 深圳市艾塔文化科技有限公司 A kind of speech recognition string processing comparison method based on phonetic
CN108881383A (en) * 2018-05-07 2018-11-23 深圳壹账通智能科技有限公司 Address book data processing method, device, computer equipment and storage medium

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102164205A (en) * 2011-03-02 2011-08-24 宇龙计算机通信科技(深圳)有限公司 Quick input method of contact person information and mobile terminal
CN102263855A (en) * 2011-06-03 2011-11-30 宇龙计算机通信科技(深圳)有限公司 Method, system, mobile terminal and browser for inputting contact information
CN105607799A (en) * 2015-12-15 2016-05-25 深圳市永兴元科技有限公司 Data processing method and apparatus
CN107016994A (en) * 2016-01-27 2017-08-04 阿里巴巴集团控股有限公司 The method and device of speech recognition
US20180109675A1 (en) * 2016-09-15 2018-04-19 International Business Machines Corporation Multi-dimensional auditory notification composition based on sender categorizations
CN107357772A (en) * 2017-07-04 2017-11-17 贵州小爱机器人科技有限公司 List filling method, device and computer equipment
CN108595141A (en) * 2018-03-22 2018-09-28 珠海市魅族科技有限公司 Pronunciation inputting method and device, computer installation and computer readable storage medium
CN108874872A (en) * 2018-04-26 2018-11-23 深圳市艾塔文化科技有限公司 A kind of speech recognition string processing comparison method based on phonetic
CN108881383A (en) * 2018-05-07 2018-11-23 深圳壹账通智能科技有限公司 Address book data processing method, device, computer equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
白俊贤: "大词汇量语音匹配方法研究", 《中国优秀硕士学位论文全文数据库》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112364212A (en) * 2020-11-04 2021-02-12 北京致远互联软件股份有限公司 Voice name recognition method based on approximate voice recognition
CN113241078A (en) * 2021-05-10 2021-08-10 杭州魔点科技有限公司 Attendance machine-based voice recognition method and system
CN113593569A (en) * 2021-07-27 2021-11-02 德邦物流股份有限公司 Electronic bill generation method and device, electronic equipment and storage medium
CN116010562A (en) * 2023-03-28 2023-04-25 之江实验室 A name matching method, device, equipment and medium based on multiple data sources
CN116010562B (en) * 2023-03-28 2023-07-07 之江实验室 A name matching method, device, equipment and medium based on multiple data sources

Similar Documents

Publication Publication Date Title
CN110310623B (en) Sample generation method, model training method, device, medium, and electronic apparatus
US11508366B2 (en) Whispering voice recovery method, apparatus and device, and readable storage medium
JP6429945B2 (en) Method and apparatus for processing audio data
JP6099556B2 (en) Voice identification method and apparatus
CN109360572B (en) Call separation method and device, computer equipment and storage medium
CN107665705A (en) Voice keyword recognition method, device, equipment and computer-readable recording medium
CN108269575B (en) Voice recognition method for updating voiceprint data, terminal device and storage medium
CN109688271A (en) The method, apparatus and terminal device of contact information input
CN109801646B (en) Voice endpoint detection method and device based on fusion features
CN113327575A (en) Speech synthesis method, device, computer equipment and storage medium
CN106210239A (en) The maliciously automatic identifying method of caller's vocal print, device and mobile terminal
CN111028845A (en) Multi-audio recognition method, device, equipment and readable storage medium
CN111640421B (en) Speech comparison method, device, equipment and computer readable storage medium
CN113129867B (en) Training method of voice recognition model, voice recognition method, device and equipment
CN109947971B (en) Image retrieval method, image retrieval device, electronic equipment and storage medium
CN110827803A (en) Method, device and equipment for constructing dialect pronunciation dictionary and readable storage medium
CN106341539A (en) Automatic evidence obtaining method of malicious caller voiceprint, apparatus and mobile terminal thereof
CN112669842A (en) Man-machine conversation control method, device, computer equipment and storage medium
CN112201275A (en) Voiceprint segmentation method, voiceprint segmentation device, voiceprint segmentation equipment and readable storage medium
CN112037772A (en) Multi-mode-based response obligation detection method, system and device
CN107731232A (en) Voice translation method and device
CN111402617B (en) Site information determination method, device, terminal and storage medium
CN113035230B (en) Authentication model training method and device and electronic equipment
CN108989551B (en) Position prompting method and device, storage medium and electronic equipment
CN111640450A (en) Multi-person audio processing method, device, equipment and readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190426

RJ01 Rejection of invention patent application after publication