[go: up one dir, main page]

CN109887498A - Scoring method of polite expressions at highway entrances - Google Patents

Scoring method of polite expressions at highway entrances Download PDF

Info

Publication number
CN109887498A
CN109887498A CN201910181668.3A CN201910181668A CN109887498A CN 109887498 A CN109887498 A CN 109887498A CN 201910181668 A CN201910181668 A CN 201910181668A CN 109887498 A CN109887498 A CN 109887498A
Authority
CN
China
Prior art keywords
model
keyword
probability
keywords
file
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910181668.3A
Other languages
Chinese (zh)
Inventor
卢朝阳
周云蝶
李静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN201910181668.3A priority Critical patent/CN109887498A/en
Publication of CN109887498A publication Critical patent/CN109887498A/en
Pending legal-status Critical Current

Links

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

本发明公开了一种高速公路口礼貌用语评分方法,主要解决现有人工监督收费员时重复枯燥且易疏忽的问题。其实现方案为:对语料库文件进行预处理,完成24维MFCC特征参数提取,并对特征参数进行训练,得到关键词和Filler并行的网络模型;对测试语音完成预处理和特征提取后,得到测试语音特征参数,并对测试语音特征参数与得到的网络模型进行匹配,得到初始检索结果;对初始检索结果与孤立词模型进行匹配,得到最终检索结果,若检索结果中包含所有关键词,则判为100分,否则,缺y个关键词,则判为100‑y*100/m分。本发明具有良好的鲁棒性,且关键词检索的正确率较高,误识率较低,适用于高速公路口管理。

The invention discloses a method for grading polite words at an expressway entrance, which mainly solves the problems that the existing manual supervision of toll collectors is boring and easy to neglect. The implementation scheme is: preprocessing the corpus file, completing the extraction of 24-dimensional MFCC feature parameters, and training the feature parameters to obtain a parallel network model of keywords and Filler; Voice feature parameters, and match the test voice feature parameters with the obtained network model to obtain the initial search results; match the initial search results with the isolated word model to obtain the final search results, if the search results contain all keywords, then judge 100 points, otherwise, if y keywords are missing, it will be judged as 100‑y*100/m points. The invention has good robustness, and the keyword retrieval accuracy rate is high, and the misrecognition rate is low, and is suitable for expressway crossing management.

Description

Highway mouth term of courtesy methods of marking
Technical field
The invention belongs to voice keyword retrieval technical field, in particular to a kind of term of courtesy methods of marking can be used for Highway mouth charge station.
Technical background
Duplicate work simple in the mankind is studied by machine replacement always people the initial power of robotic development, people Exchange with machine be current artificial intelligence one of demand for development.The mankind and machine directly " dialogue " are realized as a kind of Technology, speech recognition technology can advantageously convert voice signals into corresponding machine language very much, and then realize accessible friendship Stream.
Present human work life in, some needs of work by detection staff term of courtesy standard with It is no come evaluation work situation, such as highway mouth charge station staff just need to finish as defined in certain term of courtesy sides It is just up to standard.And the repetition class work of these testing and evaluations just alleviates manager once being replaced by machine to a certain extent Work load and improve management effect.Thus realize that the keyword speech recognition under some scenes seems especially useful, and It can score the evaluated personnel of unspecified person.
China has 5,000 years of civilization history, is known as the title of " state of ceremonies ", the Chinese nation is also with urbane style and features And it is world-famous for.An important component of the good social civility as Chinese traditional culture, content very abundant, the model being related to Enclose it is very extensive, almost permeate in society various aspects.Such as the staff of highway mouth charge station with passing department Machine just needs some terms of courtesy when exchanging, the staff of charge station whether using specific term of courtesy and using frequency be Administrator assesses the important evidence of their work.
Existing supervision evaluation work has been undertaken by video monitoring in behavior exchange, such as big magnificent highway video prison Charge station's subsystem in control system monitors the working condition of fee-collector in tollbooth, but this method with regard to continuous 24 hours whole days The courtesy movement that can only monitor fee-collector, the supervision assessment in speech exchange is still to need administrator by being accomplished manually Whole process supervision, process repetition is uninteresting, but also needs specially to be arranged for each tollbooth the position of administrator, lavishes labor on.
Summary of the invention
It is an object of the invention in view of the above shortcomings of the prior art, propose a kind of highway mouth term of courtesy scoring System, to realize the intelligence to fee-collector's voice monitoring, the supervision and assessment that permit ease of administration person works to fee-collector.
To achieve the above object, the present invention includes:
(1) select m term of courtesy of highway mouth fee-collector as keyword, selection n people as enunciator, everyone It is complete to each keyword and clearly say x times, it always there are m × n × x wav file as corpus library file;
(2) keyword models and the parallel network model of Filler model are constructed:
Preemphasis successively 2a) is carried out to the corpus library file of each keyword, framing adds the pretreatment of Hamming window, obtain one The voice data of one frame of frame extracts 24 Jan Vermeer frequency cepstral coefficient MFCC as characteristic parameter from the voice data;Using Baum-Welch algorithm is trained this feature parameter, obtains the Hidden Markov Model HMM parameter model of the keyword;
2b) using the predictable non-courtesy speech syllable of highway as non-key word, use and 2a) identical method establishes Non-key word HMM model;With method identical with 2a) to the single state HMM model of mute foundation, with non-key word model and mute Model forms Filler model;
Keyword models and Filler model are arranged parallel 2c), form the network model without linguistic constraints;
(3) k people is chosen as test speaker person, everyone says one to the m voice segments comprising 1 to m keyword respectively Time, k × m is always obtained!Wav file, as tone testing file;
(4) to tone testing file pass through and 2a) it is identical pretreatment and MFCC feature extraction, obtain tested speech feature Parameter;After the weight for adjusting network edge in the parallel network model of (2) resulting keyword models and Filler model, use Viterbi algorithm calculates the matching score of the tested speech characteristic parameter pair and each model in network model, retains matching The higher s model of score, as keyword initial retrieval result;
(5) Viterbi algorithm is used, the higher model of s score and 2a in (4) resulting network model are calculated) institute Keyword models matching score, temporally length is to s matching score normalization, and using result as corresponding to s S confidence level of a model;One threshold value is set, the confidence level of more each model and the size of the threshold value are recycled, it is s times total, It is higher than the threshold value lower than the model, confidence level is discarded if the threshold value if confidence level and just retains the model, which is just used as finally Keyword retrieval result;
(6) by the random tone testing file of someone in file obtained by (3) behind (4) and (5), if including institute M keyword of retrieval in need, then be judged to 100 points, if lacking y keyword, is judged to 100-y*100/m and divides, final To being commented the work of personnel to score.
The present invention has the advantages that
1) application scenarios of the invention are the charge stations of highway mouth, build a set of term of courtesy points-scoring system, are realized Intelligence to fee-collector's voice monitoring, supervision evaluation work of the permit ease of administration person to fee-collector;
2) present invention uses the voice keyword retrieval method based on HMM, has good robustness;
3) the present invention is based on after keyword initial retrieval as a result, realizing keyword recognition plus using confidence level, has Higher retrieval accuracy;
4) it is that personnel is commented to score that the present invention, which strictly observes code of points, does not miss a term of courtesy, and leakage knowledge rate is lower.
Detailed description of the invention
Fig. 1 is implementation flow chart of the invention;
Fig. 2 is the parallel network illustraton of model of keyword models and Filler model in the present invention.
Specific embodiment
With reference to the accompanying drawings and detailed description, the present invention is furture elucidated.
Referring to Fig.1, specific step is as follows for the present embodiment:
Step 1. acquires corpus library file.
Selected m term of courtesy of highway mouth fee-collector chooses n (n >=20) personal accomplishment enunciator as keyword, Everyone is complete to each keyword and clearly says x times, always there are m × n × x wav file as corpus library file.
Step 2. constructs keyword and the parallel network model of Filler.
Preemphasis successively 2a) is carried out to the corpus library file of each keyword, framing adds the pretreatment of Hamming window:
Preemphasis 2a1) is carried out to original signal x (n) with single order high-pass digital filter, obtains preemphasis treated letter Number are as follows:
Y (n)=x (n) -0.98x (n-1);
Framing 2a2) is carried out to preemphasis treated signal y (n) and adds Hamming window, framing is obtained and adds Hamming window treated Signal are as follows:
2b) through 2a) pretreatment after obtain voice data one by one, extracted from the voice data 24 Jan Vermeers frequency Rate cepstrum coefficient MFCC is as characteristic parameter;
2c) using Baum-Welch algorithm to 2b) resulting characteristic parameter is trained:
2c1) assume observation sequence O={ Ot, t=1,2 ..., T }, initial model λ=(π, A, B), if the initial model State set is { Si, i=1,2 ..., N }, it is q in t moment statust, observe symbol are as follows:
V={ vk, k=1,2 ..., N },
In initial model λ: π={ πi, i=1,2 ..., N },
A={ aij, i=1,2 ..., N, j=1,2 ..., N },
B={ bjk, j=1,2 ..., N, k=1,2 ..., M },
πiIndicate initial state probabilities, aijIndicate that in the state of moment t be SiAnd it shifts at the t+1 moment as state SjIt is general Rate, bjkIt indicates in state SjObserve symbol vkProbability;
2c2) in 2c1) hypothesis under, introduce two groups of probability variable εt(i, j) indicates that t moment is in state SiAnd the t+1 moment In state SjProbability, γt(i) indicate that t moment is in state SiProbability, it may be assumed that
εt(i, j)=P (qt=Si,qt+1=Sj| O, λ),
γt(i)=P (O, qt=Si|λ);
2c3) by 2c2) introduced two groups of probability variables calculate one group of new parameter:
π′i1(i),
2c4) by 2c3) resulting one group of new parameter π 'i, a 'ij, b 'j(k), revaluation obtains a new model:
λ '=(π ', A ', B '),
The probability P (O | λ ') of model λ ' generation observation sequence generates the probability P (O | λ) of observation sequence than initial model λ It is big;
2c5) repeat 2c3) and 2c4), model parameter is continuously improved, until P (O | λ ') is no longer significantly increased, model at this time λ '=(π ', A ', B ') is the Hidden Markov Model HMM parameterized template of the training keyword;
2d) using the predictable non-courtesy speech syllable of highway as non-key word, use and 2a) identical method establishes Non-key word HMM model;With method identical with 2a) to the single state HMM model of mute foundation, with non-key word model and mute Model forms Filler model;
Keyword models and Filler model are arranged parallel 2e), the network model without linguistic constraints are formed, such as Fig. 2 institute Show.
Step 3. acquires tone testing file.
K personal accomplishment test speaker person is chosen, everyone says one to the m voice segments comprising 1 to m keyword respectively Time, k × m is always obtained!Wav file, as tone testing file, wherein k > 5.
Step 4. keyword initial retrieval.
4a) tone testing file is successively passed through and 2a) identical pretreatment and and 2b) identical MFCC feature extraction, Obtain tested speech characteristic parameter;
After the weight for 4b) adjusting network edge in the resulting parallel network model of step 2, using Viterbi algorithm, meter Calculate 4a) resulting tested speech characteristic parameter is to the matching score of each model in network model:
4b1) under hypothesis identical with 2c1), if moment t is along a paths sequence Q={ q1,q2,…,qtAnd qt=Si The maximum probability for generating observation sequence is δt(Si), introduce one group of intermediate variable
4b2) initialize 4b1) set by probability variable δt(Si) and intermediate variableAre as follows:
Moment t 4b3) is set along a paths sequence Q={ q1,q2,…,qtAnd qt=SjGenerate the maximum probability of observation sequence For δt(Sj), introduce one group of intermediate variableIn 4a2) gained probability variable δt(Si) and intermediate variableBasis On, obtain the maximum probability δ of observation sequencet(Sj) and intermediate variableAre as follows:
4b4) according to 4b3) resulting one group of intermediate variableRecursive calculation
4b5) by 4b3) resulting δT(Si), calculate the observation sequence at T moment and probability P ' (Q, the O | λ) of Model Matching and State q 'T:
P ' (Q, O | λ)=max1≤i≤NT(Si)],
q′T=argmax1≤i≤NT(Si)],
P ' (Q, O | λ) is the matching score of observation sequence and model at this time;
4b6) merge 4b4) resulting one group of q '1,q′2,…,q′T-1And 4b4) resulting q 'T, obtain optimum state path Sequence:
Q '={ q '1,q′2,…,q′T};
4c) retain 4b) in the higher s model of matching score, as keyword initial retrieval result.
Step 5. realizes keyword recognition with confidence level, obtains the final search result of keyword.
5a) use and 4b) identical Viterbi algorithm, calculate 4c) resulting higher keyword models pair of s score The matching score of isolated word model, temporally length is to s matching score normalization, and normalized result respectively as right It should be in s confidence level of s model;
One threshold value 5b) is set, and circulation compares 5a) size of resulting each model confidence and the threshold value, it is s times total, It is higher than the threshold value lower than the model, confidence level is discarded if the threshold value if confidence level and just retains the model, the model of reservation is with regard to conduct The search result of final keyword.
Step 6. completes scoring.
By the random tone testing file of someone in step 3 gained file after step 4 and step 5, if packet Containing retrieval in need m keyword, then be judged to 100 points, if lacking y keyword, be judged to 100-y*100/m point, most Obtain being commented the work of personnel to score eventually, wherein 0≤y≤m.
The above is only example of the present invention, does not constitute any limitation of the invention, it is clear that for It, all may be without departing substantially from the principle of the invention, structure after having understood the content of present invention and principle for one of skill in the art In the case where, carry out various modifications and change in form and details, but these modifications and variations based on inventive concept Still within the scope of the claims of the present invention.

Claims (4)

1.一种高速公路口礼貌用语评分方法,其特征在于,包括:1. A method for scoring polite expressions at a highway entrance, characterized in that, comprising: (1)选定高速公路口收费员m个礼貌用语作为关键词,选取n人作为发音者,每个人对每个关键词完整并清晰地说x遍,总共得m×n×x条WAV文件作为语料库文件;(1) Select m polite expressions of expressway entrance toll collectors as keywords, select n people as speakers, each person speaks each keyword completely and clearly x times, and a total of m×n×x WAV files are obtained as a corpus file; (2)构建关键词模型和Filler模型并行的网络模型:(2) Build a parallel network model of the keyword model and the Filler model: 2a)对每个关键词的语料库文件依次进行预加重、分帧加汉明窗的预处理,得到一帧一帧的语音数据,从该语音数据中提取24维梅尔频率倒谱系数MFCC作为特征参数;采用Baum-Welch算法对该特征参数进行训练,得到该关键词的隐马尔科夫模型HMM参数模型;2a) Pre-emphasis, frame-by-frame and Hamming window preprocessing are carried out successively to the corpus file of each keyword to obtain the speech data of one frame and one frame, and the 24-dimensional Mel frequency cepstral coefficient MFCC is extracted from the speech data as Feature parameters; use Baum-Welch algorithm to train the feature parameters to obtain the hidden Markov model HMM parameter model of the keyword; 2b)将高速公路可预测的非礼貌语音音节作为非关键词,用与2a)相同的方法建立非关键词HMM模型;用与2a)相同的方法对静音建立单状态HMM模型,用非关键词模型和静音模型组成Filler模型;2b) Use the predictable impolite speech syllables of the highway as non-keywords, and use the same method as 2a) to build a non-keyword HMM model; use the same method as 2a) to build a single-state HMM model for silence, use the non-keyword The model and the mute model form the Filler model; 2c)将关键词模型和Filler模型并行设置,组成无语法约束的网络模型;2c) Set the keyword model and Filler model in parallel to form a network model without grammatical constraints; (3)选取k人作为测试发音者,每个人分别对包含1到m个关键词的m个语音段说一遍,总共得到k×m!条WAV文件,作为语音测试文件;(3) Select k people as test speakers, each person speaks m speech segments containing 1 to m keywords, and a total of k × m is obtained! A WAV file as a voice test file; (4)对语音测试文件经过与2a)相同的预处理和MFCC特征提取,得到测试语音特征参数;在(2)所得的关键词模型和Filler模型并行的网络模型中调整网络边的权重后,采用Viterbi算法,计算该测试语音特征参数与网络模型中每一个模型的匹配得分,保留匹配得分较高的s个模型,作为关键词初始检索结果;(4) The voice test file is subjected to the same preprocessing and MFCC feature extraction as in 2a) to obtain the test voice feature parameters; Using the Viterbi algorithm, calculate the matching score between the test speech feature parameter and each model in the network model, and retain the s models with higher matching scores as the initial search results of keywords; (5)采用与(4)相同的Viterbi算法,计算出(4)所得的网络模型中s个得分较高的模型与2a)所得关键词模型的匹配分数,按时间长度对s个匹配分数归一化,并把归一化后的结果分别作为对应于s个模型的s个置信度;设置一个阈值,循环比较每个模型的置信度和该阈值的大小,共s次,若置信度低于该阈值则弃掉该模型,置信度高于该阈值就保留该模型,保留的模型就作为最终关键词检索结果;(5) Using the same Viterbi algorithm as (4), calculate the matching scores of the s models with higher scores in the network models obtained in (4) and the keyword models obtained in 2a), and normalize the s matching scores according to the length of time. Normalize, and use the normalized results as the s confidence levels corresponding to the s models; set a threshold, and cyclically compare the confidence level of each model and the size of the threshold, a total of s times, if the confidence level is low The model is discarded at the threshold, the model is retained if the confidence is higher than the threshold, and the retained model is used as the final keyword retrieval result; (6)将(3)所得文件中某个人的随机一条语音测试文件经过(4)和(5)后,若包含所有需要检索的m个关键词,则判为100分,若缺少y个关键词,则判为100-y*100/m分,最终得到被评人员的工作评分。(6) After passing through (4) and (5), a random voice test file of a certain person in the file obtained in (3), if it contains all m keywords that need to be retrieved, it will be judged as 100 points, if there are y keywords missing word, it is judged as 100-y*100/m points, and finally the work score of the evaluated person is obtained. 2.根据权利要求1所述的方法,其特征在于,2a)中对每个关键词的语料库文件依次进行预加重、分帧加汉明窗的预处理,是在假设原始信号为x(n)的条件下,按如下步骤进行:2. method according to claim 1 is characterized in that, in 2a), carry out the pre-emphasis of pre-emphasis successively to the corpus file of each keyword, divide into frames and add the preprocessing of Hamming window, is to assume that the original signal is x (n ), proceed as follows: 2a1)用一阶高通数字滤波器对原始信号x(n)进行预加重,得到预加重处理后的信号为:2a1) Pre-emphasize the original signal x(n) with a first-order high-pass digital filter, and obtain the pre-emphasized signal as: y(n)=x(n)-0.98x(n-1);y(n)=x(n)-0.98x(n-1); 2a2)对预加重处理后的信号y(n)进行分帧加汉明窗,得到分帧加汉明窗处理后的信号为:2a2) Perform frame segmentation and Hamming window processing on the pre-emphasized signal y(n), and obtain the signal after frame segmentation and Hamming window processing as: 3.根据权利要求1所述的方法,其特征在于,2a)中采用Baum-Welch算法对特征参数进行训练,按如下步骤进行:3. method according to claim 1, is characterized in that, adopts Baum-Welch algorithm to carry out training to characteristic parameter in 2a), carries out according to the following steps: 2a3)假设观测序列O={Ot,t=1,2,...,T},初始模型λ=(π,A,B),设该初始模型的状态集合为{Si,i=1,2,...,N},在t时刻所处状态为qt,观测符号为V={vk,k=1,2,...,M},其中π={πi,i=1,2,...,N},A={aij,i=1,2,...,N,j=1,2,...,N},B={bjk,j=1,2,...,N,k=1,2,...,M},πi表示初始状态概率,aij表示在时刻t的状态为Si且在t+1时刻转移为状态Sj的概率,bjk表示在状态Sj观测到符号vk的概率;2a3) Suppose the observation sequence O={O t , t=1, 2, . . . , T}, the initial model λ=(π, A, B), set the state set of the initial model to be {S i , i= 1 , 2 , . i=1,2,...,N},A={a ij ,i=1,2,...,N,j=1,2,...,N},B={b jk , j = 1, 2, . is the probability of state S j , b jk represents the probability of observing symbol v k in state S j ; 2a4)在2a3)的假设下,引入两组概率变量εt(i,j)表示t时刻处于状态Si且t+1时刻处于状态Sj的概率,γt(i)表示t时刻处于状态Si的概率,即有εt(f,j)=P(qt=Si,qt+1=Sj|O,λ),γt(i)=P(O,qt=Si|λ);2a4) Under the assumption of 2a3), two sets of probability variables ε t (i, j) are introduced to represent the probability of being in state S i at time t and in state S j at time t+1, and γ t (i) means being in state at time t. The probability of Si, that is, ε t (f, j )=P(q t =S i , q t+1 =S j |O,λ), γ t (i)=P(O, q t =S i |λ); 2a5)由2a4)所引入的两组概率变量计算一组新参数:2a5) Calculates a new set of parameters from the two sets of probability variables introduced in 2a4): π′i=γ1(i),π′ i1 (i), 2a6)由2a5)所得的一组新参数π′i,a′ij,b′j(k),重估得一个新模型:λ′=(π′,A′,B′),此时模型λ′产生观测序列的概率P(O|λ′)比初始模型λ产生观测序列的概率P(O|λ)要大;2a6) A new set of parameters π′ i , a′ ij , b′ j (k) obtained from 2a5) is re-estimated to obtain a new model: λ′=(π′, A′, B′), at this time the model The probability P(O|λ′) that λ′ produces the observation sequence is larger than the probability P(O|λ) that the initial model λ produces the observation sequence; 2a7)重复2a5)和2a6),不断改进模型参数,直到P(O|λ′)不再明显增大,此时模型λ′=(π′,A′,B′)即为训练所得最终模型。2a7) Repeat 2a5) and 2a6), and continuously improve the model parameters until P(O|λ') no longer increases significantly. At this time, the model λ'=(π', A', B') is the final model obtained by training . 4.根据权利要求1所述的方法,其特征在于,(4)中采用Viterbi算法计算匹配得分,按如下步骤进行:4. method according to claim 1 is characterized in that, adopts Viterbi algorithm to calculate matching score in (4), carries out as follows: 4a1)在与2a3)相同的假设下,设时刻t沿一条路径序列Q={q1,q2,...,qt}且qt=Si产生观测序列的最大概率为δt(Si),引入一组中间变量 4a1) Under the same assumptions as 2a3), set time t along a path sequence Q={q 1 , q 2 ,..., q t } and q t =S i The maximum probability of producing an observation sequence is δ t ( S i ), introducing a set of intermediate variables 4a2)初始化4a1)所设的概率变量δt(Si)和中间变量为:4a2) Initialize the probability variable δ t (S i ) and the intermediate variable set in 4a1) for: 4a3)设时刻t沿一条路径序列Q={q1,q2,...,qt}且qt=Sj产生观测序列的最大概率为δt(Sj),引入一组中间变量在4a2)所得概率变量δt(Si)和中间变量的基础上,得到观测序列的最大概率δt(Sj)和中间变量为:4a3) Set time t along a path sequence Q={q 1 , q 2 ,..., q t } and q t =S j to generate the maximum probability of an observation sequence as δ t (S j ), and introduce a set of intermediate variables The probability variable δ t (S i ) obtained in 4a2) and the intermediate variable On the basis of , the maximum probability δ t (S j ) of the observation sequence and the intermediate variables are obtained for: 4a4)根据4a3)所得的一组中间变量递归计算一组q′1,q′2,...,q′T-14a4) A set of intermediate variables obtained according to 4a3) Recursively compute a set of q' 1 , q' 2 , ..., q' T-1 : 4a5)由4a3)所得的δT(Si),计算T时刻的观测序列与模型匹配的概率P′(Q,O|λ)和状态q′T4a5) Calculate the probability P'(Q, O|λ) and state q' T that the observation sequence at time T matches the model from the δ T (S i ) obtained in 4a3): P′(Q,O|λ)=max1≤i≤NT(Si)],P'(Q, O|λ)=max 1≤i≤NT (S i )], q′T=arg max1≤i≤NT(Si)],q′ T =arg max 1≤i≤NT (S i )], 此时P′(Q,O|λ)即为观测序列与模型的匹配得分;At this time, P'(Q, O|λ) is the matching score between the observation sequence and the model; 4a6)合并4a4)所得的一组q′1,q′2,...,q′T-1和4a4)所得的q′T,得到最优状态路径序列:Q′={q′1,q′2,...,q′T}。 4a6 ) Combine a set of q 1 , q2 , . q' 2 , ..., q' T }.
CN201910181668.3A 2019-03-11 2019-03-11 Scoring method of polite expressions at highway entrances Pending CN109887498A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910181668.3A CN109887498A (en) 2019-03-11 2019-03-11 Scoring method of polite expressions at highway entrances

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910181668.3A CN109887498A (en) 2019-03-11 2019-03-11 Scoring method of polite expressions at highway entrances

Publications (1)

Publication Number Publication Date
CN109887498A true CN109887498A (en) 2019-06-14

Family

ID=66931723

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910181668.3A Pending CN109887498A (en) 2019-03-11 2019-03-11 Scoring method of polite expressions at highway entrances

Country Status (1)

Country Link
CN (1) CN109887498A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112259077A (en) * 2020-10-20 2021-01-22 网易(杭州)网络有限公司 Voice recognition method, device, terminal and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7050977B1 (en) * 1999-11-12 2006-05-23 Phoenix Solutions, Inc. Speech-enabled server for internet website and method
CN101510222A (en) * 2009-02-20 2009-08-19 北京大学 Multilayer index voice document searching method and system thereof
CN103956166A (en) * 2014-05-27 2014-07-30 华东理工大学 Multimedia courseware retrieval system based on voice keyword recognition
CN106776806A (en) * 2016-11-22 2017-05-31 广东电网有限责任公司佛山供电局 The methods of marking and system of call center's quality inspection voice
CN109192224A (en) * 2018-09-14 2019-01-11 科大讯飞股份有限公司 A kind of speech evaluating method, device, equipment and readable storage medium storing program for executing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7050977B1 (en) * 1999-11-12 2006-05-23 Phoenix Solutions, Inc. Speech-enabled server for internet website and method
CN101510222A (en) * 2009-02-20 2009-08-19 北京大学 Multilayer index voice document searching method and system thereof
CN103956166A (en) * 2014-05-27 2014-07-30 华东理工大学 Multimedia courseware retrieval system based on voice keyword recognition
CN106776806A (en) * 2016-11-22 2017-05-31 广东电网有限责任公司佛山供电局 The methods of marking and system of call center's quality inspection voice
CN109192224A (en) * 2018-09-14 2019-01-11 科大讯飞股份有限公司 A kind of speech evaluating method, device, equipment and readable storage medium storing program for executing

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
张伟津: "基于多模态的呼叫中心质量监控系统的研究和实现", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
陈军霞等: "基于Baum-Welch算法HMM模型的孤词算法研究", 《河北科技大学学报》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112259077A (en) * 2020-10-20 2021-01-22 网易(杭州)网络有限公司 Voice recognition method, device, terminal and storage medium
CN112259077B (en) * 2020-10-20 2024-04-09 网易(杭州)网络有限公司 Speech recognition method, device, terminal and storage medium

Similar Documents

Publication Publication Date Title
Liu et al. Voice Conversion Across Arbitrary Speakers Based on a Single Target-Speaker Utterance.
CN107452379B (en) Dialect language identification method and virtual reality teaching method and system
CN109119072A (en) Civil aviaton's land sky call acoustic model construction method based on DNN-HMM
CN101645269A (en) Language recognition system and method
CN112466316A (en) Zero-sample voice conversion system based on generation countermeasure network
CN107039036A (en) A kind of high-quality method for distinguishing speek person based on autocoding depth confidence network
CN112951206A (en) Tibetan Tibet dialect spoken language identification method based on deep time delay neural network
CN116403561A (en) Audiobook production method, production device and storage medium
Wu et al. Multilingual text-to-speech training using cross language voice conversion and self-supervised learning of speech representations
Ohnaka et al. Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images
Chen et al. Neural fusion for voice cloning
Li et al. Spontts: modeling and transferring spontaneous style for tts
Zhao et al. Research on voice cloning with a few samples
CN114120973B (en) Training method for voice corpus generation system
CN109887498A (en) Scoring method of polite expressions at highway entrances
CN119694289A (en) A speech synthesis method, device and electronic device based on style description
Udagawa et al. Human-in-the-loop Speaker Adaptation for DNN-based Multi-speaker TTS
CN117935775A (en) A speech error correction method for dialect accent environment
Yang et al. Indonesian speech recognition based on Deep Neural Network
CN115017883B (en) Text Punctuation Restoration Method Based on Pre-training Fusion Speech Features
Zhao et al. Singing voice conversion based on wd-gan algorithm
Su et al. Adversarial diffusion probability model for cross-domain speaker verification integrating contrastive loss
Zhang The algorithm of voiceprint recognition model based DNN-RELIANCE
Zhao et al. IVCGAN: an improved GAN for voice conversion
Aura et al. Analysis of the Error Pattern of HMM based Bangla ASR

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20190614

WD01 Invention patent application deemed withdrawn after publication