CN104090918B - Sentence similarity calculation method based on information amount - Google Patents
Sentence similarity calculation method based on information amount Download PDFInfo
- Publication number
- CN104090918B CN104090918B CN201410268361.4A CN201410268361A CN104090918B CN 104090918 B CN104090918 B CN 104090918B CN 201410268361 A CN201410268361 A CN 201410268361A CN 104090918 B CN104090918 B CN 104090918B
- Authority
- CN
- China
- Prior art keywords
- formula
- sentences
- words
- sentence
- concepts
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Machine Translation (AREA)
Abstract
本发明涉及一种基于信息量的句子相似度计算方法,包括以下步骤:首先,通过两个句子词语间具有最大的信息量的概念确定词语的词义;然后利用语义网的层次结构和语料库统计来计算词语的信息量和多词语间的公共信息量;接下来应用组合数学中容斥原理计算多个词语的总信息量,从而分别得到两个句子各自的信息量,以及两个句子总共的信息量;最后根据Jaccard相似度原理定义并计算出句子的相似度。本发明能逼真的模拟人类对句子相似程度的判断,并且不需要使用语料训练参数或使用经验参数、不依赖语料库的规模、无需词性标注等其他自然语言处理技术;时间性能优秀,对一般长度的句子对,在当前主流多核PC机上获得准实时计算效率。
The present invention relates to a kind of sentence similarity calculation method based on the amount of information, comprising the following steps: first, determine the meaning of the word through the concept with the largest amount of information between two sentence words; then use the hierarchical structure of the semantic web and corpus statistics to Calculate the amount of information of words and the amount of public information between multiple words; then apply the principle of inclusion and exclusion in combinatorics to calculate the total information amount of multiple words, so as to obtain the respective information amount of the two sentences and the total information of the two sentences Quantity; Finally, the similarity of sentences is defined and calculated according to the Jaccard similarity principle. The present invention can realistically simulate human beings' judgment on the similarity of sentences, and does not need to use corpus training parameters or use experience parameters, does not depend on the scale of the corpus, and does not need other natural language processing technologies such as part-of-speech tagging; the time performance is excellent, and it is suitable for general-length sentences. The sentence is right, to obtain quasi-real-time computing efficiency on the current mainstream multi-core PC.
Description
技术领域technical field
本发明涉及一种句子相似度计算方法,具体涉及一种基于信息量的句子相似度计算方法,属于自然语言处理技术领域。The invention relates to a sentence similarity calculation method, in particular to an information-based sentence similarity calculation method, which belongs to the technical field of natural language processing.
背景技术Background technique
句子或短文本相似度计算是自然语言处理的一项重要研究内容,近年来在信息检索、机器翻译、问答系统、自动文摘等应用领域中的作用越来越重要。传统的方法多采用文档相似度的计算方法,仅把句子词语看成相互没有关联的无意义符号,对于计算含有少量词语的句子不够精确。而目前常用的混合方法通常需要在相关数据集上训练参数或者使用经验参数,其缺点是依赖训练数据集,通用性不强。Sentence or short text similarity calculation is an important research content of natural language processing. In recent years, it has played an increasingly important role in information retrieval, machine translation, question answering system, automatic summarization and other application fields. The traditional method mostly uses the calculation method of document similarity, and only regards the sentence words as meaningless symbols that are not related to each other, which is not accurate enough for the calculation of sentences containing a small number of words. However, the commonly used hybrid methods usually need to train parameters on related data sets or use empirical parameters. The disadvantage is that they rely on training data sets and are not very versatile.
发明内容Contents of the invention
本发明方法的目的在于解决上述问题,提供一种基于信息量的句子相似度计算方法,通过使用信息量这个语言的本质属性,使用容斥原理获得精确的多个词语的总信息量,从而得到更接近于人主观判断得到的句子相似度结果。The purpose of the method of the present invention is to solve the above-mentioned problem, provide a kind of sentence similarity calculation method based on the amount of information, by using the essential property of the language of the amount of information, use the tolerance-exclusion principle to obtain accurate total information amount of a plurality of words, thereby obtain It is closer to the sentence similarity result obtained by human subjective judgment.
本发明方法的思想是首先通过两个句子词语间具有最大的信息量的概念确定词语的词义;然后利用语义网(比如WordNet)的层次结构和语料库(比如BNC语料库或Brown语料库等)统计来计算词语的信息量和多词语间的公共信息量;接下来应用组合数学中容斥原理计算多个词语的总信息量,从而分别得到两个句子各自的信息量,以及两个句子总共的信息量;最后根据Jaccard相似度原理定义并计算出句子的相似度。The idea of the inventive method is first to determine the meaning of the word by the concept of the maximum amount of information between two sentence words; then utilize the hierarchical structure of the Semantic Web (such as WordNet) and the statistics of the corpus (such as the BNC corpus or the Brown corpus, etc.) to calculate The amount of information of words and the amount of public information between multiple words; then apply the principle of tolerance and exclusion in combinatorics to calculate the total information amount of multiple words, so as to obtain the respective information amount of two sentences and the total information amount of two sentences ; Finally, define and calculate the sentence similarity according to the Jaccard similarity principle.
为达到上述目的,本发明采用的技术方案是:In order to achieve the above object, the technical scheme adopted in the present invention is:
一种基于信息量的句子相似度计算方法,包括以下步骤:A method for calculating sentence similarity based on information content, comprising the following steps:
步骤1:输入待计算的两个句子sa和sb,记句子sa和sb分别为:Step 1: Input the two sentences s a and s b to be calculated, and record the sentences s a and s b as:
其中,和分别表示句子sa和sb的第i个词语,n和m分别表示句子sa和sb的词语数;in, and represent the i-th word of sentences s a and s b respectively, and n and m represent the number of words of sentences s a and s b respectively;
步骤2:对输入句子中的词语进行词义选择,过程如下:Step 2: Select the meaning of the words in the input sentence, the process is as follows:
词语的词义按照式1确定:words The meaning of the word is determined according to Formula 1:
[式1][Formula 1]
其中,subsum(c1,c2)为在语义网中包含概念c1和c2的所有概念集合,表示在语义网中所有包含词语的概念的集合,concepts(sb)表示语义网中包含句子sb中的所有词语的概念的集合,P(c)为概念c在语料库中的频率,特殊的,如果P(c)为0,则log P(c)为0,P(c)的值按照式2确定:Among them, subsum(c 1 ,c 2 ) is the collection of all concepts including concepts c 1 and c 2 in the Semantic Web, Represents all words in the Semantic Web that contain The set of concepts, concepts(s b ) represents the set of concepts in the semantic web that contains all the words in the sentence s b , P(c) is the frequency of concept c in the corpus, special, if P(c) is 0 , then log P(c) is 0, and the value of P(c) is determined according to formula 2:
[式2][Formula 2]
P(c)=Σw∈words(c)count(w)/NP(c)= Σw∈words(c) count(w)/N
其中words(c)表示语义网中概念c以及概念c的所有子概念中的所有词语的集合,count(w)为词语w在语料库中的频数,N表示语义网中全部概念的频数之和,而每个概念的频数为该概念中全部词语在语料库中的总频数之和;Among them, words(c) represents the set of all words in concept c and all sub-concepts of concept c in the semantic web, count(w) is the frequency of word w in the corpus, and N represents the sum of the frequencies of all concepts in the semantic web, The frequency of each concept is the sum of the total frequencies of all words in the concept in the corpus;
同理,将式1中替换成concepts(sb)替换成concepts(sa),可得句子sb中第i个词语的词义 Similarly, in formula 1 replace with concepts(s b ) is replaced by concepts(s a ), the meaning of the i-th word in the sentence s b can be obtained
词义确定后句子sa和sb可以记为:After the word meaning is determined, the sentences s a and s b can be recorded as:
步骤3:根据步骤2所得确定词义的句子,应用组合数学中的容斥原理计算句子sa和sb各自的信息量以及二者的总信息量,计算过程如下:Step 3: According to the sentence obtained in step 2 to determine the meaning of the word, apply the principle of inclusion and exclusion in combinatorics to calculate the respective information content of sentences s a and s b and the total information content of the two sentences. The calculation process is as follows:
句子sa的信息量IC(sa)的计算公式如式3所示:The calculation formula of the information content IC(s a ) of the sentence s a is shown in formula 3:
[式3][Formula 3]
其中,表示通过语义网的层次结构和语料库统计共同构建的语义信息空间,根据式4计算:in, Represents the semantic information space jointly constructed by the hierarchical structure of the Semantic Web and corpus statistics, Calculate according to formula 4:
[式4][Formula 4]
其中,为在语义网中包含概念的所有概念的集合;in, To include concepts in the Semantic Web The collection of all concepts of
同理,把式3和式4中所有字母a替换成b,n替换成m,可得句子sb的信息量;In the same way, replace all letters a in formula 3 and formula 4 with b, and n with m, and the information amount of sentence s b can be obtained;
把句子sa和sb中所有不重复的词语的集合看成一个新的句子,则通过式5得到句子sa和sb的总信息量IC(sa∪sb):The set of all non-repetitive words in sentences s a and s b is regarded as a new sentence, and the total information IC(s a ∪s b ) of sentences s a and s b can be obtained through formula 5:
[式5][Formula 5]
其中,p为句子sa和sb不重复的词语的总数;Among them, p is the total number of words that are not repeated in sentences s a and s b ;
步骤4:由并集和交集之间的关系定义两个句子sa和sb的公共信息量COMMONIC(sa,sb),计算公式如式6所示:Step 4: Define the public information amount COMMONIC(s a , s b ) of two sentences s a and s b from the relationship between union and intersection, and the calculation formula is shown in formula 6:
[式6][Formula 6]
COMMONIC(sa,sb)=IC(sa)+IC(sb)-IC(sa∪sb)COMMONIC(s a ,s b )=IC(s a )+IC(s b )-IC(s a ∪s b )
步骤5:根据Jaccard相关性原理,定义句子sa和sb的相似度sim(sa,sb),计算公式如式7所示:Step 5: According to the Jaccard correlation principle, define the similarity sim(s a , s b ) of sentences s a and s b , and the calculation formula is shown in formula 7:
[式7][Formula 7]
步骤6:输出两个句子的相似度sim(sa,sb)。Step 6: Output the similarity sim(s a ,s b ) of the two sentences.
有益效果Beneficial effect
对比现有技术,本发明方法进一步提高了混合方法的精确度,能逼真的模拟人类对句子相似程度的判断,并且不需要使用语料训练参数或使用经验参数、不依赖语料库的规模、无需词性标注等其他自然语言处理技术,可以直接用于句子相似度计算,具有较好的通用性;时间性能优秀,对一般长度的句子对,在当前主流多核PC机上获得准实时计算效率。Compared with the prior art, the method of the present invention further improves the accuracy of the hybrid method, can realistically simulate human judgment on the similarity of sentences, and does not need to use corpus training parameters or use experience parameters, does not depend on the size of the corpus, and does not require part-of-speech tagging and other natural language processing technologies can be directly used for sentence similarity calculation, which has good versatility; the time performance is excellent, and for sentence pairs of general length, quasi-real-time computing efficiency can be obtained on current mainstream multi-core PCs.
附图说明Description of drawings
图1是本发明方法的流程图。Figure 1 is a flow chart of the method of the present invention.
图2是本发明方法与其他方法对比图。Fig. 2 is a comparison diagram between the method of the present invention and other methods.
具体实施方式detailed description
下面将结合附图对本发明的实施过程进行详细说明。The implementation process of the present invention will be described in detail below in conjunction with the accompanying drawings.
如图1所示,本发明方法主要分为5个步骤:As shown in Figure 1, the inventive method is mainly divided into 5 steps:
步骤1:输入待计算的两个句子。记句子sa和sb分别为:Step 1: Enter the two sentences to be calculated. Remember the sentences s a and s b are:
其中,和分别表示句子sa和sb的第i个词语,n和m分别表示句子sa和sb的词语数。in, and represent the i-th word of sentences s a and s b respectively, and n and m represent the number of words of sentences s a and s b respectively.
步骤2:因词语的一词多义现象普遍存在,对输入句子的词语进行词义选择,可以消除句子语义的不确定性,从而为后续计算句子相似度做准备。具体过程如下:Step 2: Due to the common phenomenon of polysemous words, selecting the meaning of the words in the input sentence can eliminate the uncertainty of the sentence semantics, so as to prepare for the subsequent calculation of sentence similarity. The specific process is as follows:
(1)在输入的两个句子中,各选取一个词语组成词语对;(1) In the two input sentences, select a word to form a word pair;
(2)使用语义网(比如WordNet)中词语的词义(或概念)来代替词语,该词语对转换成多个词义对;(2) Use the meaning (or concept) of the word in the semantic network (such as WordNet) to replace the word, and the word pair is converted into multiple word sense pairs;
(3)计算词义对之间的公共信息量;(3) Calculate the amount of public information between word sense pairs;
(4)选取最大的公共信息量作为该词语对的公共信息量;(4) choose the maximum amount of public information as the amount of public information of the word pair;
(5)把所有词语对的公共信息量按照从大到小排序;(5) sort the public information of all word pairs in descending order;
(6)定义两个数组,用来记录每个句子中词语的词义,并初始化数组中的所有元素值为未确定词义;(6) define two arrays, be used for recording the meaning of words in each sentence, and all element values in the initialization array are undetermined meanings;
(6)按照(5)的顺序依次处理,如果词语的词义还没有确定,即词语在数组中对应的元素值仍为未确定词义,则把该信息量对应的词语的词义记录到词义数组中;(6) Process according to the order of (5), if the meaning of the word has not been determined, that is, the corresponding element value of the word in the array is still undetermined, then the meaning of the word corresponding to the amount of information is recorded in the meaning array ;
(7)直到数组元素值没有未确定词义为止,即全部词语的词义确定完毕。(7) Until the array element values have no undetermined meanings, that is, the meanings of all words are determined.
则词语的词义(或概念)按照如式1确定:words The meaning (or concept) of is determined according to formula 1:
[式1][Formula 1]
其中in
[式2][Formula 2]
P(c)=∑w∈words(c)count(w)/NP(c)= ∑w∈words(c) count(w)/N
这里,subsum(c1,c2)为在语义网中包含概念c1和c2的所有概念集合,表示在语义网中的所有包含词语的概念的集合,concepts(sb)表示语义网中包含句子sb中的所有词语的概念的集合,P(c)为概念c在某语料库中的频率。特殊的,如果P(c)为0,则logP(c)为0;Here, subsum(c 1 ,c 2 ) is the collection of all concepts including concepts c 1 and c 2 in the Semantic Web, Represents all contained words in the Semantic Web concepts(s b ) represents the set of concepts in the semantic web including all the words in sentence s b , and P(c) is the frequency of concept c in a corpus. In particular, if P(c) is 0, then logP(c) is 0;
words(c)表示语义网中概念c以及概念c的所有子概念中的所有词语的集合,count(w)为词语w在某一语料库中的频数,N表示语义网中全部概念的频数之和,而每个概念的频数为该概念中全部词语在语料库中的总频数之和。words(c) represents the collection of all words in concept c and all sub-concepts of concept c in the semantic web, count(w) is the frequency of word w in a certain corpus, and N represents the sum of the frequencies of all concepts in the semantic web , and the frequency of each concept is the sum of the total frequencies of all words in the concept in the corpus.
同理可得句子sb中第i个词语的词义词义确定后句子sa和sb可以记为:In the same way, the meaning of the i-th word in the sentence s b can be obtained After the word meaning is determined, the sentences s a and s b can be recorded as:
步骤3:根据步骤2所得确定词义的句子,应用组合数学中的容斥原理计算句子sa和sb各自的信息量,即句子sa的信息量的计算公式如式3所示:Step 3: According to the sentence whose meaning is determined in step 2, apply the principle of inclusion and exclusion in combinatorics to calculate the respective information volumes of sentences s a and s b , that is, the calculation formula of the information volume of sentence s a is shown in formula 3:
[式3][Formula 3]
其中in
[式4][Formula 4]
这里,表示通过语义网的层次结构和语料库统计共同构建的语义信息空间中概念的交集,为在语义网中包含概念的所有概念的集合,对于取值在1至n之间的任意k,表示句子sa中取k个词的一种组合。here, Represent concepts in the semantic information space jointly constructed by the hierarchical structure of the Semantic Web and corpus statistics the intersection of To include concepts in the Semantic Web The set of all concepts of , for any k between 1 and n, Indicates a combination of k words in the sentence s a .
同理,把式3和式4中所有字母a替换成b,n替换成m,可得句子sb的信息量。In the same way, replace all letters a in formula 3 and formula 4 with b, and n with m, and the information amount of sentence s b can be obtained.
把句子sa和sb中所有不重复的词语的集合看成一个新的句子,则通过式5得到句子sa和sb的总信息量:The set of all non-repetitive words in sentences s a and s b is regarded as a new sentence, then the total information amount of sentences s a and s b can be obtained through formula 5:
[式5][Formula 5]
其中,p为两个句子不重复的词语总数。Among them, p is the total number of words that are not repeated in the two sentences.
步骤4:通过步骤3得到的两个句子各自的信息量以及两个句子的总信息量。由并集和交集之间的关系得到两个句子sa和sb的公共信息量COMMONIC(sa,sb)如式6所示:Step 4: The respective information amounts of the two sentences and the total information amount of the two sentences obtained in step 3. The public information COMMONIC(s a , s b ) of the two sentences s a and s b is obtained from the relationship between the union and the intersection, as shown in Formula 6:
[式6][Formula 6]
COMMONIC(sa,sb)=IC(sa)+IC(sb)-IC(sa∪sb)COMMONIC(s a ,s b )=IC(s a )+IC(s b )-IC(s a ∪s b )
步骤5:通过步骤3和步骤4得到的两个句子的总信息量和公共信息量,根据Jaccard相关性原理,两个句子sa和sb的相似度可通过式7计算:Step 5: The total information and public information of the two sentences obtained through steps 3 and 4, according to the Jaccard correlation principle, the similarity between the two sentences s a and s b can be calculated by formula 7:
[式7][Formula 7]
如上所述,本发明提供了一种基于信息量的句子相似度计算方法。通过用户输入的真实句子对,系统将自动计算出逼真于人类对句子相似度判断的结果。As mentioned above, the present invention provides a method for calculating sentence similarity based on information content. Through the real sentence pair input by the user, the system will automatically calculate the result that is realistic to the human judgment of sentence similarity.
下面给出本发明方法与现有其他4种句子相似度计算方法的比较结果。实验中使用了的语义网WordNet和语料库BNC。评估采用线性相关的Pearson相关系数(PCC),一般排序相关的Spearman rank相关系数(SRCC)和概率排序相关的Kendall rank相关系数(KRCC)。表1为对词语对应句子对采用各种方法的打分结果。The comparison results between the method of the present invention and other four existing sentence similarity calculation methods are given below. The semantic web WordNet and corpus BNC are used in the experiment. Pearson correlation coefficient (PCC) for linear correlation, Spearman rank correlation coefficient (SRCC) for general rank correlation and Kendall rank correlation coefficient (KRCC) for probability rank correlation were used for evaluation. Table 1 shows the scoring results of the sentence pairs corresponding to the words using various methods.
表1句子对的人工打分和各种方法打分Table 1 Manual scoring of sentence pairs and scoring by various methods
从表1可得到图2。从图2中可以看到,本文方法在PCC、SRCC和KRCC上都优于其他4种方法,说明模型对句子相似度的判断更接近人类的主观判断。另外,人工数据集的单一志愿者打分与全部志愿者打分均值之间的PCC均值为0.825,最大值为0.921;而本发明方法的PCC为高于单人打分的平均值且低于单人打分的上限值;说明模型对句子相似度的判断水平高于人的平均水平,且结果可信。Figure 2 can be obtained from Table 1. As can be seen from Figure 2, the method in this paper is superior to the other four methods on PCC, SRCC and KRCC, indicating that the model's judgment of sentence similarity is closer to human subjective judgment. In addition, the average value of PCC between the scores of a single volunteer and the mean of all volunteers in the artificial data set is 0.825, and the maximum value is 0.921; while the PCC of the method of the present invention is higher than the average value of single-person scoring and lower than the average value of single-person scoring The upper limit value of ; it shows that the judgment level of the model on sentence similarity is higher than the average human level, and the result is credible.
以上所述的具体描述,对发明的目的、技术方案和有益效果进行了进一步详细说明,所应理解的是,以上所述仅为本发明的具体实施例而已,并不用于限定本发明的保护范围,凡在本发明的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The specific description above further elaborates the purpose, technical solution and beneficial effect of the invention. It should be understood that the above description is only a specific embodiment of the present invention and is not used to limit the protection of the present invention. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention shall be included in the protection scope of the present invention.
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410268361.4A CN104090918B (en) | 2014-06-16 | 2014-06-16 | Sentence similarity calculation method based on information amount |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410268361.4A CN104090918B (en) | 2014-06-16 | 2014-06-16 | Sentence similarity calculation method based on information amount |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104090918A CN104090918A (en) | 2014-10-08 |
CN104090918B true CN104090918B (en) | 2017-02-22 |
Family
ID=51638634
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410268361.4A Active CN104090918B (en) | 2014-06-16 | 2014-06-16 | Sentence similarity calculation method based on information amount |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104090918B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105354186A (en) * | 2015-11-05 | 2016-02-24 | 同济大学 | News event extraction method and system |
CN106780061A (en) * | 2016-11-30 | 2017-05-31 | 华南师范大学 | Social network user analysis method and device based on comentropy |
CN108628883B (en) * | 2017-03-20 | 2021-03-16 | 北京搜狗科技发展有限公司 | Data processing method and device and electronic equipment |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102945228A (en) * | 2012-10-29 | 2013-02-27 | 广西工学院 | Multi-document summarization method based on text segmentation |
CN103136359A (en) * | 2013-03-07 | 2013-06-05 | 宁波成电泰克电子信息技术发展有限公司 | Generation method of single document summaries |
CN103699529A (en) * | 2013-12-31 | 2014-04-02 | 哈尔滨理工大学 | Method and device for fusing machine translation systems by aid of word sense disambiguation |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4299963B2 (en) * | 2000-10-02 | 2009-07-22 | ヒューレット・パッカード・カンパニー | Apparatus and method for dividing a document based on a semantic group |
US9189482B2 (en) * | 2012-10-10 | 2015-11-17 | Abbyy Infopoisk Llc | Similar document search |
-
2014
- 2014-06-16 CN CN201410268361.4A patent/CN104090918B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102945228A (en) * | 2012-10-29 | 2013-02-27 | 广西工学院 | Multi-document summarization method based on text segmentation |
CN103136359A (en) * | 2013-03-07 | 2013-06-05 | 宁波成电泰克电子信息技术发展有限公司 | Generation method of single document summaries |
CN103699529A (en) * | 2013-12-31 | 2014-04-02 | 哈尔滨理工大学 | Method and device for fusing machine translation systems by aid of word sense disambiguation |
Non-Patent Citations (3)
Title |
---|
一种基于含糊同义词的查询扩展方法;李海芳等;《计算机应用与软件》;20111215;第28卷(第12期);第41-43,47页 * |
一种基于知网的句子相似度计算方法;程传鹏等;《计算机工程与科学》;20120215;第34卷(第2期);第172-175页 * |
多层次结构句子相似计算的应用研究;金春霞;《计算机应用与软件》;20091015;第26卷(第10期);第180-182,202页 * |
Also Published As
Publication number | Publication date |
---|---|
CN104090918A (en) | 2014-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107766324B (en) | Text consistency analysis method based on deep neural network | |
CN108549626A (en) | A kind of keyword extracting method for admiring class | |
CN102663139B (en) | Method and system for constructing emotional dictionary | |
CN102880600B (en) | Based on the phrase semantic tendency Forecasting Methodology of world knowledge network | |
CN108595425A (en) | Based on theme and semantic dialogue language material keyword abstraction method | |
CN114880461B (en) | Chinese news text abstract method combining contrast learning and pre-training technology | |
CN110598219A (en) | A sentiment analysis method for Douban movie reviews | |
CN103049435A (en) | Text fine granularity sentiment analysis method and text fine granularity sentiment analysis device | |
CN106997382A (en) | Innovation intention label automatic marking method and system based on big data | |
CN109670039A (en) | Sentiment analysis method is commented on based on the semi-supervised electric business of tripartite graph and clustering | |
CN106257455B (en) | A kind of Bootstrapping method extracting viewpoint evaluation object based on dependence template | |
CN103699626A (en) | Method and system for analysing individual emotion tendency of microblog user | |
CN103729456B (en) | Microblog multi-modal sentiment analysis method based on microblog group environment | |
CN109299865A (en) | Psychological evaluation system and method based on semantic analysis, information data processing terminal | |
CN102279844A (en) | Method and system for automatically testing Chinese composition | |
CN106294316A (en) | A kind of text emotion based on dictionary analyzes method | |
CN108280164B (en) | A short text filtering and classification method based on category-related words | |
CN103246687A (en) | Automatic Blog abstracting method based on characteristic information | |
CN110427609A (en) | One kind writing people's composition structure of an article reasonability method for automatically evaluating | |
CN102663001A (en) | Automatic blog writer interest and character identifying method based on support vector machine | |
CN101770580A (en) | Training method and classification method of cross-field text sentiment classifier | |
CN102214246A (en) | Method for grading Chinese electronic document reading on the Internet | |
CN104572877A (en) | Detection method and detection system of game public opinion | |
CN105095430A (en) | Method and device for setting up word network and extracting keywords | |
CN105488098B (en) | A New Word Extraction Method Based on Domain Difference |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |