[go: up one dir, main page]

CN113609892A - Handwritten poetry recognition method integrating deep learning with scenic spot knowledge map - Google Patents

Handwritten poetry recognition method integrating deep learning with scenic spot knowledge map Download PDF

Info

Publication number
CN113609892A
CN113609892A CN202110663733.3A CN202110663733A CN113609892A CN 113609892 A CN113609892 A CN 113609892A CN 202110663733 A CN202110663733 A CN 202110663733A CN 113609892 A CN113609892 A CN 113609892A
Authority
CN
China
Prior art keywords
text
handwritten
poetry
image
poem
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110663733.3A
Other languages
Chinese (zh)
Other versions
CN113609892B (en
Inventor
何坚
白佳豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN202110663733.3A priority Critical patent/CN113609892B/en
Publication of CN113609892A publication Critical patent/CN113609892A/en
Application granted granted Critical
Publication of CN113609892B publication Critical patent/CN113609892B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Animal Behavior & Ethology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Databases & Information Systems (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Character Discrimination (AREA)

Abstract

深度学习与景区知识图谱融合的手写诗词识别方法属于电子信息领域,解决采用文本识别方法从景区手写诗词图像中提取的诗词文本难以将诗词的主要文本涵盖的问题。本发明通过与用户智能手机进行交互,借助其集成的北斗/GPS传感器感知用户所在位置、朝向信息,用户通过智能手机的摄像头采集景区内手写诗词的图像数据;通过基于FPN的景区手写诗词检测技术对景区手写诗词图像数据中的诗词文本位置信息进行检测,提取到诗词文本的具体区域位置信息;通过基于ACE的景区手写诗词识别方法对区域内的诗词文本信息进行识别,获取到初步的诗词文本识别结果;通过景区知识图谱对初步识别的诗词文本结果进行矫正完善,将最终的诗词文本识别结果返回给用户。

Figure 202110663733

The handwritten poetry recognition method based on the fusion of deep learning and scenic spot knowledge map belongs to the field of electronic information, which solves the problem that the main text of the poem is difficult to be covered by the text of the poem extracted from the handwritten poem image of the scenic spot by the text recognition method. The invention interacts with the user's smart phone, and uses the integrated Beidou/GPS sensor to perceive the user's location and orientation information, and the user collects the image data of handwritten poems in the scenic spot through the camera of the smart phone; Detect the location information of the poem text in the image data of handwritten poems in the scenic spot, and extract the specific area location information of the poem text; identify the poem text information in the area through the ACE-based handwritten poem recognition method in the scenic spot, and obtain the preliminary poem text. Recognition results; correct and improve the preliminary recognized poetry text results through the scenic knowledge map, and return the final poetry text recognition results to the user.

Figure 202110663733

Description

Handwritten poetry recognition method integrating deep learning with scenic spot knowledge map
Technical Field
The invention belongs to the field of electronic information, and relates to a handwritten poetry recognition method based on deep learning and knowledge maps and applicable to scenic spots.
Background
With the rapid development of economy in China, scenic spot visiting and touring increasingly become indispensable contents for daily life of people. Meanwhile, in order to improve the popularity and attract passengers, different scenic spots in the scenic spots usually quote poetry works of famous poetry famous ancient and modern times so as to improve the personal information of the scenic spots. These cited poems usually copy the handwritten characters of the famous poetry of the cited poems, resulting in different styles of the handwritten poems in scenic spots and also causing difficulty in recognizing all the characters in the cited poems completely by tourists today. Therefore, the recognition method of the handwritten poetry in the scenic spot obviously becomes a key problem. At present, a mature technical method for the field exists at home and abroad, and the method can be mainly divided into two types: the first method is a method for respectively training a scene poetry image text detection and recognition model, wherein the text detection method mainly refers to an algorithm in the field of target detection and performs regression detection on a text box of a poetry text; the text recognition method mainly refers to an algorithm in the field of speech recognition, performs characteristic coding on poetry text image regions, and further decodes poetry text content information by using a model. The second is an end-to-end recognition method that combines text detection with a text recognition model, which can optimize the text detection model by the text recognition result, but has a greater computational complexity than the first method. Although the two methods have achieved good recognition effect at present, the recognition effect is poor due to the factors of complex text content fonts, poor imaging effect and complex background of handwritten poetry images in scenic spots. In view of the fact that at present, more and more scenic spots establish knowledge maps facing the scenic spots, tourists can conveniently retrieve scenic spot information in the scenic spots, and a new solution is brought to the problem that the recognition effect of handwritten poetry image texts in the scenic spots is poor.
Disclosure of Invention
In order to solve the problem that a poetry text extracted from a scenic spot handwritten poetry image by adopting a traditional text recognition method is difficult to cover a main text of the poetry, the invention provides a scenic spot handwritten poetry recognition method integrating deep learning with a scenic spot knowledge map. The method mainly interacts with a smart phone of a user, senses the position and orientation information of the user by means of an integrated Beidou/GPS sensor, and the user acquires image data of handwritten poems in a scenic spot through a camera of the smart phone; detecting poetry text position information in the scenic spot handwritten poetry image data through a FPN-based scenic spot handwritten poetry detection technology, and extracting specific area position information of poetry texts; identifying poetry text information in the region by using an ACE-based scenic spot handwritten poetry identification method to obtain a preliminary poetry text identification result; and correcting and perfecting the poetry text result primarily identified through the scenic spot knowledge map, and returning the final poetry text identification result to the user for display.
The invention comprises the following steps:
step 1, acquiring a scene region handwritten poetry text image to be identified and related attribute information of the image, including an image shooting geographical position, an image background texture and a character direction, weakening noise in the image through a spatial domain enhancement algorithm, and enabling a structural similarity index of an original image and a denoised image to be larger than 0.9 to obtain a preprocessing result.
And 2, inputting the preprocessed scenic spot handwritten poetry images into a feature extraction network based on VGG16 to perform feature extraction of poetry texts, obtaining a handwritten poetry text weight sequence of the scenic spot images through a poetry text classifier, and training the classifier.
And 3, inputting the extracted handwritten poetry text characteristic graph into a handwritten poetry detection network based on the FPN for fusion to obtain a single-character Gaussian thermodynamic diagram of the handwritten poetry text, and further obtaining the text region position information of the handwritten poetry image in the scenic spot through a multi-character text frame linking algorithm.
And 4, performing area cutting on the handwritten poetry images in the scenic region according to the extracted position information of the text images, and sequentially inputting the cut areas into an ACE (adaptive communication environment) -based handwritten poetry recognition network to perform Encoder-Decoder processing on the handwritten poetry texts to obtain recognition results of the handwritten poetry images in the scenic region.
And 5, inputting relevant attribute information of the handwritten poetry images of the scenic spot into a knowledge map of the scenic spot, carrying out graph search to obtain a search knowledge result set, and obtaining a final text recognition result of the handwritten poetry images of the scenic spot by using the recognition result obtained in the step 4 and the search knowledge result set through a matching algorithm of the handwritten poetry of the scenic spot.
Effects of the invention
The handwritten poetry recognition method based on fusion of vision and scenic spot knowledge maps designed by the smart phone can facilitate scenic spot visitors to perform text recognition on handwritten poetry in scenic spots, enhance understanding of scenic spot culture, improve scenic spot pedestrian volume and promote scenic spot relevant culture. The method comprises the steps of performing text recognition on handwritten poetry images in a scenic spot by introducing deep learning, extracting relevant poetry information from a scenic spot knowledge map according to the position and orientation information of tourists acquired by a Beidou/GPS sensor and poetry relevant attribute information, and correcting a text recognition result to assist the tourists in accurately recognizing the handwritten poetry image texts.
Invention difficulties
(1) A detection technology of scene handwritten poetry based on FPN is designed, the difficulty lies in accurate prediction of large and small text areas in scene handwritten poetry images, and meanwhile, the requirement of low delay is also guaranteed.
(2) The recognition method of the handwritten poetry in the scenic spot based on the ACE is designed, the text recognition is carried out on the detected poetry image in the text area, and the method is considered to be applied to smart phone interaction, so that the difficulty lies in ensuring the requirements on the recognition accuracy and the real-time performance of the poetry text.
(3) A poetry correcting technology based on a knowledge graph is designed, the poetry text result which is preliminarily recognized is corrected and perfected, and the difficulty lies in how to perform fusion matching calculation on the recognized text and the knowledge entity in the knowledge graph of the scenic spot.
Drawings
Fig. 1 is a frame diagram of recognition of handwritten poems in a scenic spot.
Fig. 2 is a multi-feature module network architecture.
Fig. 3 is a diagram of the overall architecture of the system.
FIG. 4 is a view of scenic spot knowledge acquisition and processing steps
Fig. 5 is a scenic spot knowledge cloud recognition service process diagram.
Fig. 6 is a flow chart of result matching of poetry in a scenic region.
FIG. 7 is a fusion architecture diagram of a knowledge graph and a scenic region poetry recognition algorithm.
Core algorithm of the invention
(1) Detection technology of handwritten poetry in scenic spot based on FPN
The structure is as shown in a scene area handwritten poetry image text region detection module in fig. 1, the whole structure is totally divided into 3 parts, which are respectively: the method comprises the steps of handwritten poetry space feature extraction network, a character key point calibration algorithm and a multi-character text box linking algorithm. Firstly, inputting handwritten poetry images in a scenic spot into a handwritten poetry space characteristic extraction network to extract handwritten poetry text characteristics; then, marking key points of single characters by extracted poetry text characteristics through a character key point calibration algorithm; and finally, processing the marked single-character key points by a multi-character text box linking algorithm to obtain region coordinate information of handwritten poetry text in the scenic spot, and transmitting the information as input to a scenic spot handwritten poetry image text character recognition module. The detailed description thereof is as follows:
the method comprises the steps of extracting the characteristics of the handwritten poetry images of the scenic spot by the VGG16 to obtain the poetry text characteristic diagram of the handwritten poetry images of the scenic spot, wherein the characteristics of the convolutional network have hierarchy, and the characteristics of different hierarchies can be subjected to information complementation, so that the network can be directly and effectively incorporated into the multilayer characteristic fusion of a single model, and the accuracy of the network is improved. The backbone network with the multi-feature module fusion uses a feature pyramid network. In order to solve the problem of multi-scale target detection, the feature pyramid network extracts features of different scales from bottom to top from the same picture by using a convolutional neural network, and all layers are connected with each other by using the existing network. The invention adopts 7 layers of pyramid networks, and the layers are connected with each other to promote the characteristic multiplexing. In the pyramid network of the present invention, data transmission is performed using a horizontal pairing of a bottom-up path and a top-down path. The network structure is shown in fig. 2.
1) A bottom-up path: refers to the upward flow of data. As shown on the left side of fig. 2, each layer includes a convolutional layer, a pooling layer, an activation function layer, and a cyclic layer. This path can result in 7 multiscale feature maps, labeled { c }1,c2,c3,c4,c5,c6,c7And recording character characteristics of different levels by different characteristic graphs, wherein the low-level characteristics reflect the characteristic boundaries of poetry characters of a shallow level, and the high-level characteristics reflect the characteristics of poetry characters of a deeper level.
2) A top-down path: refers to the downward flow of data as shown on the right side of fig. 2. By up-sampling the feature map, a high-resolution feature map with strong semantic information is provided, which is important for detecting the handwritten poetry texts in the scenic region. The outputs of the previous layers are used as the input of the current layer, and richer features are extracted. Meanwhile, a deformable convolution module Def-Incept is added, the characteristics of partial deformed texts are extracted, and then a multilayer characteristic diagram { p } is generated1,p2,p3,p4,p5,p6,p7And finally, carrying out convolution operation on the characteristic diagram once, so that the number of parameters is reduced, and the confounding effect caused by deformation is eliminated. The following equation (1) shows the process of feature extraction.
Figure BDA0003116106210000051
Where, Conv denotes a convolution operation,
Figure BDA0003116106210000052
representing feature fusion, UpSample representing an upsampling operation, and Def-inclusion representing a deformable convolution operation. The feature pyramid network module constructed by the invention can fuse shallow features and deep featuresAnd adding the deep semantic information into the shallow feature map, and fusing the features of the shallow feature map and the deep semantic information to generate more small target features, which is very helpful for improving the detection capability of the model on the small text image. And processing the output information of the operation through a Gaussian kernel function to obtain a Gaussian thermodynamic diagram of the character key nodes of the image text.
The invention designs a multi-character text box linking algorithm. The method comprises the steps of taking Gaussian thermodynamic diagrams of character key nodes of an obtained image text as a precondition, and calculating a link relation between the character key nodes to obtain a final scene handwritten poetry image text detection box. The calculation process of the multi-character text box linking algorithm will be described in detail below.
Firstly, analyzing and calculating each character in a handwritten poetry image text in a scenic region through a character key point Gaussian thermodynamic diagram to obtain the maximum diameter of the Gaussian thermodynamic diagram of each character, drawing a square box by taking the maximum diameter as a side length, and marking the position of the text box of a single character, so that the situation that the text box is difficult to completely contain character areas due to the fact that the shape of the Gaussian thermodynamic diagram has a rotation angle can be avoided. Secondly, selecting half of the length of the diagonal line of the single character text box as an initial value of the radius r of the outward radiation circle, and setting the maximum value max of the radius r according to the maximum side length of the input image. Then, the step length is taken as
Figure BDA0003116106210000061
The radius of the radiation circle is continuously increased, the search is carried out in the character direction input by the user, if another character text box is encountered, the search is stopped, and the character text box is linked with the center of the encountered text box; if the radius r reaches the maximum value, the end of the text box link is indicated. And finally, integrating the linked text boxes to obtain a final detection area position result of the handwritten poetry image text in the scenic spot.
(2) ACE-based recognition method for handwritten poems in scenic spot
The architecture is shown as a character recognition network module in fig. 1. In a text recognition network using handwritten poetry in a scenic spot, firstly, a local text area obtained by detecting the handwritten poetry text in the scenic spot is subjected to image normalization processing, so that the data is more standard; then inputting the processed image data into a handwritten poetry word character feature extraction network to perform serialized coding on poetry word text features; and finally, decoding the codes after the poetry text characteristic serialization through a character recognizer to obtain a preliminary scenic spot handwritten poetry text recognition result. The detailed description thereof is as follows:
after normalization processing is carried out on a scene handwritten poetry text image, character feature extraction is carried out on the processed text image, the invention takes a convolution cyclic Neural Network (CRNN) as a backbone Network to carry out feature extraction on the scene handwritten poetry image text, and the character feature extraction Network structure mainly comprises a convolution layer and a cyclic Network layer. Firstly, inputting a poetry text image subjected to normalization processing into a convolution layer, and extracting a convolution characteristic diagram of the image; then, inputting the extracted convolution characteristic graph into a circulation network layer for continuously extracting character sequence characteristics on the basis of the convolution characteristics, wherein the characteristics comprise context information of a poetry text, so that a character recognition result is more accurate; and finally, outputting the extracted features, so as to facilitate further research and analysis.
The method adopts an Aggregation Cross Entropy (ACE) algorithm to decode the characteristic sequence of the handwritten poetry image text in the scenic region so as to realize the recognition of the handwritten poetry image text in the scenic region. The goal of the algorithm is to multiply the gain values by the high frequency portions of the image that express the particular content and then recombine them to obtain a better image. Therefore, calculating the gain factor of the high frequency part is the core of the ACE algorithm. In the initial stage of model training, because different characters are uniformly distributed at different moments and different character categories, the gain coefficient is set to be 1; in the training stage, the probability of a certain category is far higher than that of other categories at different moments, and the gain coefficient is set to be the number of characters of the text in the image. The ACE algorithm can be well suitable for the recognition situation of long texts, can solve the alignment problem caused by sequences of indefinite length, and provides great help for decoding the feature sequences of handwritten poetry texts in scenic spots.
(3) Poetry correcting technology based on knowledge graph
The architecture is shown as a poetry rectification technology module in fig. 1. The invention provides knowledge reasoning and related content knowledge recommendation functions, which specifically comprise the following steps: and (4) searching a map by using the scenic spot knowledge map by taking poetry description information inquired by the user as a searching condition to obtain a searching result set C.
Secondly, after the search result set C and the primary recognition result x of the scenic spot handwritten poetry text image are participled through a participle algorithm f (-), the obtained search result set keyword matrix S is obtainedn×mAnd a result matrix E of primary recognition of handwritten poetry text images in scenic spots1×mWhere n represents the number of entities in the search result set, m represents the number of keywords after processing, k represents the index position of the entities in the search result set, f (-) represents the word segmentation function for segmenting the text, and C [ k ]]Representing a single text in the search result set, and x representing the preliminary recognition result of a handwritten poetry text image in the scenic spot. The calculation formulas are shown in (2) and (3).
Sk×m=f(C[k]) (2)
E1×m=f(x) (3)
Then, the entity keywords in the matrix obtained above are calculated by using the distribution model for generating the word vectors. Each entity key vector v thereofeIs shown in (4), wherein e represents a single sample in the matrix, g (-) represents a distribution model function for generating a word vector, i represents an index position of a keyword of an entity, and Sn×m[e]Representing a single sample data in a keyword matrix of a search result set, Sn×m[e][i]The method comprises the step of gradually taking entity key word data of a single sample in a key word matrix of a search result set.
Figure BDA0003116106210000081
Obtaining the vector v of the poetry texteIs subjected to normalization treatmentAnd then, obtaining a final text vector V of the poetry text, and combining the n text vectors to generate a search result vector set V. . The calculation formula is shown in (5).
Figure BDA0003116106210000082
And (4) obtaining a final vector q of the primary recognition result of the handwritten poetry text image in the scenic spot after the primary recognition result of the handwritten poetry text image in the scenic spot is processed.
And finally, calculating the similarity of two poetry text vectors by using the vector q of the primary recognition Result of the handwritten poetry text image in the scenic region and the text vector V of each entity in the search Result vector set V through VSM, outputting the corresponding text Result in the search Result set C of the vector q of the primary recognition Result of the handwritten poetry text image in the scenic region and the poetry text vector with the highest similarity in the search Result vector set V, and taking the corresponding text Result as the text recognition Result of the handwritten poetry image in the scenic region. The calculation formulas are shown as (6) and (7), wherein VSM ((-)) represents a vector space model calculation function, j represents the index position of a single search result vector, and s1And s2For two texts, the word frequency is respectively atAnd btIndicating the index position of the frequency of the text word.
Figure 100002_DEST_PATH_IMAGE001
Figure 100002_DEST_PATH_IMAGE002
And acquiring a knowledge node with a relation of 'association' with a retrieval Result knowledge node by using a knowledge reasoning method, wherein a three-tuple representation can be described as (. Compared with the prior art, the system has the advantages that the related fragmentation knowledge of the poetry in the scenic spot is more relevant, the poetry body and the poetry characteristic entity are established in a relation by extracting and expressing the knowledge of the original data, so that a user can retrieve the complete information of the poetry more accurately according to the partial content characteristics of the poetry, the scenic spot cultural knowledge and the poetry related resource recommendation service are provided by using the scenic spot poetry knowledge map, and visitors are helped to more effectively and more intuitively receive the scenic spot cultural knowledge.
Detailed Description
In order to make those skilled in the art better understand the technical solutions in the present disclosure, the technical solutions in the present disclosure will be clearly and completely described below with reference to the accompanying drawings of the present disclosure, and it is obvious that the described implementation examples are only a part of implementation examples of the present disclosure, and not all implementation examples. All other embodiments obtained by a person skilled in the art based on the embodiments in the embodiments of the present invention should fall within the scope of the protection of the embodiments of the present invention.
As pointed out in the background section above, how to improve the accuracy of retrieving and identifying poetry in scenic spots becomes a critical issue. At present, retrieval and poetry identification methods mainly comprise two methods: (1) a traditional scenic region poetry database retrieval system (2) is a scenic region poetry identification method based on computer vision. In the first method, a very rich scenic spot poetry data set is needed, the requirement on database retrieval performance is high, and the system is realized based on keyword or shallow semantic analysis, so that the result is poor and the time and labor cost are greatly consumed. Aiming at the second method, the current popular artificial intelligence method is used, and the collected poetry images in the scenic spots are used for model large-scale training and learning, but the method only uses image single-mode data, has low recognition accuracy rate and cannot meet the condition of searching and recognizing most of poetry scenes in the scenic spots.
In view of the above, the present invention provides a method for identifying handwritten poetry by combining deep learning with a scenic spot knowledge map, which can solve the problems mentioned in the related art.
The following further describes example implementations of the present invention in conjunction with the accompanying drawings.
FIG. 3 shows a system architecture diagram of an example of the present invention. The system mainly comprises a knowledge acquisition and processing module, a knowledge storage module and a knowledge application module. The basic layer comprises a knowledge acquisition and processing module, the database layer and the cache layer comprise a knowledge storage module, and the Service end and the API end comprise a knowledge application module.
And the knowledge acquisition and processing module is used for carrying out three processes of data cleaning, knowledge processing and knowledge representation on the original knowledge of the site poetry knowledge related books, websites, poetry Excel spreadsheets and XML files to obtain a relationship network between site poetry characteristic entities and site poetry bodies. FIG. 4 shows an overall step diagram of the knowledge acquisition and processing module.
For example, taking the laevous guoye scenic spot as an example, the method comprises the steps of collecting data about the laevous guoye poetry in a webpage, performing knowledge processing on the poetry in the scenic spot after data cleaning to obtain structural knowledge shown in table 1 (the data cannot be partially omitted), performing knowledge processing on the poetry in the scenic spot to obtain Chinese word segmentation, part-of-speech tagging, named entity recognition and word vector relation calculation, filtering to obtain characteristic entities of the poetry in the scenic spot as shown in table 2, and obtaining a series of relations through a triple knowledge representation method, such as ' when a doctor throws a lotus root and breaks the lotus root when casting a tassel ', belonging to ' foreign country foreign food and foreign country ', ' belonging to ' foreign country food and country ', belonging to ' foreign country ', and having a relationship (foreign country food, belonging to ' and foam ' if the laevous), and related content knowledge relations such as ' dog's (native dog), association, Ming Guo miscellaneous (miscellaneous), is included in the text of the United states of America and Japan).
The knowledge storage module provides a scenic spot poetry knowledge map storage service by using a Neo4j map database, and stores poetry ontology relations with characteristic entities and related content knowledge ontology relations, wherein poetry ontology attributes comprise ID, name, address, picture, longitude and latitude, content background and related content, and the attribute data types are shown in Table 3.
The system adopts a design mode facing micro-services to carry out platform design, divides a core service line of the system into user identity verification service, user authority control service, poetry characteristic entity extraction service, poetry knowledge reasoning and retrieval service, scenic spot poetry text image recognition service and poetry knowledge recommendation service based on an SOA architecture, and adopts Restful standard design and realizes an API interface. And storing user information and system log records by using a MySQL object relational database, and performing distributed caching by using Redis in consideration of the expandability and high concurrency support of a platform. The knowledge service application of the system platform is packaged by using a Docker container technology, so that the distributed application is convenient for deployment, and the knowledge service system has high transportability and high expandability. The Kubernetes platform management container is adopted, so that the system platform can realize automatic deployment, expansion and management, and the system has high availability.
The knowledge application module comprises common multi-user services, such as: user login, user registration and historical retrieval record query management; knowledge retrieval and reasoning, and using an image text detection and recognition model to assist the knowledge reasoning; recommending poetry content resource knowledge bodies related to the poetry body knowledge according to the poetry body knowledge, if: related books, related poems, and related scenic spots.
Table 1 scenic spot structured knowledge representation example
Figure BDA0003116106210000111
Table 2 example of representation of characteristic entities of poetry in scenic region
Figure BDA0003116106210000121
Table 3 attribute data type table
Attribute name Type (B) Description of the invention
ID int Knowledge point ID
Name string Knowledge point name
Image string Poetry picture
Address string Geographic location
LatitudeLongitude string Latitude and longitude
Background string Background
RelatedContent string Related content
Fig. 5 shows a process of performing cloud recognition service on knowledge of scenic spots by a user, the user logs in the system through identity verification to input poetry characteristic information and uploads a poetry text image of the scenic spot, the knowledge inference module and the poetry image recognition module obtain ontology knowledge of related poetry, the ontology knowledge of poetry complete resources is obtained through inference according to the ontology knowledge, the knowledge is integrated and pushed to the user, and a knowledge service flow is completed. Fig. 6 shows a flow chart of result matching of poetry in a scenic region. Fig. 7 shows a deep learning and scenic spot knowledge map fusion architecture diagram.

Claims (2)

1.深度学习与景区知识图谱融合的手写诗词识别方法,其特征在于,包括以下步骤:1. The handwritten poetry recognition method of deep learning and scenic spot knowledge map fusion, is characterized in that, comprises the following steps: 步骤1、获取需要识别的景区手写诗词文本图像以及该图像的相关属性信息,包括图像拍摄地理位置、图像背景纹理、文字方向,并将图像通过空间域增强算法减弱图像中的噪声,使原图像与去噪图像的结构相似性指数大于0.9,得到预处理结果;Step 1. Obtain the handwritten poetry text image of the scenic spot to be recognized and the relevant attribute information of the image, including the location of the image, the background texture of the image, and the text direction, and reduce the noise in the image through the spatial domain enhancement algorithm, so that the original image can be reduced. The structural similarity index with the denoised image is greater than 0.9, and the preprocessing result is obtained; 步骤2、将预处理后的景区手写诗词图像输入基于VGG16的特征提取网络进行诗词文本的特征提取,并且通过诗词文本分类器获得景区图像的手写诗词文本权重序列,同时对分类器进行训练;Step 2. Input the preprocessed scenic spot handwritten poem image into the feature extraction network based on VGG16 to extract the feature of the poem text, and obtain the handwritten poem text weight sequence of the scenic spot image through the poem text classifier, and train the classifier at the same time; 步骤3、将提取到的手写诗词文本特征图输入基于FPN的手写诗词检测网络进行融合,得到手写诗词文本的单字符高斯热力图,进而经过多字符文本框链接算法得到景区手写诗词图像的文本区域位置信息;Step 3. Input the extracted feature map of handwritten poetry text into the FPN-based handwritten poetry detection network for fusion to obtain a single-character Gaussian heat map of the handwritten poetry text, and then obtain the text area of the scenic handwritten poetry image through the multi-character text box link algorithm. location information; 步骤4、根据提取到的文本图像位置信息对景区手写诗词图像进行区域裁剪,并依次将裁剪区域输入基于ACE的手写诗词识别网络进行手写诗词文本的Encoder-Decoder处理,得到景区手写诗词图像的识别结果;Step 4. According to the extracted text image position information, perform regional cropping on the handwritten poetry image of the scenic spot, and then input the cropped region into the ACE-based handwritten poetry recognition network to perform the Encoder-Decoder processing of the handwritten poetry text, so as to obtain the recognition of the handwritten poetry image in the scenic spot. result; 步骤5、将景区手写诗词图像的相关属性信息输入到景区知识图谱中经过图搜索得到搜索知识结果集,利用步骤4得到的识别结果与搜索知识结果集通过景区手写诗词匹配算法,得出最终的景区手写诗词图像的文本识别结果。Step 5. Input the relevant attribute information of the handwritten poem image of the scenic spot into the knowledge map of the scenic spot to obtain a search knowledge result set through graph search, and use the recognition result obtained in step 4 and the search knowledge result set to pass the scenic spot handwritten poem matching algorithm to obtain the final result. Text recognition results of handwritten poetry images in scenic spots. 2.根据权利要求1所述的深度学习与景区知识图谱融合的手写诗词识别方法,其特征在于,2. the deep learning according to claim 1 and the handwritten poetry recognition method of scenic spot knowledge map fusion, it is characterised in that, (1)基于FPN的景区手写诗词的检测技术(1) Detection technology of handwritten poems in scenic spots based on FPN 一共分为3个部分,分别为:手写诗词空间特征提取网络、字符关键点标定算法以及多字符文本框链接算法;首先将景区手写诗词图像输入到手写诗词空间特征提取网络对手写诗词文本特征进行提取;然后将提取到的诗词文本特征通过字符关键点标定算法对单字符关键点进行标记;最后将标记好的单字符关键点经过多字符文本框链接算法处理后得出景区手写诗词文本区域坐标信息,将该信息作为输入传递给景区手写诗词图像文本字符识别模块;其详细描述如下:It is divided into three parts: handwritten poetry spatial feature extraction network, character key point calibration algorithm and multi-character text box link algorithm; first, the scenic handwritten poetry image is input into the handwritten poetry spatial feature extraction network to carry out the text features of handwritten poetry. Extraction; then mark the single-character key points with the extracted poem text features through the character key point calibration algorithm; finally, the marked single-character key points are processed by the multi-character text box link algorithm to obtain the coordinates of the handwritten poem text area in the scenic spot information, and pass the information as input to the scenic spot handwritten poem image text character recognition module; its detailed description is as follows: 通过VGG16对景区手写诗词图像进行特征提取,得到景区手写诗词图像的诗词文本特征图,多特征模块融合的骨干网络使用的是特征金字塔网络;采用了7层金字塔网络,层与层之间相互连接,促进特征复用;在的金字塔网络中,利用一个自下而上的路径和一个自上而下的路径水平配对进行数据传输;1)自下而上通路:是指数据的向上流动;每层均包含卷积层、池化层、激活函数层以及循环层;该通路得到7个多尺度特征图,标记为{c1,c2,c3,c4,c5,c6,c7},不同特征图记录不同层次的文字特征,低层特征反映较浅层次的诗词文字特征边界,高层特征则反映较深层次的诗词文字特征;The feature extraction of handwritten poem images in scenic spots is carried out through VGG16, and the feature maps of poem texts of handwritten poem images in scenic spots are obtained. The backbone network of multi-feature module fusion uses feature pyramid network; 7-layer pyramid network is used, and the layers are connected to each other. , to promote feature multiplexing; in the pyramid network, a bottom-up path and a top-down path are used for horizontal pairing for data transmission; 1) Bottom-up path: refers to the upward flow of data; each The layers all contain convolutional layers, pooling layers, activation function layers, and recurrent layers; this path obtains 7 multi-scale feature maps, marked as {c 1 ,c 2 ,c 3 ,c 4 ,c 5 ,c 6 ,c 7 }, different feature maps record text features at different levels, low-level features reflect shallower-level poetry and text feature boundaries, and high-level features reflect deeper-level poetry and text features; 2)自上而下通路:是指数据的向下流动,通过对特征图进行上采样,前面几层的输出作为当前层的输入,同时,增加可变形卷积模块Def-Incept,提取部分变形文本的特征,然后生成多层特征图{p1,p2,p3,p4,p5,p6,p7},最后对特征图进行一次卷积运算,减少了参数数量,消除了变形带来的混杂效应;以下公式(1)展示了特征提取的过程;2) Top-down path: refers to the downward flow of data. By upsampling the feature map, the output of the previous layers is used as the input of the current layer. At the same time, the deformable convolution module Def-Incept is added to extract partial deformation. The features of the text, and then generate a multi-layer feature map {p 1 ,p 2 ,p 3 ,p 4 ,p 5 ,p 6 ,p 7 }, and finally perform a convolution operation on the feature map, which reduces the number of parameters and eliminates the The confounding effect caused by deformation; the following formula (1) shows the process of feature extraction;
Figure FDA0003116106200000031
Figure FDA0003116106200000031
式中,Conv表示卷积操作,
Figure FDA0003116106200000032
表示特征融合,UpSample表示上采样操作,Def-Incept表示可变形卷积操作;将本操作的输出信息通过高斯核函数处理,得到图像文本的字符关键节点的高斯热力图;
where Conv represents the convolution operation,
Figure FDA0003116106200000032
Represents feature fusion, UpSample represents upsampling operation, and Def-Incept represents deformable convolution operation; the output information of this operation is processed through the Gaussian kernel function to obtain the Gaussian heat map of the key nodes of the characters of the image text;
多字符文本框链接算法是将得到图像文本的字符关键节点的高斯热力图作为前提条件,通过计算字符关键节点之间的链接关系得出最终的景区手写诗词图像文本检测框;以下将对多字符文本框链接算法的计算过程进行详细介绍;The multi-character text box linking algorithm is to obtain the Gaussian heat map of the character key nodes of the image text as a precondition, and obtain the final scenic spot handwritten poetry image text detection frame by calculating the link relationship between the character key nodes; The calculation process of the text box link algorithm is introduced in detail; 首先,通过对景区手写诗词图像文本中的每个字符通过字符关键点高斯热力图进行分析计算,得出每个字符的高斯热力图的最大直径,以最大直径为边长画出正方形框,用来标记单个字符的文本框位置;其次,选取单字符文本框的对角线长度的一半作为向外辐射圆的半径r的初始值,并根据输入图像的最大边长设定半径r的最大值max;然后,以步长为
Figure FDA0003116106200000033
的幅度不断增大辐射圆的半径,在用户输入的文字方向进行探索,若遇到另一个字符文本框则停止,并将本字符文本框与相遇文本框的中心进行链接;若半径r达到最大值则表示文本框链接结束;最后,将链接起来的文本框进行整合,得出最终的景区手写诗词图像文本检测区域位置结果;
First, by analyzing and calculating each character in the handwritten poetry image text of the scenic spot through the character key point Gaussian heat map, the maximum diameter of the Gaussian heat map of each character is obtained, and a square box is drawn with the maximum diameter as the side length. to mark the text box position of a single character; secondly, half of the diagonal length of the single character text box is selected as the initial value of the radius r of the outward radiating circle, and the maximum value of the radius r is set according to the maximum side length of the input image max; then, the step size is
Figure FDA0003116106200000033
The amplitude of the radiating circle continuously increases the radius of the radiating circle, and explores in the direction of the text input by the user. If it encounters another character text box, it stops and links the character text box with the center of the encounter text box; if the radius r reaches the maximum The value indicates the end of the text box link; finally, the linked text boxes are integrated to obtain the final location result of the scenic spot handwritten poem image text detection area;
(2)基于ACE的景区手写诗词的识别方法(2) Recognition method of handwritten poems in scenic spots based on ACE 通过使用景区手写诗词的文本识别网络中,首先将景区手写诗词文本检测得到的局部文本区域经过图像归一化处理,使数据更加规范;然后将处理后的图像数据输入到手写诗词字符特征提取网络对诗词文本特征进行序列化编码;最后将诗词文本特征序列化后的编码通过字符识别器进行解码得到初步的景区手写诗词文本识别结果;其详细描述如下:By using the text recognition network of handwritten poems in scenic spots, the local text regions detected by handwritten poems in scenic spots are first processed by image normalization to make the data more standardized; then the processed image data is input into the character feature extraction network of handwritten poems The features of the poem text are serialized and encoded; finally, the serialized code of the poem text features is decoded by the character recognizer to obtain the preliminary scenic spot handwritten poem text recognition results; the detailed description is as follows: 对景区手写诗词文本图像进行归一化处理后,将针对处理后的文本图像进行字符特征提取,将以卷积循环神经网络CRNN为主干网络对景区手写诗词图像文本进行特征提取,该字符特征提取网络结构主要由卷积层和循环网络层两部分构成;首先,将归一化处理之后的诗词文本图像输入到卷积层,提取到图像的卷积特征图;然后,将提取到的卷积特征图输入到循环网络层,用于在卷积特征的基础上继续提取文字序列特征,将提取到的特征进行输出;After normalizing the text images of handwritten poems in scenic spots, character features are extracted from the processed text images, and the convolutional recurrent neural network CRNN is used as the backbone network to extract features from the image texts of handwritten poems in scenic spots. The network structure is mainly composed of two parts: the convolution layer and the recurrent network layer; first, the normalized poem text image is input into the convolution layer, and the convolution feature map of the image is extracted; then, the extracted convolution The feature map is input to the recurrent network layer, which is used to continue to extract text sequence features based on convolution features, and output the extracted features; 将采用聚合交叉熵ACE算法对景区手写诗词图像文本的特征序列进行解码,在模型训练的初始阶段,将增益系数设置为1;在之后的训练阶段,则将增益系数设置为该图像中文本的字符个数;The aggregated cross-entropy ACE algorithm will be used to decode the feature sequence of the image text of handwritten poems in scenic spots. In the initial stage of model training, the gain coefficient is set to 1; in the subsequent training stage, the gain coefficient is set to the value of the text in the image. number of characters; (3)基于知识图谱的诗词矫正技术(3) Poetry correction technology based on knowledge graph 针对用户查询诗词描述信息作为搜索条件,利用景区知识图谱进行图搜索,得到搜索结果集C;Aiming at the user's query of poem description information as a search condition, the knowledge map of scenic spots is used to perform graph search, and the search result set C is obtained; 其次,将搜索结果集C与景区手写诗词文本图像的初步识别结果x通过分词算法f(·)进行分词后,将得到的搜索结果集关键词矩阵Sn×m以及景区手写诗词文本图像初步识别结果矩阵E1×m,其中n表示搜索结果集中的实体数量,m表示处理后的关键字数量,k表示搜索结果集中实体的索引位置,f(·)表示对文本进行分词的分词函数,C[k]表示搜索结果集中的单个文本,x表示景区手写诗词文本图像的初步识别结果;其计算公式如(2)与(3)所示;Secondly, after the search result set C and the preliminary recognition result x of the handwritten poem text image of the scenic spot are segmented by the word segmentation algorithm f( ), the keyword matrix S n×m of the obtained search result set and the handwritten poem text image of the scenic spot are preliminarily recognized. The result matrix E 1×m , where n represents the number of entities in the search result set, m represents the number of processed keywords, k represents the index position of the entity in the search result set, f( ) represents the word segmentation function for text segmentation, C [k] represents a single text in the search result set, and x represents the preliminary recognition result of the text image of handwritten poems in scenic spots; its calculation formula is shown in (2) and (3); Sk×m=f(C[k]) (2)S k×m = f(C[k]) (2) E1×m=f(x) (3)E 1×m = f(x) (3) 然后,利用产生词向量的分布模型对前文所得到的矩阵中的实体关键字进行计算;其每个实体关键字向量ve的计算公式如(4)所示,其中e表示矩阵中的单个样本,g(·)表示产生词向量的分布模型函数,i表示实体关键字的索引位置,Sn×m[e]表示搜索结果集关键词矩阵中单个样本数据,Sn×m[e][i]表示逐步取搜索结果集关键词矩阵中单个样本的实体关键字数据;Then, use the distribution model for generating word vectors to calculate the entity keywords in the matrix obtained above; the calculation formula of each entity keyword vector ve is shown in (4), where e represents a single sample in the matrix , g( ) represents the distribution model function for generating the word vector, i represents the index position of the entity keyword, S n×m [e] represents the single sample data in the keyword matrix of the search result set, S n×m [e][ i] represents step by step to obtain the entity keyword data of a single sample in the keyword matrix of the search result set;
Figure FDA0003116106200000051
Figure FDA0003116106200000051
将得到的诗词文本的向量ve经过归一化处理后,得到诗词文本的最终文本向量v,将n个文本向量进行组合生成搜索结果向量集V;其计算公式如(5)所示;After normalizing the obtained vector ve of the poem text, the final text vector v of the poem text is obtained, and the n text vectors are combined to generate the search result vector set V; the calculation formula is shown in (5);
Figure FDA0003116106200000052
Figure FDA0003116106200000052
将景区手写诗词文本图像初步识别结果也经过上述处理后,得到最终的景区手写诗词文本图像初步识别结果向量q;After the preliminary recognition result of the handwritten poem text image in the scenic spot is also processed above, the final initial recognition result vector q of the handwritten poem text image in the scenic spot is obtained; 最后,利用景区手写诗词文本图像初步识别结果向量q与搜索结果向量集V中每个实体的文本向量v通过VSM进行计算两个诗词文本向量的相似程度,取景区手写诗词文本图像初步识别结果向量q与搜索结果向量集V中相似度最高的诗词文本向量,其对应的搜索结果集C中的文本Result进行输出,作为景区手写诗词图像文本识别结果;其计算公式如(6)和(7)所示,其中VSM(·)表示向量空间模型计算函数,j表示单个搜索结果向量的索引位置,s1和s2为两个文本,其单词频率分别用at和bt表示,t表示文本单词频率的索引位置;Finally, use the initial recognition result vector q of the handwritten poem text image in the scenic spot and the text vector v of each entity in the search result vector set V to calculate the similarity between the two poem text vectors through VSM, and select the initial recognition result vector of the handwritten poem text image in the scenic spot. q is the poem text vector with the highest similarity in the search result vector set V, and the corresponding text Result in the search result set C is output as the text recognition result of the handwritten poem image text in the scenic spot; its calculation formulas are as follows (6) and (7) where VSM( ) represents the vector space model calculation function, j represents the index position of a single search result vector, s 1 and s 2 are two texts, and their word frequencies are represented by a t and b t respectively, and t represents the text the index position of the word frequency;
Figure DEST_PATH_IMAGE001
Figure DEST_PATH_IMAGE001
Figure DEST_PATH_IMAGE002
Figure DEST_PATH_IMAGE002
利用知识推理方法获取与检索结果Result知识节点关系为“关联”的知识节点,三元组表示法可描述为(?,关联,Result),通过知识推理与计算得到相关诗词知识节点信息,实现知识推荐服务。Use the knowledge reasoning method to obtain knowledge nodes whose relationship with the retrieval result Result knowledge node is "associated", and the triple representation can be described as (?, association, Result). Through knowledge reasoning and calculation, the relevant poetry knowledge node information is obtained to realize knowledge Recommended service.
CN202110663733.3A 2021-06-16 2021-06-16 Handwritten poetry recognition method based on deep learning and scenic spot knowledge graph Active CN113609892B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110663733.3A CN113609892B (en) 2021-06-16 2021-06-16 Handwritten poetry recognition method based on deep learning and scenic spot knowledge graph

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110663733.3A CN113609892B (en) 2021-06-16 2021-06-16 Handwritten poetry recognition method based on deep learning and scenic spot knowledge graph

Publications (2)

Publication Number Publication Date
CN113609892A true CN113609892A (en) 2021-11-05
CN113609892B CN113609892B (en) 2025-03-11

Family

ID=78303520

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110663733.3A Active CN113609892B (en) 2021-06-16 2021-06-16 Handwritten poetry recognition method based on deep learning and scenic spot knowledge graph

Country Status (1)

Country Link
CN (1) CN113609892B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114332896A (en) * 2021-12-31 2022-04-12 北京思图场景数据科技服务有限公司 Key text information extraction method, device, electronic device and storage medium
CN114707005A (en) * 2022-06-02 2022-07-05 浙江建木智能系统有限公司 Knowledge graph construction method and system for ship equipment
CN114792423A (en) * 2022-05-20 2022-07-26 北京百度网讯科技有限公司 Document image processing method and device and storage medium
CN116051132A (en) * 2023-04-03 2023-05-02 之江实验室 Illegal commodity identification method and device, computer equipment and storage medium
CN117132997A (en) * 2023-10-26 2023-11-28 国网江西省电力有限公司电力科学研究院 Handwriting form recognition method based on multi-head attention mechanism and knowledge graph

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108549893A (en) * 2018-04-04 2018-09-18 华中科技大学 A kind of end-to-end recognition methods of the scene text of arbitrary shape
CN111062386A (en) * 2019-11-28 2020-04-24 大连交通大学 Natural scene text detection method based on depth pyramid attention and feature fusion
CN111259897A (en) * 2018-12-03 2020-06-09 杭州翼心信息科技有限公司 Knowledge-aware text recognition method and system
CN111461114A (en) * 2020-03-03 2020-07-28 华南理工大学 Multi-scale feature pyramid text detection method based on segmentation
CN111860348A (en) * 2020-07-21 2020-10-30 国网山东省电力公司青岛供电公司 OCR recognition method of weakly supervised power drawings based on deep learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108549893A (en) * 2018-04-04 2018-09-18 华中科技大学 A kind of end-to-end recognition methods of the scene text of arbitrary shape
CN111259897A (en) * 2018-12-03 2020-06-09 杭州翼心信息科技有限公司 Knowledge-aware text recognition method and system
CN111062386A (en) * 2019-11-28 2020-04-24 大连交通大学 Natural scene text detection method based on depth pyramid attention and feature fusion
CN111461114A (en) * 2020-03-03 2020-07-28 华南理工大学 Multi-scale feature pyramid text detection method based on segmentation
CN111860348A (en) * 2020-07-21 2020-10-30 国网山东省电力公司青岛供电公司 OCR recognition method of weakly supervised power drawings based on deep learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
FAGU I LI U等: "FTPN: Scene Text Detection With Feature Pyramid Based Text Proposal Network", IEEE ACCESS, vol. 7, 9 April 2019 (2019-04-09), pages 44219 - 44228, XP011718955, DOI: 10.1109/ACCESS.2019.2908933 *
WEI JIA等: "A CNN-based Approach to Detecting Text from Images of Whiteboards and Handwritten Notes", 2018 16TH INTERNATIONAL CONFERENCE ON FRONTIERS IN HANDWRITING RECOGNITION, 5 August 2018 (2018-08-05), pages 1 - 6, XP033468553, DOI: 10.1109/ICFHR-2018.2018.00010 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114332896A (en) * 2021-12-31 2022-04-12 北京思图场景数据科技服务有限公司 Key text information extraction method, device, electronic device and storage medium
CN114792423A (en) * 2022-05-20 2022-07-26 北京百度网讯科技有限公司 Document image processing method and device and storage medium
CN114707005A (en) * 2022-06-02 2022-07-05 浙江建木智能系统有限公司 Knowledge graph construction method and system for ship equipment
CN114707005B (en) * 2022-06-02 2022-10-25 浙江建木智能系统有限公司 Knowledge graph construction method and system for ship equipment
CN116051132A (en) * 2023-04-03 2023-05-02 之江实验室 Illegal commodity identification method and device, computer equipment and storage medium
CN117132997A (en) * 2023-10-26 2023-11-28 国网江西省电力有限公司电力科学研究院 Handwriting form recognition method based on multi-head attention mechanism and knowledge graph
CN117132997B (en) * 2023-10-26 2024-03-12 国网江西省电力有限公司电力科学研究院 Handwriting form recognition method based on multi-head attention mechanism and knowledge graph

Also Published As

Publication number Publication date
CN113609892B (en) 2025-03-11

Similar Documents

Publication Publication Date Title
CN113283551B (en) Training method and training device of multi-mode pre-training model and electronic equipment
CN113609892A (en) Handwritten poetry recognition method integrating deep learning with scenic spot knowledge map
CN111460247B (en) Automatic detection method for network picture sensitive characters
CN112949647B (en) Three-dimensional scene description method and device, electronic equipment and storage medium
CN111461039B (en) Landmark identification method based on multi-scale feature fusion
CN107203775B (en) A method, apparatus and device for image classification
CN113378609B (en) Recognition method and device for agent's signature
CN116226785A (en) Target object recognition method, multi-mode recognition model training method and device
CN110490081A (en) A kind of remote sensing object decomposition method based on focusing weight matrix and mutative scale semantic segmentation neural network
CN109299303B (en) Hand-drawn sketch retrieval method based on deformable convolution and deep network
CN108734159B (en) Method and system for detecting sensitive information in image
CN116975615A (en) Task prediction method and device based on video multi-mode information
CN109684928A (en) Chinese document recognition methods based on Internal retrieval
CN117931858B (en) Data query method, device, computer equipment and storage medium
CN113468332A (en) Classification model updating method and corresponding device, equipment and medium
CN113468891A (en) Text processing method and device
CN105678244A (en) Approximate video retrieval method based on improvement of editing distance
CN111651674B (en) Bidirectional searching method and device and electronic equipment
CN113469272A (en) Target detection method of hotel scene picture based on fast R-CNN-FFS model
CN109766918A (en) Salient object detection method based on multi-level context information fusion
CN116955686A (en) Image processing method, device, computer equipment and storage medium
CN115775349A (en) False news detection method and device based on multi-mode fusion
CN112801217B (en) Text similarity judgment method and device, electronic equipment and readable storage medium
CN119131803A (en) A dual-modal image-text fusion target recognition method and system
CN113961733A (en) Image and text retrieval method, device, electronic device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant