[go: up one dir, main page]

CN111428749B - Pre-verification method, device and equipment for image annotation task and storage medium - Google Patents

Pre-verification method, device and equipment for image annotation task and storage medium Download PDF

Info

Publication number
CN111428749B
CN111428749B CN202010106306.0A CN202010106306A CN111428749B CN 111428749 B CN111428749 B CN 111428749B CN 202010106306 A CN202010106306 A CN 202010106306A CN 111428749 B CN111428749 B CN 111428749B
Authority
CN
China
Prior art keywords
image information
text box
anchor point
data
equation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010106306.0A
Other languages
Chinese (zh)
Other versions
CN111428749A (en
Inventor
王健宗
李佳琳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ping An Technology Shenzhen Co Ltd
Original Assignee
Ping An Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ping An Technology Shenzhen Co Ltd filed Critical Ping An Technology Shenzhen Co Ltd
Priority to CN202010106306.0A priority Critical patent/CN111428749B/en
Publication of CN111428749A publication Critical patent/CN111428749A/en
Priority to PCT/CN2020/117379 priority patent/WO2021164251A1/en
Application granted granted Critical
Publication of CN111428749B publication Critical patent/CN111428749B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The embodiment of the application belongs to the technical field of artificial intelligence, and relates to a pre-checking method of an image annotation task, which comprises the following steps: receiving a pre-verification request sent by a user terminal, wherein the pre-verification request at least carries original image information and data annotation information; performing shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information; judging whether the data annotation information meets the preset anchor point requirement or not; and if the preset anchor point requirement is met, outputting a pre-verification success signal to the user terminal. The application also provides a pre-verification device, computer equipment and a storage medium for the image annotation task. According to the application, whether the data marking information of the user meets the requirement of the anchor point is judged by acquiring the anchor point data, so that the accuracy of the data marking information submitted by the user can be improved, the workload of a data processing platform is further reduced, and the working efficiency of the data processing platform is improved.

Description

Pre-verification method, device and equipment for image annotation task and storage medium
Technical Field
The present application relates to the field of artificial intelligence technologies, and in particular, to a method and apparatus for pre-verifying an image labeling task, a computer device, and a storage medium.
Background
In recent years, with the continuous development of a data processing platform, the types of image annotation tasks are more and more diversified. In order to meet the accuracy requirements of different types of image annotation tasks, a targeted task circulation mechanism and a verification mechanism are often required.
The existing verification method of the image annotation task is that after data annotation information submitted by a user is received, the data annotation information is cut, the cut annotation information is compared by a verification system, and data annotations with the same content are screened out.
However, the traditional verification method is generally not intelligent, when a user receives an image annotation task, the image annotation task can be arbitrarily annotated, and the data annotation information submitted by the user is generally low in accuracy, so that the workload of a data processing platform is increased, and the working efficiency of the data processing platform is further reduced.
Disclosure of Invention
The embodiment of the application aims to provide a pre-calibration method of an image labeling task, and aims to solve the problems of low general accuracy and low working efficiency of the conventional image labeling task calibration method.
In order to solve the technical problems, the embodiment of the application provides a pre-checking method for an image annotation task, which adopts the following technical scheme:
Receiving a pre-verification request sent by a user terminal, wherein the pre-verification request at least carries original image information and data annotation information;
performing shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information;
judging whether the data annotation information meets the preset anchor point requirement or not;
and if the preset anchor point requirement is met, outputting a pre-verification success signal to the user terminal.
In order to solve the technical problems, the embodiment of the application also provides a pre-verification device for the image annotation task, which adopts the following technical scheme:
The request receiving module is used for receiving a pre-verification request sent by the user terminal, wherein the pre-verification request at least carries original image information and data annotation information;
The anchor point acquisition module is used for carrying out shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information;
the marking judging module is used for judging whether the data marking information meets the preset anchor point requirement;
and the result output module is used for outputting a pre-verification success signal to the user terminal if the preset anchor point requirement is met.
In order to solve the above technical problems, the embodiment of the present application further provides a computer device, which adopts the following technical schemes:
Comprising a memory and a processor, the computer-readable storage medium having stored thereon a computer program which, when executed by the processor, performs the steps of the pre-verification method of the image annotation task as described above.
In order to solve the above technical problems, an embodiment of the present application further provides a computer readable storage medium, which adopts the following technical schemes:
The computer readable storage medium has stored thereon a computer program which, when executed by a processor, implements the steps of the pre-verification method of the image annotation task described above.
Compared with the prior art, the embodiment of the application has the following main beneficial effects:
The invention provides a pre-checking method of an image labeling task, which is used for receiving a pre-checking request sent by a user terminal, wherein the pre-checking request at least carries original image information and data labeling information; performing shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information; judging whether the data annotation information meets the preset anchor point requirement or not; and if the preset anchor point requirement is met, outputting a pre-verification success signal to the user terminal. Whether the data marking information of the user meets the requirement of the anchor point is judged by acquiring the anchor point data, so that the accuracy of the data marking information submitted by the user can be improved, the workload of a data processing platform is further reduced, and the working efficiency of the data processing platform is improved.
Drawings
In order to more clearly illustrate the solution of the present application, a brief description will be given below of the drawings required for the description of the embodiments of the present application, it being apparent that the drawings in the following description are some embodiments of the present application, and that other drawings may be obtained from these drawings without the exercise of inventive effort for a person of ordinary skill in the art.
FIG. 1 is a flowchart of a pre-calibration method for image annotation task according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a data annotation according to a first embodiment of the present invention;
FIG. 3 is a schematic diagram of another data annotation information provided in accordance with an embodiment of the present invention;
FIG. 4 is a schematic diagram of an anchor requirement provided by a first embodiment of the present invention;
FIG. 5 is a flow chart of an implementation of step S102 in FIG. 1;
FIG. 6 is a flow chart of an implementation of step S103 in FIG. 1;
FIG. 7 is a flow chart of an implementation of step S303 in FIG. 6;
FIG. 8 is a flowchart of an implementation of step S402 in FIG. 7;
Fig. 9 is a schematic structural diagram of a pre-verification device for image labeling task according to a second embodiment of the present invention;
Fig. 10 is a schematic structural diagram of an anchor point acquisition module according to a second embodiment of the present invention;
FIG. 11 is a schematic structural view of one embodiment of a computer device according to the present application.
Detailed Description
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs; the terminology used in the description of the applications herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the application; the terms "comprising" and "having" and any variations thereof in the description of the application and the claims and the description of the drawings above are intended to cover a non-exclusive inclusion. The terms first, second and the like in the description and in the claims or in the above-described figures, are used for distinguishing between different objects and not necessarily for describing a sequential or chronological order.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the application. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those of skill in the art will explicitly and implicitly appreciate that the embodiments described herein may be combined with other embodiments.
In order to make the person skilled in the art better understand the solution of the present application, the technical solution of the embodiment of the present application will be clearly and completely described below with reference to the accompanying drawings.
According to the pre-verification method for the image annotation task, which is provided by the embodiment of the invention, the accuracy of the data annotation information submitted by the user can be improved by acquiring the anchor point data and judging whether the data annotation information of the user meets the anchor point, so that the workload of a data processing platform is reduced, and the working efficiency of the data processing platform is improved.
Example 1
Fig. 1 shows a flowchart of an implementation of a pre-verification method for an image labeling task according to an embodiment of the present invention, and for convenience of explanation, only a portion relevant to the present invention is shown.
In step S101, a pre-verification request sent by a user terminal is received, where the pre-verification request at least carries original image information and data labeling information.
In the embodiment of the present invention, the user terminal may be a mobile terminal such as a mobile phone, a smart phone, a notebook computer, a digital broadcast receiver, a PDA (personal digital assistant), a PAD (tablet computer), a PMP (portable multimedia player), a navigation device, etc., and a fixed terminal such as a digital TV, a desktop computer, etc., it should be understood that the examples of the user terminal are only for convenience and are not limited to the present invention.
In the embodiment of the invention, the original image information refers to image information of some designated positions, such as positions of designated characters in the drawings, which need to be marked by a user in a line drawing manner, and generally limits the user to drawing only straight lines or different shapes composed of line segments.
In the embodiment of the invention, the data marking information refers to different shapes formed by combining the coordinates of the end points of the straight line and each line segment. As an example, as shown in fig. 2, the data annotation information is [ (2, 6), (5, 5) ]; as further shown in FIG. 3, the data labels are [ (2, 6), (5, 5) ], [ (5, 5), (7, 7) ], [ (7, 7), (4, 8) ] and [ (4, 8), (2, 6) ], and it should be understood that the examples of the data labels herein are for convenience only and are not intended to limit the present invention.
In step S102, a shunt identification operation is performed on the original image information based on a shunt identification algorithm, and anchor point data corresponding to the original image information is obtained.
In the embodiment of the invention, the shunt recognition algorithm refers to performing row cutting, full text recognition and column cutting operations on the original image information respectively, so that the original image is cut by taking characters in the original image information as units, and the recognized characters are taken as anchor point data.
In the embodiment of the invention, the anchor point refers to a mark arranged on the original image information and is used for setting a certain threshold range for standard data annotation so as to determine whether the data annotation information submitted by a user meets the threshold range.
In step S103, it is determined whether the data annotation information meets a preset anchor point requirement.
In the embodiment of the present invention, the anchor point is required to detect whether the data annotation information submitted by the user passes through a group of anchor points, as shown in fig. 4, and the anchor point requirements are (2, 5) and (5, 7), so that the data annotation information [ (2, 6), (5, 5) ] passes between the anchor points (2, 5) and (5, 7), that is, the data annotation information [ (2, 6), (5, 5) ] meets the anchor point requirement, and it should be understood that the anchor point requirement is merely for convenience in understanding and is not limited to the present invention.
In step S104, if the preset anchor point requirement is met, a pre-verification success signal is output to the user.
In the embodiment of the invention, the pre-verification signal can be output to the user through being displayed on a display device of the user terminal or through being displayed in a short message mode, and the method is not limited by the invention.
In the embodiment of the invention, the application provides a pre-checking method of an image labeling task, which is used for receiving a pre-checking request sent by a user terminal, wherein the pre-checking request at least carries original image information and data labeling information; performing shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information; judging whether the data annotation information meets the preset anchor point requirement or not; and if the preset anchor point requirement is met, outputting a pre-verification success signal to the user terminal. By acquiring anchor point data and judging whether the data marking information of the user meets the anchor point, the accuracy of the data marking information submitted by the user can be improved, the workload of a data processing platform is further reduced, and the working efficiency of the data processing platform is improved.
With continued reference to fig. 5, a flow chart for the implementation of step S102 in fig. 1 is shown, only the portions relevant to the present invention being shown for ease of illustration.
In some optional implementations of the first embodiment, the step S102 specifically includes: step S201, step S202, step S203, step S204, and step S205.
In step S201, a line cutting operation is performed on the original image information, and line image information is acquired.
In the embodiment of the present invention, the line cutting operation refers to dividing the original image in a line manner, where the height of the line is not specifically limited, and a user can adapt to the actual situation.
In the embodiment of the present invention, the line image information refers to image information after the original image is divided by a line form.
In step S202, the line image information is subjected to an identification operation, and a text box is acquired.
In the embodiment of the invention, the identification operation refers to identifying the text content in the line image information, and when the text content is identified, the text content position is identified in the form of a text box.
In step S203, a column cutting operation is performed on the original image information according to the text box, so as to obtain a plurality of columns of column image information carrying the text box.
In the embodiment of the present invention, the column cutting operation refers to dividing the original image by a column form, where, since the recognized text box is obtained in step S202, when the column cutting operation is performed on the original image information, the text box in the original image information is divided by taking the text box as a reference, so that the text content in the original image information is divided.
In step S204, a text box is selected from the column image information of each two adjacent columns to form a text box group.
In the embodiment of the invention, when only one text box exists in the column image information, the text box which exists only is selected to form the text box group; when there are a plurality of text boxes in the column image information, the user may obtain the text boxes forming the text box group in a limited selection manner according to the actual situation, for example, obtain the text boxes by a random selection manner, and select the text boxes by a relative distance manner, and it should be understood that the example of the obtaining manner is only convenient to understand, and is not limited to the embodiment of the present invention.
In step S205, the center point of the text box in the text box group is taken as the anchor point data.
In the embodiment of the invention, the line image information is acquired by carrying out line cutting operation on the original image information, so that the character content in the image information is conveniently identified; then, the text box is obtained through identification operation on the line image information, so that the specific position of the Chinese content in the original image information is determined; and finally, anchor point data is determined based on the text box, so that the effective range of the user annotation data is limited by the specification, and the annotation operation of the user is effectively specified.
In some optional implementations of the first embodiment of the present invention, the step S203 specifically includes the following steps:
The cutting is performed with a straight line having the smallest number of text boxes passing longitudinally as a column cut line.
In the embodiment of the invention, the lines with the least number of the text boxes passing through are selected as the column cutting lines for cutting, so that the number of the cut effective text boxes is ensured to be as large as possible, and the accuracy of anchor point setting is further ensured.
In some optional implementations of the first embodiment of the present invention, the step S204 specifically includes:
And taking two text boxes which are farthest from each other in the vertical axis direction in the column image information of the two adjacent columns as the text box group.
Fig. 6 is a flowchart showing the implementation of step S103 in fig. 1, and only the portions relevant to the present invention are shown for convenience of explanation.
In some optional implementations of the first embodiment of the present invention, the step S103 specifically includes a step S301, a step S302, a step S303, and a step S304.
In step S301, a first equation and a second equation corresponding to the data labeling information and the anchor point data are obtained in the same coordinate system.
In the embodiment of the invention, it is assumed that an algorithm finds a group of anchor points in original image information, and points A and B, and a wire in the annotation data annotation submitted by a user must pass between the points A and B, i.e. a line segment AB must have an intersection point with a line segment in the user data annotation.
Establishing a first equation and a second equation:
A first mode: assuming that the coordinates of point A are (x A,yA), the coordinates of point B are (x B,yB), and x A≤xB, the equation of a line segment can be calculated from the coordinates of two points on the line segment :(yB-yA)x+(xA-xB)y+xB*yA-xA*yB=0(xA≤x≤xB).
The second equation: according to the end point coordinates of each line segment in the data label submitted by the user, the equation of each line segment can be solved, for example, the coordinates of the end point M in the line segment MN (x M,yM), the coordinates of the end point N are (x N,yN) and x N≤xM, and the equation of the line segment MN is :(yN-yM)x+(xM-xN)y+xN*yM-xM*yN=0(xN≤x≤xM).
In step S302, it is determined whether an intersection exists on the line corresponding to the second equation and the first equation.
And (3) establishing a simultaneous equation set of the line segment AB and each line segment drawn by the user, and if the equation set is in a definition domain and only one solution exists, indicating that the line segment passes between the anchor points A and B. Taking line segment MN as an example, the method for judging whether the line segment MN passes between anchor points a and B is as follows:
① Simultaneous equations:
② The abscissa of the unique solution when the system of equations does not consider the domain is:
③ If the abscissa of the unique solution meets the definition field requirement x A≤x*≤xB and x N≤x*≤xM, the line segment MN and the line segment AB are considered to have an intersection point in the definition field, that is, a wired path between the anchor point a and the anchor point B.
In the embodiment of the present invention, x * meets the requirements of definition field x A≤x*≤xB and x N≤x*≤xM, that is, there is an intersection point on the second equation intersecting the first equation, which indicates that the data submitted by the user is marked with a line passing between point a and point B.
In step S303, if there is an intersection point on the line corresponding to the second equation and the first equation, the pre-verification success signal is output.
In step S304, if there is no intersection point on the line corresponding to the second equation and the first equation, a pre-verification failure signal is output.
In the embodiment of the invention, the judgment formula is formed by judging whether the two line segments are intersected or not and forming the coordinate and equation set, so that the answers submitted by the user can be well standardized, the probability of cheating the crowd-sourced user to get the points is reduced, the crowd-sourced platform can get more accurate answers in a shorter time, and the reliability of the answers of the image labeling class crowd-sourced task is improved.
Fig. 7 is a flowchart showing the implementation of step S303 in fig. 6, and only the portions relevant to the present invention are shown for convenience of explanation.
In some optional implementations of the first embodiment of the present invention, the step S303 specifically includes: step S401, step S402, and step S403.
In step S401, if the determination result is yes, it is determined whether the lines corresponding to the first equation and the second equation are parallel or overlap with each other.
In the embodiment of the present invention, taking the above-mentioned simultaneous equations as an example, the method may be implemented by determiningWhether or not to be equal toThereby determining whether the first equation and the second equation are parallel or coincident with each other.
In step S402, if the lines corresponding to the first equation and the second equation are not parallel or overlap, the pre-verification success signal is output.
In step S403, if the lines corresponding to the first equation and the second equation are parallel or overlap, the pre-verification failure signal is output.
Fig. 8 is a flowchart showing the implementation of step S402 in fig. 7, and only the portions relevant to the present invention are shown for convenience of explanation.
In some optional implementations of the first embodiment of the present invention, the step S402 specifically includes: step S501, step S502, step S503, and step S504. The original image information labeling task also carries invalid anchor point data.
In step S501, a third program and a fourth program corresponding to the data labeling information and the invalid anchor data are acquired in the same coordinate system.
In the embodiment of the invention, the invalid anchor point data refers to an area which cannot be passed through by the original image information marking task in advance, for example, the positions of a file signature, a file lifting head and the like of the original image information are only used for marking the file, the file is not associated with the original image information marking task, and when a user marks at the position, the requirement of the original image information marking task is obviously not met, so that the position of the invalid anchor point needs to be preset in the original image information marking task.
In the embodiment of the present invention, the third party program refers to an equation of no anchor point data co-ordination, and assuming that the coordinate of the point C of the invalid anchor point is (x C,yC), the coordinate of the point D is (x D,yD), and x C≤xD, the third party program for calculating a line segment according to the two-point coordinate on the line segment is :(yD-yC)x+(xC-xD)y+xD*yC-xC*yD=0(xC≤x≤xD).
In the embodiment of the present invention, the fourth aspect refers to an equation associated with the end points of the data annotation information submitted by the user, and it is assumed that the two end points of the data annotation information submitted by the user are Q (x Q,yQ),P(xP,yP):
then the fourth mode is :(yP-yQ)x+(xQ-xP)y+xP*yQ-xQ*yP=0(xQ≤x≤xP).
In step S502, it is determined whether there is an intersection on the line corresponding to the fourth program and the third program.
① Simultaneous equations:
② The abscissa of the unique solution when the system of equations does not consider the domain is:
③ If the abscissa of the unique solution meets the definition field requirement x C≤x*≤xD and x Q≤x*≤xP, then the intersection point of the line segment QP and the line segment CD is considered to exist in the definition field, that is, the line passes between the anchor point C and the anchor point D.
In step S503, if there is an intersection point on the line corresponding to the fourth program and the third program, the pre-verification failure signal is output.
In step S504, if there is no intersection point on the lines corresponding to the fourth program and the third program, the pre-verification success signal is output.
In summary, the present invention provides a pre-checking method for an image labeling task, which receives a pre-checking request sent by a user terminal, where the pre-checking request at least carries original image information and data labeling information; performing shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information; judging whether the data annotation information meets the preset anchor point requirement or not; and if the preset anchor point requirement is met, outputting a pre-verification success signal to the user terminal. By acquiring anchor point data and judging whether the data marking information of the user meets the anchor point requirement, the accuracy of the data marking information submitted by the user can be improved, the workload of a data processing platform is further reduced, and the working efficiency of the data processing platform is improved. Meanwhile, line image information is obtained through line cutting operation on original image information, so that character content in the image information is conveniently identified, a text box is obtained through identification operation on the line image information, so that the specific position of the character content in the original image information is determined, anchor point data are finally determined based on the text box, the effective range of user annotation data is standardized and limited, and the annotation operation of a user is effectively standardized; the judgment formula is formed by judging whether the two line segments are intersected or not and forming the mode of the coordinates and the equation set, so that answers submitted by users can be well standardized, the probability of cheating points by crowdsourcing users is reduced, a crowdsourcing platform can take more accurate answers in a shorter time, and the reliability of answers of labeling tasks of image labeling class crowdsourcing tasks is improved.
Those skilled in the art will appreciate that implementing all or part of the above-described methods in accordance with the embodiments may be accomplished by way of a computer program stored in a computer-readable storage medium, which when executed, may comprise the steps of the embodiments of the methods described above. The storage medium may be a nonvolatile storage medium such as a magnetic disk, an optical disk, a Read-Only Memory (ROM), or a random access Memory (Random Access Memory, RAM).
It should be understood that, although the steps in the flowcharts of the figures are shown in order as indicated by the arrows, these steps are not necessarily performed in order as indicated by the arrows. The steps are not strictly limited in order and may be performed in other orders, unless explicitly stated herein. Moreover, at least some of the steps in the flowcharts of the figures may include a plurality of sub-steps or stages that are not necessarily performed at the same time, but may be performed at different times, the order of their execution not necessarily being sequential, but may be performed in turn or alternately with other steps or at least a portion of the other steps or stages.
Example two
With further reference to fig. 9, as an implementation of the method shown in fig. 1, the present application provides a pre-verification device for an image labeling task, where an embodiment of the device corresponds to the embodiment of the method shown in fig. 1, and the device may be specifically applied to various electronic devices.
As shown in fig. 9, the pre-verification device 100 for image labeling task provided in the second embodiment of the present invention includes: the device comprises a request receiving module 101, an anchor point obtaining module 102, a marking judging module 103 and a result outputting module 104. Wherein:
A request receiving module 101, configured to receive a pre-verification request sent by a user terminal, where the pre-verification request at least carries original image information and data annotation information;
The anchor point obtaining module 102 is configured to perform a shunt identification operation on the original image information based on a shunt identification algorithm, and obtain anchor point data corresponding to the original image information;
The marking judging module 103 is used for judging whether the data marking information meets the preset anchor point requirement;
And the result output module 104 is configured to output a pre-verification success signal to the user terminal if the preset anchor point requirement is met.
In the embodiment of the present invention, the user terminal may be a mobile terminal such as a mobile phone, a smart phone, a notebook computer, a digital broadcast receiver, a PDA (personal digital assistant), a PAD (tablet computer), a PMP (portable multimedia player), a navigation device, etc., and a fixed terminal such as a digital TV, a desktop computer, etc., it should be understood that the examples of the user terminal are only for convenience and are not limited to the present invention.
In the embodiment of the invention, the original image information refers to image information of some designated positions, such as positions of designated characters in the drawings, which need to be marked by a user in a line drawing manner, and generally limits the user to drawing only straight lines or different shapes composed of line segments.
In the embodiment of the invention, the data marking information refers to different shapes formed by combining the coordinates of the end points of the straight line and each line segment. As an example, as shown in fig. 2, the data annotation information is [ (2, 6), (5, 5) ]; as further shown in FIG. 3, the data labels are [ (2, 6), (5, 5) ], [ (5, 5), (7, 7) ], [ (7, 7), (4, 8) ] and [ (4, 8), (2, 6) ], and it should be understood that the examples of the data labels herein are for convenience only and are not intended to limit the present invention.
In the embodiment of the invention, the shunt recognition algorithm refers to performing row cutting, full text recognition and column cutting operations on the original image information respectively, so that the original image is cut by taking characters in the original image information as units, and the recognized characters are taken as anchor point data.
In the embodiment of the invention, the anchor point refers to a mark arranged on the original image information and is used for setting a certain threshold range for standard data annotation so as to determine whether the data annotation information submitted by a user meets the threshold range.
In the embodiment of the present invention, before the anchor point is required to detect whether the data annotation information submitted by the user passes through a group of anchor points, as shown in fig. 4, the anchor point requirements are (2, 5) and (5, 7), and then the data annotation information [ (2, 6), (5, 5) ] passes between the anchor points (2, 5) and (5, 7), that is, the data annotation information [ (2, 6), (5, 5) ] meets the anchor point requirement, which is to be understood that the anchor point requirement is merely for convenience in understanding and is not to be used for limiting the present invention.
In the embodiment of the invention, the pre-verification signal can be output to the user through being displayed on a display device of the user terminal or through being displayed in a short message mode, and the method is not limited by the invention.
In an embodiment of the present invention, the present application provides a pre-verification device for an image labeling task, including: the request receiving module is used for receiving a pre-verification request sent by the user terminal, wherein the pre-verification request at least carries original image information and data annotation information; the anchor point acquisition module is used for carrying out shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information; the marking judging module is used for judging whether the data marking information meets the preset anchor point requirement; and the result output module is used for outputting the pre-verification success signal to the user terminal if the preset anchor point requirement is met. By acquiring anchor point data and judging whether the data marking information of the user meets the anchor point, the accuracy of the data marking information submitted by the user can be improved, the workload of a data processing platform is further reduced, and the working efficiency of the data processing platform is improved.
In some optional implementations of the second embodiment of the present invention, as shown in fig. 10, the anchor point obtaining module 102 includes: a row image acquisition sub-module 1021, a text box acquisition sub-module 1022, a column image acquisition sub-module 1023, a text box group acquisition sub-module 1024, and an anchor point determination sub-module 1025. Wherein:
A line image obtaining submodule 1021, configured to perform a line cutting operation on the original image information, and obtain line image information;
a text box obtaining sub-module 1022, configured to perform an identifying operation on the line image information, and obtain a text box;
A column image obtaining submodule 1023, configured to perform column cutting operation on the original image information according to the text box, and obtain a plurality of columns of column image information carrying the text box;
a text box group obtaining sub-module 1024, configured to select a text box from the column image information of every two adjacent columns to form a text box;
and an anchor point determining sub-module 1025, configured to use a center point of the text box in the text box group as the anchor point data.
In the embodiment of the present invention, the line cutting operation refers to dividing the original image in a line manner, where the height of the line is not specifically limited, and a user can adapt to the actual situation.
In the embodiment of the present invention, the line image information refers to image information after the original image is divided by a line form.
In the embodiment of the invention, the identification operation refers to identifying the text content in the line image information, and when the text content is identified, the text content position is identified in the form of a text box.
In the embodiment of the present invention, the column cutting operation refers to dividing the original image by a column form, where, since the recognized text box is obtained in step S202, when the column cutting operation is performed on the original image information, the text box in the original image information is divided by taking the text box as a reference, so that the text content in the original image information is divided.
In the embodiment of the invention, when only one text box exists in the column image information, the text box which exists only is selected to form the text box group; when there are a plurality of text boxes in the column image information, the user may obtain the text boxes forming the text box group in a limited selection manner according to the actual situation, for example, obtain the text boxes by a random selection manner, and select the text boxes by a relative distance manner, and it should be understood that the example of the obtaining manner is only convenient to understand, and is not limited to the embodiment of the present invention.
In the embodiment of the invention, the line image information is acquired by carrying out line cutting operation on the original image information, so that the character content in the image information is conveniently identified; then, the text box is obtained through identification operation on the line image information, so that the specific position of the Chinese content in the original image information is determined; and finally, anchor point data is determined based on the text box, so that the effective range of the user annotation data is limited by the specification, and the annotation operation of the user is effectively specified.
In some optional implementations of the second embodiment of the present invention, the text box group obtaining sub-module 1024 includes:
and the text box group determining submodule is used for taking two text boxes which are farthest from each other in the vertical axis direction in the column image information of the two adjacent columns as the text box group.
In some optional implementations of the second embodiment of the present invention, the column image capturing submodule 1023 includes:
And the column cutting line determining sub-module is used for cutting by taking a straight line with the least number of text boxes passing longitudinally as a column cutting line.
In some optional implementations of the second embodiment of the present invention, the labeling determining module 103 includes:
The first equation acquisition sub-module is used for acquiring a first equation and a second equation corresponding to the data marking information and the anchor point data;
A first equation judging sub-module for judging whether an intersection point intersecting with the first equation exists on the second equation;
A first signal output sub-module, configured to output the pre-verification success signal if an intersection point intersecting the first equation exists on the second equation;
And the second signal output sub-module is used for outputting a pre-verification failure signal if the intersection point which is intersected with the first equation does not exist on the second equation.
In some optional implementations of the second embodiment of the present invention, the first signal output submodule specifically includes:
the second equation judging sub-module is used for judging whether the first equation and the second equation are parallel or coincident with each other if the judging result is yes;
A third signal output sub-module, configured to output the pre-verification success signal if the first equation and the second equation are not parallel or coincident with each other;
And the fourth signal output sub-module is used for outputting the pre-verification failure signal if the first equation and the second equation are parallel or coincident with each other.
In some optional implementations of the second embodiment of the present invention, the third signal output submodule specifically includes:
a second equation obtaining sub-module, configured to obtain a third program and a fourth program corresponding to the data labeling information and the invalid anchor data;
a third-party program judging sub-module for judging whether an intersection point intersecting with the third-party program exists on the fourth program;
A fifth signal output sub-module, configured to output the pre-verification failure signal if an intersection point intersecting with a third party program exists on the fourth party program;
and a sixth signal output sub-module, configured to output the pre-verification success signal if there is no intersection point intersecting with the third party program on the fourth party program.
Example III
In order to solve the technical problems, the embodiment of the application also provides computer equipment. Referring specifically to fig. 11, fig. 11 is a basic structural block diagram of a computer device according to the present embodiment.
The computer device 11 comprises a memory 111, a processor 112, a network interface 113 communicatively coupled to each other via a system bus. It should be noted that only computer device 11 having components 111-113 is shown in the figures, but it should be understood that not all of the illustrated components are required to be implemented and that more or fewer components may be implemented instead. It will be appreciated by those skilled in the art that the computer device herein is a device capable of automatically performing numerical calculation and/or information processing according to a preset or stored instruction, and its hardware includes, but is not limited to, a microprocessor, an Application SPECIFIC INTEGRATED Circuit (ASIC), a Programmable gate array (Field-Programmable GATE ARRAY, FPGA), a digital Processor (DIGITAL SIGNAL Processor, DSP), an embedded device, and the like.
The computer equipment can be a desktop computer, a notebook computer, a palm computer, a cloud server and other computing equipment. The computer equipment can perform man-machine interaction with a user through a keyboard, a mouse, a remote controller, a touch pad or voice control equipment and the like.
The memory 111 includes at least one type of readable storage medium including flash memory, hard disk, multimedia card, card memory (e.g., SD or DX memory, etc.), random Access Memory (RAM), static Random Access Memory (SRAM), read Only Memory (ROM), electrically Erasable Programmable Read Only Memory (EEPROM), programmable Read Only Memory (PROM), magnetic memory, magnetic disk, optical disk, etc. In some embodiments, the storage 111 may be an internal storage unit of the computer device 11, such as a hard disk or a memory of the computer device 11. In other embodiments, the memory 111 may also be an external storage device of the computer device 11, such as a plug-in hard disk, a smart memory card (SMART MEDIA CARD, SMC), a Secure Digital (SD) card, a flash memory card (FLASH CARD) or the like, which are provided on the computer device 11. Of course, the memory 111 may also include both an internal storage unit of the computer device 11 and an external storage device thereof. In this embodiment, the memory 111 is generally used to store an operating system and various application software installed on the computer device 11, such as program codes of a pre-verification method of an image labeling task. In addition, the memory 111 may be used to temporarily store various types of data that have been output or are to be output.
The processor 112 may be a central processing unit (Central Processing Unit, CPU), controller, microcontroller, microprocessor, or other data processing chip in some embodiments. The processor 112 is typically used to control the overall operation of the computer device 11. In this embodiment, the processor 112 is configured to execute a program code stored in the memory 111 or process data, for example, a program code of a pre-verification method for executing the image labeling task.
The network interface 113 may comprise a wireless network interface or a wired network interface, which network interface 113 is typically used to establish communications connections between the computer device 11 and other electronic devices.
The present application also provides another embodiment, namely, a computer readable storage medium, where a pre-verification program of an image labeling task is stored, where the pre-verification program of the image labeling task is executable by at least one processor, so that the at least one processor performs the steps of the pre-verification method of the image labeling task as described above.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present application may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the method according to the embodiments of the present application.
It is apparent that the above-described embodiments are only some embodiments of the present application, but not all embodiments, and the preferred embodiments of the present application are shown in the drawings, which do not limit the scope of the patent claims. This application may be embodied in many different forms, but rather, embodiments are provided in order to provide a thorough and complete understanding of the present disclosure. Although the application has been described in detail with reference to the foregoing embodiments, it will be apparent to those skilled in the art that modifications may be made to the embodiments described in the foregoing description, or equivalents may be substituted for elements thereof. All equivalent structures made by the content of the specification and the drawings of the application are directly or indirectly applied to other related technical fields, and are also within the scope of the application.

Claims (8)

1. A method of pre-calibration of an image annotation task, the method comprising the steps of:
Receiving a pre-verification request sent by a user terminal, wherein the pre-verification request at least carries original image information and data annotation information;
performing shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information;
judging whether the data annotation information meets the preset anchor point requirement or not;
If the preset anchor point requirement is met, outputting a pre-verification success signal to the user terminal;
the step of obtaining anchor point data corresponding to the original image information by performing a shunt identification operation on the original image information based on a shunt identification algorithm specifically comprises the following steps:
performing line cutting operation on the original image information to obtain line image information;
performing identification operation on the line image information to obtain a text box;
Performing column cutting operation on the original image information according to the text box to obtain a plurality of columns of column image information carrying the text box;
selecting a text box from the column image information of every two adjacent columns to form a text box group;
taking the center point of the text box in the text box group as the anchor point data;
the step of selecting a text box from the column image information of every two adjacent columns to form a text box group comprises the following steps:
And taking two text boxes which are farthest from each other in the vertical axis direction in the column image information of the two adjacent columns as the text box group.
2. The method for pre-checking an image labeling task according to claim 1, wherein the step of performing a column cutting operation on the original image information according to the text box to obtain a plurality of columns of column image information carrying the text box specifically comprises the following steps:
The cutting is performed with a straight line having the smallest number of text boxes passing longitudinally as a column cut line.
3. The method for pre-checking an image labeling task according to claim 1, wherein the step of determining whether the data labeling information meets a preset anchor point requirement specifically comprises the following steps:
Acquiring a first equation and a second equation corresponding to the data marking information and the anchor point data in the same coordinate system;
Judging whether an intersection point exists on a line corresponding to the second equation and the first equation;
outputting the pre-verification success signal if the intersection point exists on the line corresponding to the second equation and the first equation;
If the second equation and the first equation have no intersection point on the corresponding line, outputting a pre-verification failure signal.
4. The method for pre-checking an image labeling task according to claim 3, wherein the step of outputting the pre-checking success signal if the determination result is yes comprises the following steps:
If the judgment result is yes, judging whether the lines corresponding to the first equation and the second equation are parallel or coincident with each other;
Outputting the pre-verification success signal if the lines corresponding to the first equation and the second equation are not parallel or coincident with each other;
Outputting the pre-verification failure signal if the lines corresponding to the first equation and the second equation are parallel or coincident with each other.
5. The method for pre-checking an image labeling task according to claim 4, wherein the image labeling task further carries invalid anchor point data, and the step of outputting the pre-checking success signal if the judgment result is negative comprises the following steps:
Acquiring a third program and a fourth program corresponding to the data marking information and the invalid anchor point data in the same coordinate system;
judging whether an intersection point exists on a line corresponding to the fourth program and the third program;
outputting the pre-verification failure signal if the intersection points exist on the lines corresponding to the fourth program and the third program;
And if the intersection points do not exist on the lines corresponding to the fourth program and the third program, outputting the pre-verification success signal.
6. A pre-verification device for an image annotation task, the device comprising:
The request receiving module is used for receiving a pre-verification request sent by the user terminal, wherein the pre-verification request at least carries original image information and data annotation information;
The anchor point acquisition module is used for carrying out shunt identification operation on the original image information based on a shunt identification algorithm to acquire anchor point data corresponding to the original image information;
the marking judging module is used for judging whether the data marking information meets the preset anchor point requirement;
The result output module is used for outputting a pre-verification success signal to the user terminal if the preset anchor point requirement is met;
the anchor point acquisition module includes: a row image acquisition sub-module, a text box acquisition sub-module, a column image acquisition sub-module, a text box group acquisition sub-module, and an anchor point determination sub-module, wherein:
The line image acquisition sub-module is used for carrying out line cutting operation on the original image information to acquire line image information;
the text box obtaining sub-module is used for carrying out identification operation on the line image information to obtain a text box;
The column image acquisition sub-module is used for performing column cutting operation on the original image information according to the text box to acquire a plurality of columns of column image information carrying the text box;
the text box group acquisition submodule is used for respectively selecting one text box from the column image information of every two adjacent columns to form a text box group;
The anchor point determining sub-module is used for taking the center point of the text box in the text box group as the anchor point data;
The text box group acquisition submodule comprises: the text box group determines a sub-module, wherein:
the text box group determining submodule is used for taking two text boxes which are farthest from each other in the vertical axis direction in the column image information of the two adjacent columns as the text box group.
7. A computer device comprising a memory and a processor, the memory having stored therein a computer program, the processor, when executing the computer program, performing the steps of the pre-verification method of an image annotation task according to any of claims 1 to 5.
8. A computer-readable storage medium, on which a computer program is stored which, when executed by a processor, implements the steps of the pre-verification method of an image annotation task according to any of claims 1 to 5.
CN202010106306.0A 2020-02-21 2020-02-21 Pre-verification method, device and equipment for image annotation task and storage medium Active CN111428749B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202010106306.0A CN111428749B (en) 2020-02-21 2020-02-21 Pre-verification method, device and equipment for image annotation task and storage medium
PCT/CN2020/117379 WO2021164251A1 (en) 2020-02-21 2020-09-24 Image annotation task pre-verification method and apparatus, device, and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010106306.0A CN111428749B (en) 2020-02-21 2020-02-21 Pre-verification method, device and equipment for image annotation task and storage medium

Publications (2)

Publication Number Publication Date
CN111428749A CN111428749A (en) 2020-07-17
CN111428749B true CN111428749B (en) 2024-07-02

Family

ID=71547037

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010106306.0A Active CN111428749B (en) 2020-02-21 2020-02-21 Pre-verification method, device and equipment for image annotation task and storage medium

Country Status (2)

Country Link
CN (1) CN111428749B (en)
WO (1) WO2021164251A1 (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111259980B (en) * 2020-02-10 2023-10-03 北京小马慧行科技有限公司 Method and device for processing annotation data
CN111428749B (en) * 2020-02-21 2024-07-02 平安科技(深圳)有限公司 Pre-verification method, device and equipment for image annotation task and storage medium
CN112766254B (en) * 2021-01-19 2024-02-02 中国工商银行股份有限公司 Automatic test method and device for inputting credential information
CN112836732B (en) * 2021-01-25 2024-04-19 深圳市声扬科技有限公司 Verification method and device for data annotation, electronic equipment and storage medium
CN115410165A (en) * 2021-05-27 2022-11-29 上海商汤临港智能科技有限公司 Image labeling method, image processing device and computer equipment
CN113888732A (en) * 2021-10-20 2022-01-04 江南造船(集团)有限责任公司 Ship installation information labeling method and system, electronic equipment and storage medium
CN115577379B (en) * 2022-11-09 2023-05-09 中孚安全技术有限公司 Hierarchical protection security analysis method, system and equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109190631A (en) * 2018-08-31 2019-01-11 阿里巴巴集团控股有限公司 The target object mask method and device of picture
CN110726971A (en) * 2019-10-22 2020-01-24 Oppo(重庆)智能科技有限公司 Visible light positioning method, device, terminal and storage medium

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9639740B2 (en) * 2007-12-31 2017-05-02 Applied Recognition Inc. Face detection and recognition
JP5664346B2 (en) * 2011-03-04 2015-02-04 富士ゼロックス株式会社 Image processing apparatus, information providing system, and program
CN108073914B (en) * 2018-01-10 2022-02-18 成都品果科技有限公司 Animal face key point marking method
CN108549893B (en) * 2018-04-04 2020-03-31 华中科技大学 An End-to-End Recognition Method for Scene Texts of Arbitrary Shapes
CN110245716B (en) * 2019-06-20 2021-05-14 杭州睿琪软件有限公司 Sample labeling auditing method and device
CN111428749B (en) * 2020-02-21 2024-07-02 平安科技(深圳)有限公司 Pre-verification method, device and equipment for image annotation task and storage medium

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109190631A (en) * 2018-08-31 2019-01-11 阿里巴巴集团控股有限公司 The target object mask method and device of picture
CN110726971A (en) * 2019-10-22 2020-01-24 Oppo(重庆)智能科技有限公司 Visible light positioning method, device, terminal and storage medium

Also Published As

Publication number Publication date
WO2021164251A1 (en) 2021-08-26
CN111428749A (en) 2020-07-17

Similar Documents

Publication Publication Date Title
CN111428749B (en) Pre-verification method, device and equipment for image annotation task and storage medium
US20210357710A1 (en) Text recognition method and device, and electronic device
CN103455264B (en) A kind of handwriting Chinese character input method and there is the electronic installation of the method
CN107256379A (en) Information collecting method, mobile terminal and storage medium based on image recognition
WO2013047980A1 (en) Method for inputting figures into an electronic document in a direct input manner and computer readable recording medium recording a program for inputting figures into an electronic document in direct input manner
CN113177133B (en) Image retrieval method, device, equipment and storage medium
CN111680480A (en) Template-based job correction method, device, computer equipment and storage medium
CN106570549A (en) Coding pattern generation and identification methods and coding pattern generation and identification devices
CN112395450B (en) Picture character detection method and device, computer equipment and storage medium
CN111709346B (en) Historical building identification and detection method based on deep learning and high-resolution images
CN110766938B (en) Road network topological structure construction method and device, computer equipment and storage medium
CN107122093B (en) Information frame display method and device
CN112637541A (en) Audio and video labeling method and device, computer equipment and storage medium
CN111368915A (en) Drawing verification method, device, equipment and storage medium
CN102592108B (en) Picture label information extracting method and device
CN115357155A (en) Window identification method, device, equipment and computer readable storage medium
CN112396048B (en) Picture information extraction method and device, computer equipment and storage medium
CN113139472B (en) Method, device, readable storage medium and electronic device for correcting calculation problems
CN112990142B (en) Video guide generation method, device and equipment based on OCR (optical character recognition), and storage medium
CN110263310B (en) Data graph generation method and device and computer readable storage medium
HK40031280A (en) Pre-verification method, device, equipment and storage medium for image annotation task
CN111783180A (en) Drawing splitting method and related device
CN113934875B (en) Electrophoresis data identification method and system, computer storage medium and electronic equipment
CN116704528A (en) Bill identification verification method, device, computer equipment and storage medium
CN114996542A (en) Data distribution-based horizontal histogram display method and related equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
REG Reference to a national code

Ref country code: HK

Ref legal event code: DE

Ref document number: 40031280

Country of ref document: HK

SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant