[go: up one dir, main page]

CN114612699B - A method and device for processing image data - Google Patents

A method and device for processing image data Download PDF

Info

Publication number
CN114612699B
CN114612699B CN202210235982.7A CN202210235982A CN114612699B CN 114612699 B CN114612699 B CN 114612699B CN 202210235982 A CN202210235982 A CN 202210235982A CN 114612699 B CN114612699 B CN 114612699B
Authority
CN
China
Prior art keywords
data
sample
result data
initial training
similarity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210235982.7A
Other languages
Chinese (zh)
Other versions
CN114612699A (en
Inventor
张建虎
王林芳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jingdong Technology Information Technology Co Ltd
Original Assignee
Jingdong Technology Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jingdong Technology Information Technology Co Ltd filed Critical Jingdong Technology Information Technology Co Ltd
Priority to CN202210235982.7A priority Critical patent/CN114612699B/en
Publication of CN114612699A publication Critical patent/CN114612699A/en
Application granted granted Critical
Publication of CN114612699B publication Critical patent/CN114612699B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种图像数据处理的方法和装置,涉及计算机技术领域。该方法的一具体实施方式包括:利用初始训练样本集训练得到的图像标注模型对待处理数据集进行标注,获得标注结果数据集;确定所述标注结果数据集与所述初始训练样本集之间的相似度;根据所述相似度更新所述初始训练样本集。该实施方式能够高效地更新、纠正初始训练样本集中的错标签,提高初始训练样本集的标注质量,进而提升训练得到的图像标注模型的性能。

The present invention discloses a method and device for image data processing, and relates to the field of computer technology. A specific implementation of the method includes: annotating a data set to be processed using an image annotation model trained with an initial training sample set to obtain an annotation result data set; determining the similarity between the annotation result data set and the initial training sample set; and updating the initial training sample set according to the similarity. This implementation can efficiently update and correct wrong labels in the initial training sample set, improve the annotation quality of the initial training sample set, and thus improve the performance of the trained image annotation model.

Description

Image data processing method and device
Technical Field
The present invention relates to the field of computer technologies, and in particular, to a method and an apparatus for processing image data.
Background
With the continuous development of artificial intelligence technology, the deep learning technology has more and more prominent effect in various industries, for example, the deep learning technology can be applied to the processing of image data. When training the deep learning model, a large amount of training data needs to be obtained, and the quality of labels in the training data can influence the performance of the model.
In the prior art, the scheme of weak supervision and semi-supervised learning is mainly adopted to improve the labeling speed and efficiency of data, but the labeling accuracy is lower due to the fact that the error labels of the original data are difficult to update and correct efficiently, and the labeling accuracy of a model obtained through training is lower.
Disclosure of Invention
In view of the above, the embodiment of the invention provides a method and a device for processing image data, which can efficiently update and correct the error label in the initial training sample set, improve the labeling quality of the updated initial training sample set, and further improve the performance of the image labeling model obtained by training.
To achieve the above object, according to one aspect of an embodiment of the present invention, there is provided a method of image data processing, including:
Labeling the data set to be processed by using an image labeling model obtained by training an initial training sample set to obtain a labeling result data set;
Determining the similarity between the labeling result data set and the initial training sample set;
and updating the initial training sample set according to the similarity.
Optionally, after obtaining the labeling result data set, correcting the labeling result data set is further included.
Optionally, determining the similarity between the labeling result dataset and the initial training sample set includes:
and for any one of the labeling result data in the labeling result data set, respectively calculating the similarity between the labeling result data and each piece of sample data in the initial training sample set.
Optionally, updating the initial training sample set according to the similarity includes:
For any one of the labeling result data, a sample data subset with similarity meeting a preset condition between the initial training sample set and the any one of the labeling result data is screened out;
and updating the initial training sample set according to the sample data subsets corresponding to all the labeling result data and the labeling result data set.
Optionally, before updating the initial training sample set according to the sample data subsets corresponding to all the labeling result data and the labeling result data sets, the method further comprises correcting the sample data meeting the preset condition and/or
And de-duplicating the sample data meeting the preset conditions.
Optionally, calculating the similarity between the labeling result data and each sample data in the initial training sample set includes:
determining sample characteristics corresponding to each sample data and labeling result characteristics corresponding to any labeling result data;
and determining the similarity between the labeling result data and each piece of sample data according to the distance between the labeling result feature and each sample feature.
Optionally, determining a sample feature corresponding to each sample data includes:
Respectively determining M sub-features corresponding to the sample data by using M preset models, wherein M is a positive integer greater than or equal to 1;
and splicing the M sub-features to obtain sample features corresponding to each piece of sample data, or taking the M sub-features as M sample features corresponding to each piece of sample data.
In another aspect of an embodiment of the present invention, there is provided an apparatus for processing image data, including:
The model labeling module is used for labeling the data set to be processed by using an image labeling model obtained by training the initial training sample set to obtain a labeling result data set;
The determining module is used for determining the similarity between the labeling result data set and the initial training sample set;
And the updating module is used for updating the initial training sample set according to the similarity.
According to still another aspect of an embodiment of the present invention, there is provided an electronic apparatus including:
One or more processors;
storage means for storing one or more programs,
The one or more programs, when executed by the one or more processors, cause the one or more processors to implement the methods of image data processing provided by the present invention.
According to still another aspect of an embodiment of the present invention, there is provided a computer-readable medium having stored thereon a computer program which, when executed by a processor, implements the method of image data processing provided by the present invention.
The image data processing method has the advantages that the image annotation model obtained by training the initial training sample set is utilized to annotate the data set to be processed to obtain the annotation result data set, the similarity between the annotation result data set and the initial training sample set is determined, the sample data subset with the similarity meeting the preset condition is screened or searched from the initial training sample set, and the initial training sample set is updated by utilizing the annotation result data set and the initial training sample set. The method provided by the embodiment of the invention can be used for efficiently updating and correcting the error labels in the initial training sample set, so that the labeling quality of the initial training sample set is improved, the performance of the image labeling model obtained by training is further improved, and the labeling accuracy of the image labeling model is improved.
Further effects of the above-described non-conventional alternatives are described below in connection with the embodiments.
Drawings
The drawings are included to provide a better understanding of the invention and are not to be construed as unduly limiting the invention. Wherein:
FIG. 1 is a schematic diagram of the main flow of a method of image data processing according to an embodiment of the present invention;
FIG. 2 is a schematic flow chart of another method of image data processing according to an embodiment of the present invention;
FIG. 3 is a schematic flow chart of a method of processing image data according to still another embodiment of the present invention;
FIG. 4 is a flow chart of a method of image data processing according to an embodiment of the present invention;
fig. 5 is a schematic diagram of main modules of an apparatus for image data processing according to an embodiment of the present invention;
FIG. 6 is an exemplary system architecture diagram in which embodiments of the present invention may be applied;
fig. 7 is a schematic diagram of a computer system suitable for use in implementing an embodiment of the invention.
Detailed Description
Exemplary embodiments of the present invention will now be described with reference to the accompanying drawings, in which various details of the embodiments of the present invention are included to facilitate understanding, and are to be considered merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the invention. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
Fig. 1 is a schematic diagram of main flow of a method for processing image data according to an embodiment of the present invention, as shown in fig. 1, the method for processing image data is applied to a server, and includes the following steps:
S101, marking a data set to be processed by using an image marking model obtained by training an initial training sample set to obtain a marking result data set;
Step S102, determining the similarity between the labeling result data set and the initial training sample set;
And step 103, updating the initial training sample set according to the similarity.
In an embodiment of the present invention, the initial training sample set includes a plurality of pieces of sample data, each piece of sample data including image data and labeling data (i.e., labels). The initial training sample set can be a purely manually marked image data set, and can also be a noise data set obtained through a semi-supervised or weakly supervised learning method. The error labels exist in the initial training sample set, the labeling quality of the error labels is low, and therefore the performance of an image labeling model obtained by training the initial training sample set containing the error labels is poor, and the labeling accuracy is low, so that the error labels in the initial training sample set need to be updated and corrected.
And training the neural network model by using the initial training sample set to obtain an image annotation model, wherein the neural network model can be a CNN (convolutional neural network) model of resnet50 0 (an image recognition model structure) or other neural network models, such as EFFICIENTNET, GOOGLENET and other neural network models.
The method comprises the steps of marking an unlabeled data set to be processed by using an image marking model to obtain a marking result data set, wherein each marking result data set comprises image data and model marking data obtained by marking the image marking model. The source of the data set to be processed is not particularly limited, and may be online data of some services, or data crawled from a third party or the internet.
In an optional implementation manner of the embodiment of the present invention, after obtaining the labeling result data set in step S101, correction is further included on the labeling result data set. Since the model annotation data in the annotation result data set may have an inaccurate condition, the erroneous or inaccurate model annotation data may be corrected by correction. The correction can be in a manual correction mode or an automatic correction mode so as to correct the wrong label in the model labeling data.
Optionally, when correcting the labeling result data set, in order to improve the correction efficiency and accuracy, the labeling result data set may be sampled and then corrected, for example, a part of labeling result data is extracted from the labeling result data set to correct, so that the similarity between the labeling result data set and the initial training sample set may be determined according to the part of labeling result data in the labeling result set. For example, 50 pieces of labeling result data are extracted from a labeling result data set containing 100 pieces of labeling result data for manual correction, the 50 pieces of labeling result data after manual correction are used as the labeling result data set, and the similarity between the 50 pieces of labeling result data and the initial training sample set is determined.
In the embodiment of the invention, in step S102, the similarity between the labeling result data set and the initial training sample set is determined, wherein the similarity between the labeling result data and each piece of sample data in the initial training sample set is calculated for any labeling result data in the labeling result data set.
And when the similarity between the labeling result data set and the initial training sample set is determined, calculating the similarity between each piece of sample data in the initial training sample set and the labeling result data according to any one of the labeling result data. For example, if the initial training sample set includes K pieces of sample data, K is an integer greater than or equal to 1, then, for any labeling result data, K similarity results may be calculated, and if the labeling result data set includes L pieces of labeling result data, then, a total of l×k similarity results may be calculated. The result of obtaining the l×k similarities by calculation may be used as the similarity between the labeling result set and the initial training sample set.
In an alternative implementation manner of the embodiment of the present invention, the similarity may be a distance, for example, may be at least one of a Cosine distance (Cosine distance, also referred to as Cosine similarity), a euclidean distance, and the like.
In an embodiment of the present invention, as shown in fig. 2, updating an initial training sample set according to the similarity includes:
Step S201, for any labeling result data, a sample data subset with similarity meeting a preset condition between the initial training sample set and any labeling result data is screened out;
Step S201, updating the initial training sample set according to the sample data subsets corresponding to all the labeling result data and the labeling result data sets.
In the embodiment of the invention, for any labeling result data, after calculating the similarity between any labeling result data and each sample data in the initial training sample set, judging whether the similarity meets a preset condition, and forming a sample data subset according to the sample data meeting the preset condition, wherein the similarity is the similarity between standard result data and sample data, different expression modes of the similarity are different, and the preset conditions are different. When the similarity is a distance, sample data with a distance smaller than a preset distance threshold (such as 0.1) from any one of the labeling result data is used as sample data meeting a preset condition, and the sample data meeting the preset condition can be that the distances between each piece of sample data and any one of the labeling result data are ordered according to the order from small to large for any one of the labeling result data, and the sample data corresponding to the distance N before the ordering is used as the sample data meeting the preset condition, wherein N is a positive integer greater than 1.
For each piece of labeling result data in the labeling result data set, one sample data subset is corresponding, and for the labeling result data set containing L pieces of labeling result data, L sample data subsets can be obtained, namely, sample data subsets corresponding to all the labeling result data are obtained. The initial training sample set may be updated according to the sample data subsets and the labeling result data sets corresponding to all the labeling result data, i.e. the updated initial training sample set contains the sample data subsets and the labeling result data sets corresponding to all the labeling result data. The initial training sample set may also be updated by selecting a portion of the sample data in the sample data subset corresponding to all of the annotation result data and all or a portion of the annotation result data.
In the embodiment of the invention, before updating the initial training sample set according to the sample data subsets corresponding to all the labeling result data and the labeling result data sets, the method further comprises correcting the sample data meeting the preset condition and/or
And de-duplicating the sample data meeting the preset condition.
In order to further improve the accuracy of labeling labels in the sample data subsets, before updating the initial training sample set according to the sample data subsets corresponding to all labeling result data and the labeling result data sets, correcting the sample data meeting preset conditions to obtain corrected sample data subsets, correcting error labels, updating the initial training sample set by using the corrected sample data subsets and the labeling result data sets, and improving the labeling quality of the updated initial training sample set, so that the performance of an image labeling model is improved.
In the embodiment of the invention, each piece of labeling result data corresponds to one sample data subset, the sample data subsets comprise sample data meeting the preset condition, all the labeling result data correspond to a plurality of sample data subsets, and the same data possibly exist in the plurality of sample data subsets, so that the same data in the sample data subsets corresponding to all the labeling result data can be de-duplicated to simplify the data.
In an alternative implementation manner of the embodiment of the present invention, as shown in fig. 3, calculating a similarity between the labeling result data and each sample data in the initial training sample set includes:
Step 301, determining sample characteristics corresponding to each piece of sample data and labeling result characteristics corresponding to any labeling result data;
Step S301, determining the similarity between the labeling result data and each piece of sample data according to the distance between the labeling result feature and each sample feature.
In the embodiment of the invention, when the similarity between the labeling result data and each sample data in the initial training sample set is calculated, the labeling result feature corresponding to the labeling result data and the sample feature corresponding to each sample data can be extracted, wherein the labeling result feature and the sample feature can be represented by adopting feature vectors, and the labeling result feature and the sample feature can be extracted by adopting an image labeling model or other ubiquity open source models.
Extracting the labeling result features from the labeling result data set to form a labeling result feature set, extracting the sample features from the initial training sample set to form a sample feature set, and taking the similarity between the labeling result feature set and the sample feature set as the similarity between the labeling result data and the initial training sample set. And aiming at any marking result feature, taking the calculated distance between any marking result feature and each sample feature as the similarity between any marking result data and each sample data, determining the sample feature with the distance meeting the preset condition from any marking result feature, mapping the sample feature meeting the preset condition back to the initial training sample set, and determining the corresponding sample data meeting the preset condition, namely screening a sample data subset meeting the preset condition from the initial training sample set.
In the embodiment of the invention, the distance between the labeling result feature and the sample feature is calculated, and the method can be realized by selecting an open source library faiss (an open source search library aiming at clustering and similarity).
In an embodiment of the present invention, determining a sample feature corresponding to each sample data includes:
respectively determining M sub-features corresponding to the sample data by using M preset models, wherein M is a positive integer greater than or equal to 1;
And splicing the M sub-features to obtain sample features corresponding to each piece of sample data, or taking the M sub-features as the M sample features corresponding to each piece of sample data.
In the embodiment of the invention, the M preset models can select an image annotation model obtained by training an initial training sample set, and can also select one or more of other generalization open source models, such as VGG, resnet, inceptionV and other convolution network models. The extraction of the labeling result features and the sample features can be performed by adopting one image labeling model or a plurality of preset models.
In the embodiment of the invention, when determining the sample characteristics corresponding to each piece of sample data, each preset model can extract one sample sub-characteristic corresponding to the piece of sample data for each piece of sample data, and M sample sub-characteristics can be obtained by respectively extracting M preset models.
In one implementation manner of the embodiment of the present invention, after determining M sample sub-features corresponding to the strip of sample data, each sample sub-feature is used as a sample feature corresponding to the strip of sample data, that is, M sample features may be obtained for one strip of sample data.
Similarly, when determining the labeling result features corresponding to any one piece of labeling result data, each preset model can extract one labeling result sub-feature for each piece of labeling result data, M preset models are adopted to extract M labeling result sub-features respectively, and each labeling result sub-feature is used as the labeling result feature corresponding to the piece of sample data, namely M labeling result features can be obtained for one piece of labeling result data.
In one implementation manner of the embodiment of the present invention, when determining sample features satisfying a preset condition, for any labeling result feature, each preset model may determine N sample features satisfying the preset condition, and for M preset models, m×n sample features satisfying the preset condition may be determined, so that sample data corresponding to m×n sample features may be used as sample data satisfying the preset condition, so that a sample data subset may be determined, and then the initial training sample set may be updated according to the sample data subset and the labeling result data set.
In another implementation manner of the embodiment of the present invention, after determining M sample sub-features corresponding to the piece of sample data, the M sample sub-features are spliced to form a sample feature, and the sample feature is used as the sample feature corresponding to the piece of sample data, for example, if m=3, 3 sample sub-features extracted from 3 preset models respectively for a certain piece of sample data are (a, b), (c, d) and (e, f), then the 3 sample sub-features may be spliced to form a sample feature (a, b, c, d, e, f), so as to obtain a sample data subset with higher similarity from the initial training sample set.
Similarly, M preset models are adopted to extract the marking result features, and each preset model can extract one marking result sub-feature for each piece of marking result data, so that M marking result sub-features can be obtained, and the M marking result sub-features are spliced into one marking result feature serving as the marking result feature corresponding to the piece of marking result data.
In another implementation manner of the embodiment of the present invention, when determining sample features satisfying a preset condition, N sample features satisfying the preset condition may be determined for one labeling result feature, so that sample data corresponding to the N sample features satisfying the preset condition may be used as sample data satisfying the preset condition, so that a sample data subset may be determined, and then the initial training sample set may be updated according to the sample data subset and the labeling result data set.
In the embodiment of the invention, the updated initial training sample set comprises the labeling result data set and the sample data subset, the image labeling model obtained by training the updated initial training sample set has better performance, and the labeling accuracy of the image labeling model is higher.
Fig. 4 shows a method for processing image data according to an embodiment of the present invention, which includes:
step S401, obtaining an initial training sample set;
step S402, training by using an initial training sample set to obtain an image annotation model;
step S403, acquiring a data set to be processed, and marking the data set to be processed by using an image marking model to acquire a marking result data set;
Step S404, manually correcting the labeling result data set after sampling to obtain a corrected labeling result data set;
Step S405, extracting the labeling result characteristics in the corrected labeling result data set by using the image labeling model, and extracting the sample characteristics in the initial training sample set;
Step S406, for any labeling result feature, calculating the cosine distance between each sample feature and any labeling result feature, and determining sample data corresponding to the sample features with the cosine distances smaller than a preset distance threshold as sample data meeting preset conditions;
step S407, constructing a sample data subset according to sample data which meet preset conditions and correspond to all the labeling result features;
Step S408, manually correcting the sample data subset to obtain a corrected sample data subset;
step S409, adding the corrected sample data subset and the corrected labeling result data set to obtain an updated initial training sample set, and returning to the step S401.
The image data processing method includes the steps of marking a data set to be processed by using an image marking model obtained through training of an initial training sample set to obtain a marking result data set, correcting the marking result data set to achieve correction of a wrong label in the marking result data set, determining similarity between the marking result data set and the initial training sample set, screening or searching sample data subsets with the similarity meeting preset conditions from the initial training sample set, and updating the initial training sample set by using the marking result data set and the initial training sample set. The method provided by the embodiment of the invention can be used for efficiently updating and correcting the error labels in the initial training sample set, so that the labeling quality of the initial training sample set is improved, the performance of the image labeling model obtained by training is further improved, and the labeling accuracy of the image labeling model is improved.
As shown in fig. 5, another aspect of an embodiment of the present invention provides an apparatus 500 for processing image data, including:
The model labeling module 501 labels the data set to be processed by using an image labeling model obtained by training an initial training sample set to obtain a labeling result data set;
a determining module 502, configured to determine a similarity between the labeling result dataset and the initial training sample set;
an updating module 503 updates the initial training sample set according to the similarity.
In the embodiment of the present invention, the model labeling module 501 is further configured to correct the labeling result dataset after obtaining the labeling result dataset.
In the embodiment of the present invention, the determining module 502 is further configured to calculate, for any labeling result data in the labeling result data set, a similarity between the labeling result data and each sample data in the initial training sample set, respectively.
In the embodiment of the invention, the updating module 503 is further configured to screen, for any labeling result data, a sample data subset with similarity satisfying a preset condition with any labeling result data from the initial training sample set, and update the initial training sample set according to the sample data subsets corresponding to all the labeling result data and the labeling result data sets.
In the embodiment of the present invention, the updating module 503 is further configured to correct the sample data that meets the preset condition before updating the initial training sample set according to the sample data subsets corresponding to all the labeling result data and the labeling result data sets, and/or
And de-duplicating the sample data meeting the preset condition.
In the embodiment of the present invention, the determining module 502 is further configured to determine a sample feature corresponding to each piece of sample data and a labeling result feature corresponding to any one of the labeling result data, and determine a similarity between the labeling result data and each piece of sample data according to a distance between the labeling result feature and each sample feature.
In the embodiment of the present invention, the determining module 502 is further configured to determine M sub-features corresponding to sample data by using M preset models, where M is a positive integer greater than or equal to 1, splice the M sub-features to obtain sample features corresponding to each piece of sample data, or use the M sub-features as M sample features corresponding to each piece of sample data.
In a further aspect, an embodiment of the present invention provides an electronic device, including one or more processors, and a storage device configured to store one or more programs, where the one or more programs are executed by the one or more processors, and cause the one or more processors to implement a method for processing image data according to an embodiment of the present invention.
The embodiment of the invention also provides a computer readable medium, on which a computer program is stored, which when executed by a processor implements the method of image data processing according to the embodiment of the invention.
Fig. 6 shows an exemplary system architecture 600 of a method of image data processing or an apparatus of image data processing to which embodiments of the invention may be applied.
As shown in fig. 6, the system architecture 600 may include terminal devices 601, 602, 603, a network 604, and a server 605. The network 604 is used as a medium to provide communication links between the terminal devices 601, 602, 603 and the server 605. The network 604 may include various connection types, such as wired, wireless communication links, or fiber optic cables, among others.
A user may interact with the server 605 via the network 604 using the terminal devices 601, 602, 603 to receive or send messages, etc. Various communication client applications such as shopping class applications, web browser applications, search class applications, instant messaging tools, mailbox clients, social platform software, etc. (by way of example only) may be installed on the terminal devices 601, 602, 603.
The terminal devices 601, 602, 603 may be various electronic devices having a display screen and supporting web browsing, including but not limited to smartphones, tablets, laptop and desktop computers, and the like.
The server 605 may be a server providing various services, such as a background management server (by way of example only) providing support for shopping-type websites browsed by users using terminal devices 601, 602, 603. The background management server may analyze and process the received data such as the product information query request, and feedback the processing result (e.g., the target push information, the product information—only an example) to the terminal device.
It should be noted that, the method for processing image data provided in the embodiment of the present invention is generally executed by the server 605, and accordingly, the device for processing image data is generally disposed in the server 605.
It should be understood that the number of terminal devices, networks and servers in fig. 6 is merely illustrative. There may be any number of terminal devices, networks, and servers, as desired for implementation.
Referring now to FIG. 7, there is illustrated a schematic diagram of a computer system 700 suitable for use in implementing an embodiment of the present invention. The terminal device shown in fig. 7 is only an example, and should not impose any limitation on the functions and the scope of use of the embodiment of the present invention.
As shown in fig. 7, the computer system 700 includes a Central Processing Unit (CPU) 701, which can perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 702 or a program loaded from a storage section 708 into a Random Access Memory (RAM) 703. In the RAM 703, various programs and data required for the operation of the system 700 are also stored. The CPU 701, ROM 702, and RAM 703 are connected to each other through a bus 704. An input/output (I/O) interface 705 is also connected to bus 704.
Connected to the I/O interface 705 are an input section 706 including a keyboard, a mouse, and the like, an output section 707 including a Cathode Ray Tube (CRT), a Liquid Crystal Display (LCD), and the like, a speaker, and the like, a storage section 708 including a hard disk, and the like, and a communication section 709 including a network interface card such as a LAN card, a modem, and the like. The communication section 709 performs communication processing via a network such as the internet. The drive 710 is also connected to the I/O interface 705 as needed. A removable medium 711 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is mounted on the drive 710 as necessary, so that a computer program read therefrom is mounted into the storage section 708 as necessary.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method shown in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via the communication portion 709, and/or installed from the removable medium 711. The above-described functions defined in the system of the present invention are performed when the computer program is executed by a Central Processing Unit (CPU) 701.
The computer readable medium shown in the present invention may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples of a computer-readable storage medium may include, but are not limited to, an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In the present invention, however, the computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams or flowchart illustration, and combinations of blocks in the block diagrams or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The modules involved in the embodiments of the present invention may be implemented in software or in hardware. The described modules may also be provided in a processor, for example, a processor may be described as including a model annotation module, a determination module, and an update module. The names of these modules do not constitute a limitation on the module itself in some cases, and for example, the update module may also be described as "a module that updates the initial training sample set according to similarity".
As a further aspect, the invention also provides a computer readable medium which may be comprised in the device described in the above embodiments or may be present alone without being fitted into the device. The computer readable medium carries one or more programs which, when executed by the apparatus, cause the apparatus to include labeling a data set to be processed using an image labeling model obtained by training an initial training sample set to obtain a labeling result data set, determining a similarity between the labeling result data set and the initial training sample set, and updating the initial training sample set according to the similarity.
According to the technical scheme, an image annotation model obtained by training an initial training sample set is utilized to annotate a data set to be processed to obtain an annotation result data set, the annotation result data set is corrected to correct a wrong label in the annotation result data set, similarity between the annotation result data set and the initial training sample set is determined, a sample data subset with the similarity meeting preset conditions is screened from the initial training sample set, and the initial training sample set is updated by utilizing the annotation result data set and the initial training sample set. The method provided by the embodiment of the invention can be used for efficiently updating and correcting the error labels in the initial training sample set, so that the labeling quality of the initial training sample set is improved, the performance of the image labeling model obtained by training is further improved, and the labeling accuracy of the image labeling model is improved.
The above embodiments do not limit the scope of the present invention. It will be apparent to those skilled in the art that various modifications, combinations, sub-combinations and alternatives can occur depending upon design requirements and other factors. Any modifications, equivalent substitutions and improvements made within the spirit and principles of the present invention should be included in the scope of the present invention.

Claims (7)

1.一种图像数据处理的方法,其特征在于,包括:1. A method for image data processing, comprising: 利用初始训练样本集训练得到的图像标注模型对待处理数据集进行标注,获得标注结果数据集;The image annotation model trained with the initial training sample set is used to annotate the dataset to be processed, and annotate the result dataset; 确定所述标注结果数据集与所述初始训练样本集之间的相似度;Determining the similarity between the labeled result data set and the initial training sample set; 根据所述相似度更新所述初始训练样本集;Updating the initial training sample set according to the similarity; 其中,获得标注结果数据集之后,还包括:对所述标注结果数据集进行校正;After obtaining the labeling result data set, the method further includes: correcting the labeling result data set; 确定所述标注结果数据集与所述初始训练样本集之间的相似度,包括:对于所述标注结果数据集中的任一标注结果数据,分别计算所述标注结果数据与所述初始训练样本集中每条样本数据之间的相似度;Determining the similarity between the labeled result data set and the initial training sample set includes: for any labeled result data in the labeled result data set, respectively calculating the similarity between the labeled result data and each sample data in the initial training sample set; 根据所述相似度更新所述初始训练样本集,包括:针对所述任一标注结果数据,从所述初始训练样本集中筛选出与所述任一标注结果数据之间的相似度满足预设条件的样本数据子集;根据所有标注结果数据对应的样本数据子集和所述标注结果数据集更新所述初始训练样本集;所述预设条件为相似度大于预设相似度阈值。Updating the initial training sample set according to the similarity includes: for any of the annotation result data, screening out from the initial training sample set a sample data subset whose similarity with any of the annotation result data meets a preset condition; updating the initial training sample set according to the sample data subset corresponding to all the annotation result data and the annotation result data set; the preset condition is that the similarity is greater than a preset similarity threshold. 2.根据权利要求1所述的方法,其特征在于,根据所有标注结果数据对应的样本数据子集和所述标注结果数据集更新所述初始训练样本集之前,还包括:对所述满足预设条件的样本数据进行校正;和/或2. The method according to claim 1 is characterized in that before updating the initial training sample set according to the sample data subset corresponding to all the labeled result data and the labeled result data set, it also includes: correcting the sample data that meets the preset conditions; and/or 对所述满足预设条件的样本数据进行去重。De-duplicate the sample data that meets the preset conditions. 3.根据权利要求1所述的方法,其特征在于,计算所述标注结果数据与所述初始训练样本集中每条样本数据之间的相似度,包括:3. The method according to claim 1, characterized in that calculating the similarity between the labeled result data and each sample data in the initial training sample set comprises: 确定每条所述样本数据对应的样本特征和所述任一标注结果数据对应的标注结果特征;Determine a sample feature corresponding to each piece of the sample data and a labeling result feature corresponding to any of the labeling result data; 根据所述标注结果特征与每个所述样本特征之间的距离确定所述标注结果数据与每条样本数据之间的相似度。The similarity between the annotation result data and each piece of sample data is determined according to the distance between the annotation result feature and each of the sample features. 4.根据权利要求3所述的方法,其特征在于,确定每条所述样本数据对应的样本特征,包括:4. The method according to claim 3, characterized in that determining the sample feature corresponding to each piece of the sample data comprises: 利用M个预设模型分别确定与所述样本数据对应的M个子特征,所述M为大于等于1的正整数;Using M preset models, respectively determine M sub-features corresponding to the sample data, where M is a positive integer greater than or equal to 1; 拼接所述M个所述子特征,获得与所述每条样本数据对应的样本特征;或,将所述M个子特征作为与所述每条样本数据对应的M个样本特征。The M sub-features are concatenated to obtain a sample feature corresponding to each piece of sample data; or the M sub-features are used as M sample features corresponding to each piece of sample data. 5.一种图像数据处理的装置,其特征在于,包括:5. A device for processing image data, comprising: 模型标注模块,利用初始训练样本集训练得到的图像标注模型对待处理数据集进行标注,获得标注结果数据集;The model annotation module uses the image annotation model trained with the initial training sample set to annotate the dataset to be processed and obtain an annotation result dataset; 确定模块,确定所述标注结果数据集与所述初始训练样本集之间的相似度;A determination module is used to determine the similarity between the annotation result data set and the initial training sample set; 更新模块,根据所述相似度更新所述初始训练样本集;An updating module, which updates the initial training sample set according to the similarity; 其中,所述模型标注模块,还用于:获得标注结果数据集之后,对所述标注结果数据集进行校正;Wherein, the model annotation module is further used to: after obtaining the annotation result data set, calibrate the annotation result data set; 确定模块,进一步用于:对于所述标注结果数据集中的任一标注结果数据,分别计算所述标注结果数据与所述初始训练样本集中每条样本数据之间的相似度;The determination module is further used to: for any labeled result data in the labeled result data set, respectively calculate the similarity between the labeled result data and each sample data in the initial training sample set; 更新模块,进一步用于:针对所述任一标注结果数据,从所述初始训练样本集中筛选出与所述任一标注结果数据之间的相似度满足预设条件的样本数据子集;根据所有标注结果数据对应的样本数据子集和所述标注结果数据集更新所述初始训练样本集;所述预设条件为相似度大于预设相似度阈值。The updating module is further used to: for any of the annotation result data, screen out a sample data subset whose similarity with any of the annotation result data meets a preset condition from the initial training sample set; update the initial training sample set according to the sample data subset corresponding to all the annotation result data and the annotation result data set; the preset condition is that the similarity is greater than a preset similarity threshold. 6.一种电子设备,其特征在于,包括:6. An electronic device, comprising: 一个或多个处理器;one or more processors; 存储装置,用于存储一个或多个程序,a storage device for storing one or more programs, 当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如权利要求1-4中任一所述的方法。When the one or more programs are executed by the one or more processors, the one or more processors implement the method according to any one of claims 1 to 4. 7.一种计算机可读介质,其上存储有计算机程序,其特征在于,所述程序被处理器执行时实现如权利要求1-4中任一所述的方法。7. A computer-readable medium having a computer program stored thereon, wherein when the program is executed by a processor, the method according to any one of claims 1 to 4 is implemented.
CN202210235982.7A 2022-03-10 2022-03-10 A method and device for processing image data Active CN114612699B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210235982.7A CN114612699B (en) 2022-03-10 2022-03-10 A method and device for processing image data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210235982.7A CN114612699B (en) 2022-03-10 2022-03-10 A method and device for processing image data

Publications (2)

Publication Number Publication Date
CN114612699A CN114612699A (en) 2022-06-10
CN114612699B true CN114612699B (en) 2025-02-28

Family

ID=81863241

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210235982.7A Active CN114612699B (en) 2022-03-10 2022-03-10 A method and device for processing image data

Country Status (1)

Country Link
CN (1) CN114612699B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115050002A (en) * 2022-07-11 2022-09-13 零束科技有限公司 Image annotation model training method and device, electronic equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108009589A (en) * 2017-12-12 2018-05-08 腾讯科技(深圳)有限公司 Sample data processing method, device and computer-readable recording medium
CN110689038A (en) * 2019-06-25 2020-01-14 深圳市腾讯计算机系统有限公司 Training method, device and medical image processing system for neural network model
CN111104479A (en) * 2019-11-13 2020-05-05 中国建设银行股份有限公司 Data labeling method and device

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10719301B1 (en) * 2018-10-26 2020-07-21 Amazon Technologies, Inc. Development environment for machine learning media models
CN109492684A (en) * 2018-10-31 2019-03-19 西安同瑞恒达电子科技有限公司 Data processing method and device
CN110580290B (en) * 2019-09-12 2022-12-13 北京小米智能科技有限公司 Method and device for optimizing training set for text classification
CN111723870B (en) * 2020-06-22 2024-04-09 中国平安人寿保险股份有限公司 Artificial intelligence-based data set acquisition method, apparatus, device and medium
CN112036491A (en) * 2020-09-01 2020-12-04 北京推想科技有限公司 Method and device for determining training sample and method for training deep learning model
CN112560993B (en) * 2020-12-25 2024-07-23 北京百度网讯科技有限公司 Data screening method and device, electronic equipment and storage medium
CN113064993B (en) * 2021-03-23 2023-07-21 南京视察者智能科技有限公司 Design method, optimization method and labeling method of automatic text classification labeling system based on big data
CN113095444B (en) * 2021-06-07 2021-09-17 北京智芯微电子科技有限公司 Image labeling method, device and storage medium
CN113780367A (en) * 2021-08-19 2021-12-10 北京三快在线科技有限公司 Classification model training and data classification method and device, and electronic equipment
CN113673607B (en) * 2021-08-24 2025-05-13 支付宝(杭州)信息技术有限公司 Image annotation model training and image annotation method and device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108009589A (en) * 2017-12-12 2018-05-08 腾讯科技(深圳)有限公司 Sample data processing method, device and computer-readable recording medium
CN110689038A (en) * 2019-06-25 2020-01-14 深圳市腾讯计算机系统有限公司 Training method, device and medical image processing system for neural network model
CN111104479A (en) * 2019-11-13 2020-05-05 中国建设银行股份有限公司 Data labeling method and device

Also Published As

Publication number Publication date
CN114612699A (en) 2022-06-10

Similar Documents

Publication Publication Date Title
JP7301922B2 (en) Semantic retrieval method, device, electronic device, storage medium and computer program
US11062089B2 (en) Method and apparatus for generating information
CN108628830B (en) Semantic recognition method and device
JP2021197154A (en) Form image recognition method and device, electronic apparatus, storage medium, and computer program
US20180365257A1 (en) Method and apparatu for querying
CN108830329B (en) Picture processing method and device
CN111274815A (en) Method and device for mining entity attention points in text
CN111291765A (en) Method and device for determining similar pictures
CN111340220B (en) Method and apparatus for training predictive models
CN111104479A (en) Data labeling method and device
CN114792089B (en) Method, apparatus and program product for managing a computer system
CN109858045B (en) Machine translation method and device
CN112712795A (en) Method, device, medium and electronic equipment for determining label data
CN110705271B (en) System and method for providing natural language processing service
WO2022148239A1 (en) Method and apparatus for information output, and electronic device
CN113408507B (en) Named Entity Recognition Method, Device and Electronic Device Based on History File
CN109635184B (en) Financial product recommendation method, device and computer equipment based on data analysis
CN111008213A (en) Method and apparatus for generating language conversion model
CN114612699B (en) A method and device for processing image data
CN112528103B (en) Method and device for recommending an object
CN111078849A (en) Method and apparatus for outputting information
CN113239259A (en) Method and device for determining similar stores
CN110807097A (en) Method and device for analyzing data
CN113076254A (en) Test case set generation method and device
CN113704442B (en) A method and device for processing consultation request

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant