US11120337B2 - Self-training method and system for semi-supervised learning with generative adversarial networks - Google Patents
Self-training method and system for semi-supervised learning with generative adversarial networks Download PDFInfo
- Publication number
- US11120337B2 US11120337B2 US15/789,628 US201715789628A US11120337B2 US 11120337 B2 US11120337 B2 US 11120337B2 US 201715789628 A US201715789628 A US 201715789628A US 11120337 B2 US11120337 B2 US 11120337B2
- Authority
- US
- United States
- Prior art keywords
- sample
- samples
- training dataset
- generated
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/088—Non-supervised learning, e.g. competitive learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G06N3/0454—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/086—Learning methods using evolutionary algorithms, e.g. genetic algorithms or genetic programming
Definitions
- the present disclosure relates to generative adversarial networks, and in particular to semi-supervised learning in generative adversarial networks.
- GANs generative adversarial networks
- a GAN includes a generator to produce data and a discriminator to classify data, engaged in a two-player minimax game to improve the performance of each other.
- the GAN may, for example, employ a least square as a loss function.
- the original GAN has been trained for unconditional semi-supervised learning and with triple GANs, for conditionally semi-supervised learning.
- GANs can provide powerful capacity for generating realistic synthetic images. Much of the current research in this domain is focused on improving the visual quality of generated images and the training stability of GANs. However, the issue of how to further exploit the visual characteristics of these generated images, in particular in the context of semi-supervised GAN, is still unclear.
- Self-training has previously been studied in the machine learning domain, including for example the Yarowsky algorithm which has been applied to word sense disambiguation.
- traditional machine learning typically uses labelled and unlabelled data and lacks appropriate mechanisms for generating and then using synthetic data to improve classification performance.
- Example aspects of the present disclosure provide a method and system for self-training using a GAN that uses unlabelled and generated datasets for estimating data distribution and classification.
- the methods and system described herein introduce self-training capacity to GANs that are enabled for self-supervised training.
- the methods and systems described herein provide a mechanism to automatically augment a training dataset by expanding the size of a training dataset stored in a databank to include additional labelled samples, there facilitating evolutionary learning.
- the self-learning GAN methods and systems described in this document may offer one or more benefits over existing methods and systems in at least some applications.
- at least some embodiments described herein provide a method of using self-training in an adversarial network for both data generation and label classification. This is in contrast to traditional methods that use self-training only for label classification.
- embodiments described herein use not only unlabelled data, but also generated unlabelled data, thereby substantially increasing the volume of data for training the discriminator.
- adversarial training is combined with self-training, and thus the overall system becomes evolutionary.
- the training data set includes labelled samples and unlabelled samples.
- the method includes: receiving generated samples generated using a first neural network of the GAN and the unlabelled samples of training dataset; determining a decision value for a sample from a decision function, wherein the sample is a generated sample of the generated samples or an unlabelled sample of the unlabelled samples of the training dataset; comparing the decision value to a threshold; in response to determining that the decision value exceeds the threshold: predicting a label for a sample; assigning the label to the sample; and augmenting the training dataset to include the sample with the assigned label as a labelled sample.
- determining a decision value for the sample, comparing the decision value to a threshold, predicting the label, assigning the label, and augmenting the training dataset are repeated for a plurality of the generated samples and unlabelled samples.
- the method includes updating unlabelled samples to remove any unlabelled samples to which labels have been assigned, and augmenting the generated samples to remove any generated samples to which labels have been assigned.
- the method includes, prior to predicting the label, training the GAN using the training dataset. In some aspects, the method includes, after augmenting the training dataset, training the GAN using the augmented training dataset. In some aspects, the method includes receiving new generated samples generated using the first neural network of the GAN and unlabelled samples of the augment training dataset, and determining a decision value for the sample, comparing the decision value the threshold, predicting the label for the sample, assigning the label, and augmenting the training dataset are repeated for a plurality of the new generated samples and the unlabelled samples of the augment training dataset.
- training of the GAN, determining a decision value for the sample, predicting the label, assigning the label, and augmenting the training dataset are repeated until a validation error for the GAN stops decreasing in respect of the augmented training data set.
- the decision function depends on information about the sample.
- the decision value generated from the decision function is a posterior probability for the sample generated by the second neural network.
- a system for augmenting a training dataset for a generative adversarial network the training dataset including labelled samples and unlabelled samples.
- the system includes a processing device and a memory coupled to the processing device.
- the memory stores computer-executable instructions that, when executed by the processing device, cause the system to: receive generated samples generated using a first neural network of the GAN and the unlabelled samples of training dataset; determine a decision value for a sample from a decision function, wherein the sample is a generated sample of the generated samples or an unlabelled sample of the unlabelled samples of the training dataset; compare the decision value to a threshold; responsive to determining that the decision value exceeds the threshold: predict a label for a sample; assign the label to the sample; and augment the training dataset to include the sample with the assigned label as a labelled sample.
- GAN generative adversarial network
- the computer program product for augmenting a training dataset for a generative adversarial network (GAN).
- the training dataset comprising labelled samples and unlabelled samples.
- the computer program product includes a computer readable medium storing program code, wherein the program code, when run on a computer, causes the computer to: receive generated samples generated using a first neural network of the GAN and the unlabelled samples of training dataset; determine a decision value for a sample from a decision function, wherein the sample is a generated sample of the generated samples or an unlabelled sample of the unlabelled samples of the training dataset; compare the decision value to a threshold; responsive to determining that the decision value exceeds the threshold: predict a label for a sample; assign the label to the sample; and augment the training dataset to include the sample with the assigned label as a labelled sample.
- FIG. 1 is a block diagram of a self-training, semi-supervised learning system that incorporates a Generative Adversarial Network (GAN) according to a first example embodiment;
- GAN Generative Adversarial Network
- FIG. 2 is a flow diagram of a self-training, semi-supervised learning method that can be implemented using the system of FIG. 1 , according to example embodiments;
- FIG. 3 is a flow diagram of initialization and training phases of the method of FIG. 2 according to an example embodiment
- FIG. 4 is a flow diagram of an augmentation phase of the method of FIG. 2 according to an example embodiment
- FIG. 5 is block diagram of a discriminator of the GAN of FIG. 1 according to an example embodiment
- FIG. 6 is block diagram of a generator of the GAN of FIG. 1 , according to an example embodiment.
- FIG. 7 is a block diagram illustrating an example processing system suitable for implementing examples disclosed herein.
- GANs generative adversarial networks
- the generator takes in a random variable, z, with a distribution P z (z) and attempts to map the random variable z to provide a realistic generated sample within a data distribution P data (x).
- the discriminator is expected to discern real samples from generated samples by giving the output of 1 or 0 respectively.
- the generator and discriminator are used to generate samples and classify them respectively to improve the performance of each other in an adversarial manner.
- a GAN implements a two-player minimax game with the objective of deriving a Nash-equilibrium point.
- the following adversarial loss function has previously been employed in training a generator and discriminator: min G max D ⁇ E x ⁇ P data (x) [log D ( x )]+ E z ⁇ p z (z) [log(1 ⁇ D ( G ( z ))] ⁇ . Equation (1)
- a first neural network (generally referred to hereinafter as a generator) is used to generate synthetic data (referred to herein as generated samples).
- a second neural network (generally referred to hereinafter as a discriminator) is configured to receive as inputs a set of training data (referred to hereinafter as a training dataset).
- the training dataset includes a set of labelled training data (referred to hereinafter as a labelled training dataset) comprising labelled training data (referred to hereinafter as labelled samples), a larger set of unlabelled training data (referred to herein as an unlabelled training dataset) comprising unlabelled training data (referred to hereinafter as unlabelled samples), and the generated samples.
- the unlabelled training dataset includes at least 10 times as many samples as the labelled training dataset.
- the discriminator is configured to discriminate the generated samples from the training samples (e.g. the labelled samples and the unlabelled samples) and also predict labels for the unlabelled samples and the generated samples. Based on the outputs from the discriminator, a data augmentation unit is used to compare the posterior probability of a label for each unlabelled sample and each generated sample.
- the data augmentation unit can be implemented in the GAN or can be a separate module coupled to the GAN. When the posterior probability for the label for a given sample (e.g. the unlabelled sample or the generated sample) exceeds a threshold confidence level, the given sample is assigned the label and converted to a labelled training sample.
- the newly labelled sample is merged into the labeled training dataset, thereby augmenting the labelled training dataset and expanding the size of the labelled training dataset.
- the growing labeled training data set is used with newly generated data samples to further train the semi-supervised GAN. Augmenting the training dataset by adding labelled samples to the training dataset using the method and system described herein improves the performance of the GAN.
- FIG. 1 shows a schematic architecture of a semi-supervised, self-training system 90 (referred to hereinafter as self-training system 90 ) according to example embodiments.
- the self-training system 90 includes a GAN 100 , a data augmentation unit 106 , and a data bank 108 .
- the GAN 100 includes two feedforward neural networks, namely a generator feedforward neural network G(z) 102 (referred to hereinafter as generator G(z) 102 ) and a classification/discriminator feedforward neural network D 104 (referred to hereinafter as discriminator D 104 ).
- data augmentation unit 106 of the self-training system 90 can be selectively connected to the data bank 108 to add newly labelled samples to the training dataset stored in the data bank 108 to augment the training dataset as described in further detail below. Further, the data augmentation unit 106 can be selectively disconnected from the data bank 108 to train the GAN 100 using the augmented training dataset as described in further detail below.
- the generator G(z) 102 is configured to map a random noise vector z that has been drawn from a uniform or normal noise distribution p z (z) to produce generated samples x gen that simulate real samples.
- the generated data samples x gen are added to a dataset X gen of generated samples that are stored in the data bank 108 .
- Data bank 108 also includes a training dataset X train .
- Training dataset X train includes a labelled dataset X lab that includes labelled training samples x lab and an unlabelled dataset X unl that includes unlabelled training samples x unl .
- the discriminator D 104 receives generated samples x gen , along with labelled training samples x lab and unlabelled training data samples x unl , and tries to discriminate the generated samples x gen as fake from the labelled and unlabelled training data samples x lab , x unl .
- the discriminator D 104 is also configured to perform a classification function to determine probabilities for the different class labels y i to y k that can be applied to an unlabelled sample (which can be a generated sample X gen or an unlabelled training sample x unl ).
- discriminator D 104 is also configured to distinguish between K possible label classes. Each of the i th component of the K-dimensional output of the discriminator D(x) 104 in FIG.
- Discriminator D 104 is also configured to generate a posterior probability value P(y i
- FIG. 2 illustrates an overview of a self-training, semi-supervised learning method 200 that can be implemented using system 90 according to example embodiments.
- Method 200 includes an initialization phase 201 , followed by a training phase 208 and a data augmentation phase 223 .
- the method 200 is implemented on a processing system 600 (described in greater detail below) using the Python programming language and the Python libraries Theano and Lasagne.
- the Adam optimization algorithm is used to iteratively update the discriminator D 104 and generator G 102 networks during the training phase 208 .
- other programming languages, libraries and optimization algorithms may be used.
- the data augmentation phase 223 is carried out by data augmentation unit 106 in combination with the generator G 102 and discriminator D 104 . As illustrated in FIG. 2 , the training phase 208 and data augmentation phase 223 are alternatively and repetitively performed on an evolving training dataset until a validation error for a validation dataset stops decreasing (block 222 ).
- FIG. 3 illustrates the initialization phase 201 and training phase 208 of FIG. 2 in detail.
- the initialization phase 201 commences with defining and initializing the feedforward neural networks used to implement the discriminator D 104 and the generator G 102 (block 202 ).
- FIGS. 5 and 6 Non-limiting examples of possible architectures for the feedforward neural networks used to implement the discriminator D 104 and the generator G 102 are illustrated in FIGS. 5 and 6 respectively.
- the discriminator D 104 contains an input layer 302 followed by five successive dense layers 304 ( 1 )- 304 ( 5 ).
- the generator G 102 includes two successive dense layers 404 ( 1 )- 404 ( 2 ).
- Initial generator weighting parameters ⁇ and discriminator weighing parameters ⁇ are set as part of the initialization phase 201 . It will be appreciated although FIG. 5 and FIG. 6 illustrate feedforward neural networks used to implement the discriminator D 104 and the generator G 102 , any suitable neural network architecture may be used to implement the discriminator D 104 and the generator G 102 .
- data augmentation unit 106 is also defined and initialized as part of the initialization phase 201 .
- data augmentation unit 106 is initialized and trained to perform the thresholding operations discussed in greater detail below.
- the loss functions loss D and loss G for the discriminator D 104 and generator G 102 are defined as part of the initialization phase 201 .
- E x ⁇ plab(x) [ ⁇ D(x) ⁇ y ⁇ 2 ] is the supervised loss, in which:
- loss D E x ⁇ plab(x) [ ⁇ D ( x ) ⁇ y ⁇ 2 ] 2 +E z ⁇ pz(z),x ⁇ punl(x) [ D ( G ( z )) ⁇ D ( x )] 2 Equation (3-A)
- loss G E z ⁇ pz(z),x ⁇ punl(x) [( D ( G ( z )) ⁇ D ( x )) 2 ] 2 Equation (4-A)
- the training phase 208 starts, as shown at block 206 in FIG. 3 .
- the training phase 208 begins with dataset X gen stored in data bank 108 being initiated to zero so that the training phase 208 starts with an empty generated dataset X gen .
- the training phase also begins with a current training dataset X train(j) , where j denotes an iteration number for the training dataset.
- the training dataset X train(j) is pre-populated with unlabelled training dataset X unl and labelled training dataset X lab .
- the unlabelled training dataset X unl includes at least 10 times as many samples as the labelled training dataset X lab .
- the training phase 208 is an iterative phase during which the generator G 102 and discriminator D 104 are trained using the current training dataset X train(j) until a validation error for a validation dataset stops decreasing, as shown by blocks 210 to 221 .
- an adversarial game is played for improving the discrimination and classification performance by the discriminator D 104 and the data generation performance by generator G 102 simultaneously.
- each iteration of the training phase 208 begins with the establishment of a subset or batch of training data that includes: a batch of labelled samples (x lab , y lab ) that is a subset of the labelled training dataset X lab ; a batch of unlabelled samples (x unl , y unl ) that is a subset of the unlabelled training dataset X unl , and a batch of generated samples (x gen ) that are generated by generator G 102 .
- the generated samples (x gen ) are merged into the generated dataset X gen for the current training phase 208 .
- the discriminator D 104 's loss is determined based on its defined loss function loss D and the subset of samples x lab , y lab , x unl and x gen and the resulting error backpropagated to adjust the weighting parameters ⁇ of the feedforward neural networks used to implement discriminator D 104 .
- the generator G 102 's loss is determined based on its defined loss function loss G and the subset of samples x unl and x gen , and the error backpropagated to adjust the weighting parameters ⁇ of the feedforward neural networks used to implement discriminator G 102 .
- the generator G 102 is trained by applying a feature matching loss that uses an intermediate layer output of discriminator D 104 for the unlabelled training samples (X unl ) and generated samples (X gen ) respectively.
- a feature matching loss that uses an intermediate layer output of discriminator D 104 for the unlabelled training samples (X unl ) and generated samples (X gen ) respectively.
- an Adam optimization is used to back propagate the discriminator weighting parameters ⁇ , followed by another Adam optimization to back propagate the generator weighting parameters ⁇ .
- the validation dataset is a predetermined dataset that is used to determine when training of the discriminator D 104 has reached a level where the validation error reaches its minimal level. If the validation error is still decreasing then the GAN 100 has still not been optimally trained using the current training dataset X train(j) . Thus, if the validation error has not yet stopped decreasing, the training phase 208 enters another iteration using the same training dataset X train(j) with an additional set of generated samples x gen , and the actions described above in respect of blocks 210 to 221 are repeated. As indicated by the dashed line in FIG. 3 , in some example embodiments, blocks 210 to 220 are repetitively performed for sets of batches of samples (for example m sets of batches of n samples) prior to the validation testing of block 221 .
- the validation error determined in respect of the current training dataset X train(j) is compared against the validation error determined in respect of one or more previous training datasets X train(j ⁇ 1) to determine if the validation error is still declining as the training dataset evolves. In the event that the validation error is no longer decreasing, the self-training/semi-supervised method 200 is concluded.
- data augmentation phase 223 is performed to further evolve the training dataset X train(j) to produce a new, augmented training set X train(j+1) for the next iteration of the training phase 208 .
- Data augmentation phase 223 that is illustrated in FIG. 4 .
- the actions represented by blocks 224 to 234 of the data augmentation phase 223 are reiterated for each sample x that belongs to the current unlabelled training dataset X unl and the generated dataset X gen .
- data augmentation phase 223 begins with the data augmentation unit 106 receiving the generated samples of the generated dataset X gen , current unlabelled training dataset x unl from the data bank 108 , and a posterior probability P(y i
- x) is an output vector that corresponds to the Softmax of the output of the neural network of the discriminator D 104 , which is determined using the following equation: P ( y i
- K is equal to the number of classes (i.e. the number different possible labels that can be assigned to a sample), and i is the index of the class.
- a value for a decision function T (f, x) for the sample x is determined and compared against a threshold.
- the decision function T (f, x) returns a value (referred to hereinafter as a decision value) that indicates whether the sample x should be added to the labelled dataset of the training dataset based on information about the sample x.
- the information about the sample is denoted by fin the decision function T (f, x).
- the information about the sample x may be an output of the neural network used to implement the discriminator 104 , the posterior probability P(y i
- a few subsets of unlabelled or generated samples may be labelled and new GANs may be trained using each of these subsets to see which subset of unlabelled or generated samples gives the best GAN.
- the validation score that the GAN obtains while being trained with the sample x (or a set where sample x belongs) is considered information f since the validation score can be used to know whether the sample should be assigned a label and added to the labelled dataset as described in further detail below.
- the decision value generated from the decision function T (f, x) is the posterior probability P(y i
- the decision function generates a decision value of 1 if the sample is to be assigned a label and added to the training dataset, and 0 otherwise.
- the threshold has a value of 0.
- a sample x can be added to the labelled training set if T (f, x)>1 (e.g. the threshold has a value of 1 and 1 is the previous validation score (e.g. the validation score when the GAN 100 is not trained on a subset containing the sample x).
- the sample x is assigned a label that corresponds to the possible label y i having the highest posterior probability P(y i
- the training dataset (X train(j+1) ) is updated to add the newly labelled sample (x,y*) to the labelled training dataset X lab stored in data bank 108 , as indicated by: X lab ⁇ X lab U ⁇ (x,y*) ⁇ .
- the newly labelled sample (x,y*) can in example embodiments be removed from the generated dataset X gen if the sample was originally a generated sample x gen or removed from the unlabelled training dataset X unl if the sample was originally an unlabelled training sample x unl .
- the samples x is ignored and not assigned a label.
- the actions indicated in blocks 224 to 234 of the data augmentation phase 223 are repeated for each of the samples x, or each subset of the samples x, in the unlabelled training dataset X unl and the generated dataset X gen .
- the newly augmented training dataset X train(j+1) is then set as the current training dataset and the training phase 208 is repeated.
- the self-training/semi-supervised method 200 illustrated in FIGS. 2 to 4 includes training phase 208 for training discriminator D 104 and generator G 102 using a current training dataset.
- the training phase 208 is a data augmentation phase 223 that iterates through all samples x or all subsets of the samples x of the unlabelled training dataset x unl and generated X gen dataset to determine if a decision function T (f, x) for the sample x, exceeds a threshold. If so, the predicted label y* is assigned to the sample x and the now labelled sample (x,y*) is added to the labelled training dataset X lab .
- the sample x is removed from either the unlabelled training dataset X unl or generated dataset X gen from which the sample originated.
- the dataset X gen stored in data bank 108 is reinitialized as indicated at step 206 for every new training phase, and the new augmented training dataset that includes labelled and unlabelled training datasets X lab and X unl is then used to retrain the discriminator D 104 and generator G 102 .
- the dataset X gen stored in data bank 108 is not reinitialized for every new training phase.
- new generated samples generated by the generator G 102 for every new training phase are added to the dataset X gen stored in the data bank 108 .
- the training and augmentation phases 208 , 223 continue repeating until the validation error for the GAN 100 stops decreasing between evolutions of the augmented training data set X train .
- a self-learning mechanism is used to grow the labelled training dataset X lab of data bank 108 for semi-supervised training of GAN 100 .
- a decision function T (f, x) is used.
- Using a decision function T (f, x) can enable the method 200 to be used with a wide variety of different decision schemes, which includes hard and soft thresholding. For example, a desired distribution model can be selected as the threshold.
- the described embodiments introduce a self-training capacity to semi-supervised training with GAN and provides a mechanism to automatically expand the size of the training dataset stored in the data bank 108 , thus being able to make learning evolutionary in at least some applications.
- system 90 once the system 90 has been trained it can be used to generate training datasets of labelled data that can then be used to train other artificial intelligence (AI) systems.
- AI artificial intelligence
- samples can include front camera images and associated labelling can include a steering angle applied to each of the images.
- object detection the samples may include a camera image and the associated label can include one or more class labels that identify objects included in the image.
- the samples could include ultrasound images and the associated labels could include class labels that identify if an anomaly is present in the ultrasound images.
- consumer electronics the samples could include measured RF activity for one or more channels and the associated labels could include an RF channel suitability indicator.
- the configurations of the neural networks used to implement the GAN architecture described above are not critical, but instead used just as one example. Many revisions to the layers of the neural networks, such as weights, activation functions and normalization methods, leading to a different neural net would be possible.
- FIG. 7 is a block diagram of an example simplified processing system 600 , which may be used to implement embodiments disclosed herein, and provides a higher level implementation example.
- the method 200 may be implemented using the example processing system 600 , or variations of the processing system 600 .
- the processing system 600 could be a server or a desktop terminal, for example, or any suitable processing system.
- Other processing systems suitable for implementing embodiments described in the present disclosure may be used, which may include components different from those discussed below.
- FIG. 7 shows a single instance of each component, there may be multiple instances of each component in the processing system 600 .
- the processing system 600 may include one or more processing devices 602 , such as a graphics processing unit, a processor, a microprocessor, an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), a dedicated logic circuitry, or combinations thereof.
- the processing system 600 may also include one or more input/output (I/O) interfaces 604 , which may enable interfacing with one or more appropriate input devices 614 and/or output devices 616 .
- the processing system 600 may include one or more network interfaces 606 for wired or wireless communication with a network (e.g., an intranet, the Internet, a P2P network, a WAN and/or a LAN) or other node.
- the network interfaces 606 may include wired links (e.g., Ethernet cable) and/or wireless links (e.g., one or more antennas) for intra-network and/or inter-network communications.
- the processing system 600 may also include one or more storage units 608 , which may include a mass storage unit such as a solid state drive, a hard disk drive, a magnetic disk drive and/or an optical disk drive.
- the processing system 600 may include one or more memories 610 , which may include a volatile or non-volatile memory (e.g., a flash memory, a random access memory (RAM), and/or a read-only memory (ROM)).
- the non-transitory memory(ies) 610 may store instructions for execution by the processing device(s) 602 , such as to carry out examples described in the present disclosure, for example to perform encoding or decoding.
- the memory(ies) 610 may include other software instructions, such as for implementing an operating system for the processing system 600 and other applications/functions.
- one or more data sets and/or modules may be provided by an external memory (e.g., an external drive in wired or wireless communication with the processing system 600 ) or may be provided by a transitory or non-transitory computer-readable medium.
- Examples of non-transitory computer readable media include a RAM, a ROM, an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM), a flash memory, a CD-ROM, or other portable memory storage.
- the processing system 600 may also include a bus 612 providing communication among components of the processing system 600 , including the processing device(s) 602 , I/O interface(s) 604 , network interface(s) 606 , storage unit(s) 608 and/or memory(ies) 610 .
- the bus 612 may be any suitable bus architecture including, for example, a memory bus, a peripheral bus or a video bus.
- the input device(s) 614 e.g., a keyboard, a mouse, a microphone, a touchscreen, and/or a keypad
- output device(s) 616 e.g., a display, a speaker and/or a printer
- the input device(s) 614 and/or the output device(s) 616 may be included as a component of the processing system 600 .
- there may not be any input device(s) 614 and output device(s) 616 in which case the I/O interface(s) 604 may not be needed.
- the memory(ies) 610 may include computer-executable instructions for a self-training/semi-supervised module 618 that, when executed, cause the processing system 600 to perform the self-training/semi-supervised method 200 .
- the memory(ies) 610 may further store training data including the datasets of databank 108 .
- the neural networks used to implement generator G(z) 102 and discriminator D 104 may be implemented by any suitable processing unit, including the processing system 600 or variant thereof. Further, any suitable neural network, including variations such as recurrent neural networks long short-term memory (LSTM) neural networks, or any other neural network, may be used.
- LSTM long short-term memory
- the present disclosure may be described, at least in part, in terms of methods, a person of ordinary skill in the art will understand that the present disclosure is also directed to the various components for performing at least some of the aspects and features of the described methods, be it by way of hardware components, software or any combination of the two. Accordingly, the technical solution of the present disclosure may be embodied in the form of a software product.
- a suitable software product may be stored in a pre-recorded storage device or other similar non-volatile or non-transitory computer readable medium, including DVDs, CD-ROMs, USB flash disk, a removable hard disk, or other storage media, for example.
- the software product includes instructions tangibly stored thereon that enable a processing system (e.g., a personal computer, a server, or a network device) to execute examples of the methods disclosed herein.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
Abstract
Description
minG maxD {E x˜P
lossD =E x˜plab(x)[∥D(x)−y∥ 2]+E z˜pz(z),x˜punl(x)[D(G(z))−D(x)] Equation (3)
-
- plab(x) is the distribution of the labelled samples xlab;
- ∥D(x)−y∥2 is the square of the distance between the estimated label D(x) for a labelled sample xlab and the ground truth label y for the labelled sample xlab;
-
- pz(z) is the distribution of the noise z input into generator G;
- punl(x) is the distribution of the unlabelled samples xlab; and
- [D(G(z))−D(x)] is the confidence score difference between a generated sample and a real sample (e.g. training sample) belonging to a certain class.
lossG =E z˜pz(z),x˜punl(x)[(D(G(z))−D(x))2] Equation (4)
lossD =E x˜plab(x)[∥D(x)−y∥ 2]2 +E z˜pz(z),x˜punl(x)[D(G(z))−D(x)]2 Equation (3-A)
lossG =E z˜pz(z),x˜punl(x)[(D(G(z))−D(x))2]2 Equation (4-A)
P(y i |x)=[exp(outputi)]/[ΣK k=1 exp(outputk)] Equation (5)
Claims (20)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/789,628 US11120337B2 (en) | 2017-10-20 | 2017-10-20 | Self-training method and system for semi-supervised learning with generative adversarial networks |
PCT/CN2017/108191 WO2019075771A1 (en) | 2017-10-20 | 2017-10-28 | Self-training method and system for semi-supervised learning with generative adversarial networks |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/789,628 US11120337B2 (en) | 2017-10-20 | 2017-10-20 | Self-training method and system for semi-supervised learning with generative adversarial networks |
Publications (2)
Publication Number | Publication Date |
---|---|
US20190122120A1 US20190122120A1 (en) | 2019-04-25 |
US11120337B2 true US11120337B2 (en) | 2021-09-14 |
Family
ID=66170689
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/789,628 Active 2040-07-15 US11120337B2 (en) | 2017-10-20 | 2017-10-20 | Self-training method and system for semi-supervised learning with generative adversarial networks |
Country Status (2)
Country | Link |
---|---|
US (1) | US11120337B2 (en) |
WO (1) | WO2019075771A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11710035B2 (en) | 2018-09-28 | 2023-07-25 | Apple Inc. | Distributed labeling for supervised learning |
Families Citing this family (61)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10706327B2 (en) * | 2016-08-03 | 2020-07-07 | Canon Kabushiki Kaisha | Information processing apparatus, information processing method, and storage medium |
US11593632B2 (en) * | 2016-12-15 | 2023-02-28 | WaveOne Inc. | Deep learning based on image encoding and decoding |
JP7023669B2 (en) * | 2017-10-26 | 2022-02-22 | 株式会社Preferred Networks | Image generation method, image generation device, and image generation program |
CA3022998A1 (en) * | 2017-11-02 | 2019-05-02 | Royal Bank Of Canada | Method and device for generative adversarial network training |
CN109784325A (en) * | 2017-11-10 | 2019-05-21 | 富士通株式会社 | Opener recognition methods and equipment and computer readable storage medium |
CN111712835B (en) * | 2018-01-02 | 2023-09-01 | 诺基亚技术有限公司 | Channel modeling in a data transmission system |
US10565475B2 (en) * | 2018-04-24 | 2020-02-18 | Accenture Global Solutions Limited | Generating a machine learning model for objects based on augmenting the objects with physical properties |
US11615208B2 (en) * | 2018-07-06 | 2023-03-28 | Capital One Services, Llc | Systems and methods for synthetic data generation |
US11537277B2 (en) * | 2018-07-19 | 2022-12-27 | Palo Alto Research Center Incorporated | System and method for generating photorealistic synthetic images based on semantic information |
CN109300107B (en) * | 2018-07-24 | 2021-01-22 | 深圳先进技术研究院 | Plaque processing method, apparatus and computing device for magnetic resonance blood vessel wall imaging |
JP7014100B2 (en) * | 2018-08-27 | 2022-02-01 | 日本電信電話株式会社 | Expansion equipment, expansion method and expansion program |
JP7279368B2 (en) * | 2019-01-17 | 2023-05-23 | 富士通株式会社 | LEARNING METHOD, LEARNING PROGRAM AND LEARNING DEVICE |
JP7163786B2 (en) | 2019-01-17 | 2022-11-01 | 富士通株式会社 | LEARNING METHOD, LEARNING PROGRAM AND LEARNING DEVICE |
US11907675B2 (en) * | 2019-01-18 | 2024-02-20 | Uber Technologies, Inc. | Generating training datasets for training neural networks |
US11315021B2 (en) * | 2019-01-28 | 2022-04-26 | StradVision, Inc. | Method and device for on-device continual learning of a neural network which analyzes input data, and method and device for testing the neural network to be used for smartphones, drones, vessels, or military purpose |
US12046038B2 (en) * | 2019-03-22 | 2024-07-23 | The Regents Of The University Of California | System and method for generating visual analytics and player statistics |
CN110074813B (en) * | 2019-04-26 | 2022-03-04 | 深圳大学 | Ultrasonic image reconstruction method and system |
WO2020227418A1 (en) * | 2019-05-06 | 2020-11-12 | Google Llc | Semi-supervised training of machine learning models using label guessing |
CN110097130B (en) * | 2019-05-07 | 2022-12-13 | 深圳市腾讯计算机系统有限公司 | Training method, device and equipment for classification task model and storage medium |
DE102019206720B4 (en) * | 2019-05-09 | 2021-08-26 | Volkswagen Aktiengesellschaft | Monitoring of an AI module of a driving function of a vehicle |
CN110188172B (en) * | 2019-05-31 | 2022-10-28 | 清华大学 | Text-based event detection method and device, computer equipment and storage medium |
CN110598530A (en) * | 2019-07-30 | 2019-12-20 | 浙江工业大学 | Small sample radio signal enhanced identification method based on ACGAN |
CN112348040B (en) * | 2019-08-07 | 2023-08-29 | 杭州海康威视数字技术股份有限公司 | Model training method, device and equipment |
CN110298415B (en) * | 2019-08-20 | 2019-12-03 | 视睿(杭州)信息科技有限公司 | A kind of training method of semi-supervised learning, system and computer readable storage medium |
KR20220047851A (en) * | 2019-08-22 | 2022-04-19 | 구글 엘엘씨 | Active Learning with Sample Concordance Assessment |
EP3786845A1 (en) * | 2019-08-29 | 2021-03-03 | Robert Bosch GmbH | Difficulty-adaptive training for machine learning modules |
US11152785B1 (en) * | 2019-09-17 | 2021-10-19 | X Development Llc | Power grid assets prediction using generative adversarial networks |
CN110617966A (en) * | 2019-09-23 | 2019-12-27 | 江南大学 | Bearing fault diagnosis method based on semi-supervised generation countermeasure network |
US11481623B2 (en) | 2019-09-25 | 2022-10-25 | International Business Machines Corporation | Systems and methods for training a model using a few-shot classification process |
CN112307860B (en) * | 2019-10-10 | 2025-02-28 | 北京沃东天骏信息技术有限公司 | Image recognition model training method and device, image recognition method and device |
CN110853703A (en) * | 2019-10-16 | 2020-02-28 | 天津大学 | Semi-supervised learning prediction method for protein secondary structure |
US20210133596A1 (en) * | 2019-10-30 | 2021-05-06 | International Business Machines Corporation | Ranking image sources for transfer learning |
EP4060572A4 (en) * | 2019-12-26 | 2023-07-19 | Telefónica, S.A. | Computer-implemented method for accelerating convergence in the training of generative adversarial networks (gan) to generate synthetic network traffic, and computer programs of same |
IT202000000664A1 (en) * | 2020-01-15 | 2021-07-15 | Digital Design S R L | GENERATIVE SYSTEM FOR THE CREATION OF DIGITAL IMAGES FOR PRINTING ON DESIGN SURFACES |
CN111460156B (en) * | 2020-03-31 | 2024-05-14 | 深圳前海微众银行股份有限公司 | Sample expansion method, device, equipment and computer readable storage medium |
CN113554177B (en) * | 2020-04-02 | 2023-10-03 | 北京航空航天大学 | Satellite power supply system autonomous fault diagnosis method based on soft decision |
CN111553267B (en) * | 2020-04-27 | 2023-12-01 | 腾讯科技(深圳)有限公司 | Image processing method, image processing model training method and device |
CN111651937B (en) * | 2020-06-03 | 2023-07-25 | 苏州大学 | Method for diagnosing faults of in-class self-adaptive bearing under variable working conditions |
CN111861924B (en) * | 2020-07-23 | 2023-09-22 | 成都信息工程大学 | A cardiac magnetic resonance image data enhancement method based on evolutionary GAN |
CN111898696B (en) * | 2020-08-10 | 2023-10-27 | 腾讯云计算(长沙)有限责任公司 | Pseudo tag and tag prediction model generation method, device, medium and equipment |
CN111794741B (en) * | 2020-08-11 | 2023-08-18 | 中国石油天然气集团有限公司 | Method for realizing sliding directional drilling simulator |
CN112183577B (en) * | 2020-08-31 | 2025-03-25 | 华为技术有限公司 | A training method, image processing method and device for a semi-supervised learning model |
CN112115467A (en) * | 2020-09-04 | 2020-12-22 | 长沙理工大学 | Intrusion detection method based on semi-supervised classification of ensemble learning |
US11270164B1 (en) * | 2020-09-24 | 2022-03-08 | Ford Global Technologies, Llc | Vehicle neural network |
CN112162515B (en) * | 2020-10-10 | 2021-08-03 | 浙江大学 | An Adversarial Attack Method for Process Monitoring System |
CN114462628A (en) * | 2020-11-09 | 2022-05-10 | 华为技术有限公司 | Data enhancement method, device, computing equipment and computer readable storage medium |
CN112418304B (en) * | 2020-11-19 | 2021-10-29 | 北京云从科技有限公司 | OCR (optical character recognition) model training method, system and device |
CN112786003A (en) * | 2020-12-29 | 2021-05-11 | 平安科技(深圳)有限公司 | Speech synthesis model training method and device, terminal equipment and storage medium |
CN112669298A (en) * | 2020-12-31 | 2021-04-16 | 武汉科技大学 | Foundation cloud image cloud detection method based on model self-training |
CN112818767B (en) * | 2021-01-18 | 2023-07-25 | 深圳市商汤科技有限公司 | Data set generation and forgery detection methods and devices, electronic equipment and storage medium |
CN112836802B (en) * | 2021-02-03 | 2024-07-12 | 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) | Semi-supervised learning method, lithology prediction method and storage medium |
CN112949646B (en) * | 2021-02-26 | 2023-12-19 | 平安科技(深圳)有限公司 | Semantic segmentation method, device, equipment and medium for electron microscopic fault data |
CN113111947B (en) * | 2021-04-16 | 2024-04-09 | 北京沃东天骏信息技术有限公司 | Image processing method, apparatus and computer readable storage medium |
CN113269228B (en) * | 2021-04-20 | 2022-06-10 | 重庆邮电大学 | A training method, device, system and electronic device for a graph network classification model |
DE102021205273A1 (en) | 2021-05-21 | 2022-11-24 | Robert Bosch Gesellschaft mit beschränkter Haftung | Method for training a discriminator to recognize a road condition |
US11599794B1 (en) * | 2021-10-20 | 2023-03-07 | Moffett International Co., Limited | System and method for training sample generator with few-shot learning |
CN113887725B (en) * | 2021-10-26 | 2024-09-20 | 南通大学 | Cervical cell image semi-supervised learning method based on generation of countermeasure network |
WO2024216545A1 (en) * | 2023-04-19 | 2024-10-24 | Huawei Technologies Co., Ltd. | Method and server for generating training data for training object detect |
CN117076871B (en) * | 2023-10-16 | 2023-12-29 | 南京邮电大学 | Battery fault classification method based on unbalanced semi-supervised countermeasure training framework |
CN117649528B (en) * | 2024-01-29 | 2024-05-31 | 山东建筑大学 | Semi-supervised image segmentation method, system, electronic equipment and storage medium |
CN117934958B (en) * | 2024-01-31 | 2024-10-11 | 贵州大学 | A classification method for active pulmonary tuberculosis images based on data enhancement |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030046297A1 (en) | 2001-08-30 | 2003-03-06 | Kana Software, Inc. | System and method for a partially self-training learning system |
US20090099988A1 (en) | 2007-10-12 | 2009-04-16 | Microsoft Corporation | Active learning using a discriminative classifier and a generative model to detect and/or prevent malicious behavior |
US20120158624A1 (en) | 2010-12-21 | 2012-06-21 | International Business Machines Corporation | Predictive modeling |
US8781782B2 (en) | 2010-09-30 | 2014-07-15 | Siemens Aktiengesellschaft | System and method for conditional multi-output regression for machine condition monitoring |
US20150199609A1 (en) | 2013-12-20 | 2015-07-16 | Xurmo Technologies Pvt. Ltd | Self-learning system for determining the sentiment conveyed by an input text |
CN105279519A (en) | 2015-09-24 | 2016-01-27 | 四川航天系统工程研究所 | Remote sensing image water body extraction method and system based on cooperative training semi-supervised learning |
US9342793B2 (en) | 2010-08-31 | 2016-05-17 | Red Hat, Inc. | Training a self-learning network using interpolated input sets based on a target output |
WO2016077127A1 (en) | 2014-11-11 | 2016-05-19 | Massachusetts Institute Of Technology | A distributed, multi-model, self-learning platform for machine learning |
CN106096627A (en) | 2016-05-31 | 2016-11-09 | 河海大学 | The Polarimetric SAR Image semisupervised classification method that considering feature optimizes |
WO2017158058A1 (en) | 2016-03-15 | 2017-09-21 | Imra Europe Sas | Method for classification of unique/rare cases by reinforcement learning in neural networks |
US20180211164A1 (en) * | 2017-01-23 | 2018-07-26 | Fotonation Limited | Method of training a neural network |
US20180336439A1 (en) * | 2017-05-18 | 2018-11-22 | Intel Corporation | Novelty detection using discriminator of generative adversarial network |
US20190327501A1 (en) * | 2018-07-06 | 2019-10-24 | Capital One Services, Llc | Real-time synthetically generated video from still frames |
-
2017
- 2017-10-20 US US15/789,628 patent/US11120337B2/en active Active
- 2017-10-28 WO PCT/CN2017/108191 patent/WO2019075771A1/en active Application Filing
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030046297A1 (en) | 2001-08-30 | 2003-03-06 | Kana Software, Inc. | System and method for a partially self-training learning system |
US20090099988A1 (en) | 2007-10-12 | 2009-04-16 | Microsoft Corporation | Active learning using a discriminative classifier and a generative model to detect and/or prevent malicious behavior |
US9342793B2 (en) | 2010-08-31 | 2016-05-17 | Red Hat, Inc. | Training a self-learning network using interpolated input sets based on a target output |
US8781782B2 (en) | 2010-09-30 | 2014-07-15 | Siemens Aktiengesellschaft | System and method for conditional multi-output regression for machine condition monitoring |
US20120158624A1 (en) | 2010-12-21 | 2012-06-21 | International Business Machines Corporation | Predictive modeling |
US20150199609A1 (en) | 2013-12-20 | 2015-07-16 | Xurmo Technologies Pvt. Ltd | Self-learning system for determining the sentiment conveyed by an input text |
WO2016077127A1 (en) | 2014-11-11 | 2016-05-19 | Massachusetts Institute Of Technology | A distributed, multi-model, self-learning platform for machine learning |
CN105279519A (en) | 2015-09-24 | 2016-01-27 | 四川航天系统工程研究所 | Remote sensing image water body extraction method and system based on cooperative training semi-supervised learning |
WO2017158058A1 (en) | 2016-03-15 | 2017-09-21 | Imra Europe Sas | Method for classification of unique/rare cases by reinforcement learning in neural networks |
CN106096627A (en) | 2016-05-31 | 2016-11-09 | 河海大学 | The Polarimetric SAR Image semisupervised classification method that considering feature optimizes |
US20180211164A1 (en) * | 2017-01-23 | 2018-07-26 | Fotonation Limited | Method of training a neural network |
US20180336439A1 (en) * | 2017-05-18 | 2018-11-22 | Intel Corporation | Novelty detection using discriminator of generative adversarial network |
US20190327501A1 (en) * | 2018-07-06 | 2019-10-24 | Capital One Services, Llc | Real-time synthetically generated video from still frames |
Non-Patent Citations (24)
Title |
---|
Bousmalis, Konstantinos, et al. "Unsupervised pixel-level domain adaptation with generative adversarial networks." Proceedings of the IEEE conference on computer vision and pattern recognition. Jul. 21-26, 2017. (Year: 2017). * |
C. Rosenberg, M. Hebert, and H. Schneiderman, "Semi-supervised self-training of object detection models", Application of Computer Vision, 2005. WACV/MOTIONS '05 vol. 1. Seventh IEEE Workshops on, 2005 2005. |
D Wu, Y Yin, H Jiang, "Large-Margin Estimation of Hidden Markov Models With Second-Order Cone Programming for Speech Recognition", IEEE Transactions on Audio, Speech, and Language Processing 19 (6), 1652-1664 2010. |
D. W. Farley and Y. Bengio, "Improving Generative Adversarial Networks with Denoising Feature Matching", ICLR 2017. 2017. |
D. Wu, "Parameter estimation for α-GMM based on maximum likelihood criterion", Neural computation 21 (6), 1776-1795 2009. |
Ian Goodfellow et al., "Generative adversarial nets", NIPS 2014. 2014. |
Kingma, D. et al. "Semi-Supervised Learning with Deep Generative Models", Online: https://arxiv.org/pdf/1406.5298 2014. |
Li, Chongxuan, et al. "Triple Generative Adversarial Nets." arXiv preprint arXiv:1703.02291 (2017). 2017. |
M. Arjovsky and L. Bottou, "Towards Principled Methods for Training Generative Adversarial Network", ICLR 2017 2017. |
M. Arjovsky et al., "Wasserstein GAN", https://arxiv.org/pdf/1701.07875.pdf 2017. |
Mao, X. et al. "Least Squares Generative Adversarial Networks", https://arxiv.org/abs/1611.04076 2017. |
Odena, A, "Semi-Supervised Learning with Generative Adversarial Networks", https://arxiv.org/pdf/1606.01583.pdf 2016. |
Rajat Raina, Alexis Battle, Honglak Lee, Benjamin Packer, Andrew Y. Ng, "Self-taught learning: transfer learning from unlabeled data", ICML '07 Proceedings of the 24th international conference on Machine learning, pp. 759-766, 2007 2007. |
Redford, A. et al. "Unsupervised representation learning with deep convolutional generative adversarial networks", ICLR 2016, https://arxiv.org/pdf/1511.06434.pdf 2016. |
S. Abney, "Understanding the Yarowsky algorithm", Journal of Computational Linguistics, vol. 30, No. 3, 2004, pp. 365-395 2004. |
Shrivastava, A. et al. "Learning from Simulated and Unsupervised Images through Adversarial Training", Apple Inc. 2016. |
Sixt, L. et al. "RENDERGAN: Generating Realistic Labeled Data", ICLR 2017, https://arxiv.org/pdf/1611.01331.pdf 2017. |
Springenberg, J.T. "Unsupervised and semi-supervised learning with categorical generative adversarial networks" ICLR 2016, https://arxiv.org/pdf/1511.06390.pdf 2016. |
Sudo, A. et al. "Associative Memory for Online Learning in Noisy Environments Using Self-Organizing Incremental Neural Network", IEEE Transactions on Neural Networks, vol. 20, No. 6, pp. 964-972 2009. |
T. Salimans et al., "Improved techniques for Training GANs", https://arxiv.org/pdf/1606.03498.pdf 2016. |
Teng, T-H, et al. "Self-Organizing Neural Networks Integrating Domain Knowledge and Reinforcement Learning", IEEE Transactions on Neural Networks and Learning Systems, vol. 26, No. 5 pp. 889-902 2015. |
Yan Zhou, Murat Kantarcioglu, and Bhavani Thuraisingham, "Self-Training with Selection-by-Rejection", 2012 IEEE 12th International Conference on Data Mining, 2012 2012. |
Yarowsky, D. "Unsupervised Word Sense Disambiguation Rivaling Supervised Methods". Proceedings of the 33rd Annual Meeting of the Association for Computational Linguistics. Cambridge, MA, pp. 189-196, 1995 1995. |
Zhi He et al, Generative Adversarial Networks-Based Semi-Supervised Learning for Hyperspectral Image Classification. Remote Sens. 2017, 9, 1042; Oct. 12, 2017, 27 pages. |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11710035B2 (en) | 2018-09-28 | 2023-07-25 | Apple Inc. | Distributed labeling for supervised learning |
US12260331B2 (en) | 2018-09-28 | 2025-03-25 | Apple Inc. | Distributed labeling for supervised learning |
Also Published As
Publication number | Publication date |
---|---|
WO2019075771A1 (en) | 2019-04-25 |
US20190122120A1 (en) | 2019-04-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11120337B2 (en) | Self-training method and system for semi-supervised learning with generative adversarial networks | |
US11003995B2 (en) | Semi-supervised regression with generative adversarial networks | |
KR102788531B1 (en) | Method and apparatus for generating fixed point neural network | |
US11423282B2 (en) | Autoencoder-based generative adversarial networks for text generation | |
US11816183B2 (en) | Methods and systems for mining minority-class data samples for training a neural network | |
US11663483B2 (en) | Latent space and text-based generative adversarial networks (LATEXT-GANs) for text generation | |
US11429860B2 (en) | Learning student DNN via output distribution | |
US11610097B2 (en) | Apparatus and method for generating sampling model for uncertainty prediction, and apparatus for predicting uncertainty | |
KR20220062065A (en) | Robust training in the presence of label noise | |
Altun et al. | Gaussian process classification for segmenting and annotating sequences | |
US11651214B2 (en) | Multimodal data learning method and device | |
JP7304488B2 (en) | Reinforcement Learning Based Locally Interpretable Models | |
WO2022217856A1 (en) | Methods, devices and media for re-weighting to improve knowledge distillation | |
US20230274150A1 (en) | Performing Inference And Training Using Sparse Neural Network | |
CN112465043A (en) | Model training method, device and equipment | |
CN114424212A (en) | Distance-based learning confidence model | |
Dinakaran et al. | Ensemble method of effective AdaBoost algorithm for decision tree classifiers | |
US20220405570A1 (en) | Post-hoc loss-calibration for bayesian neural networks | |
KR102308752B1 (en) | Method and apparatus for tracking object | |
CN111523308B (en) | Chinese word segmentation method and device and computer equipment | |
CN112396069B (en) | Semantic edge detection method, device, system and medium based on joint learning | |
Wang et al. | Adaptive contrastive learning for learning robust representations under label noise | |
JP2021197165A (en) | Information processing equipment, information processing methods and computer-readable storage media | |
CN115700615A (en) | Computer-implemented method, apparatus, and computer program product | |
CN114842920A (en) | Molecular property prediction method and device, storage medium and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: HUAWEI TECHNOLOGIES CO., LTD., CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WU, DALEI;HAIDAR, MD AKMAL;REZAGHOLIZADEH, MEHDI;AND OTHERS;REEL/FRAME:044517/0054 Effective date: 20171023 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |