[go: up one dir, main page]

Next Article in Journal
Diffusion Maximum Correntropy Criterion Based Robust Spectrum Sensing in Non-Gaussian Noise Environments
Next Article in Special Issue
Multiple Sclerosis Identification Based on Fractional Fourier Entropy and a Modified Jaya Algorithm
Previous Article in Journal
Entropy-Based Video Steganalysis of Motion Vectors
Previous Article in Special Issue
Deconstructing Cross-Entropy for Probabilistic Binary Classifiers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Graph Multi-Label Learning Based on Entropy

College of Computer Science and Engineering, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Entropy 2018, 20(4), 245; https://doi.org/10.3390/e20040245
Submission received: 25 March 2018 / Revised: 30 March 2018 / Accepted: 30 March 2018 / Published: 2 April 2018
(This article belongs to the Special Issue Entropy-based Data Mining)

Abstract

:
Recently, Multi-Graph Learning was proposed as the extension of Multi-Instance Learning and has achieved some successes. However, to the best of our knowledge, currently, there is no study working on Multi-Graph Multi-Label Learning, where each object is represented as a bag containing a number of graphs and each bag is marked with multiple class labels. It is an interesting problem existing in many applications, such as image classification, medicinal analysis and so on. In this paper, we propose an innovate algorithm to address the problem. Firstly, it uses more precise structures, multiple Graphs, instead of Instances to represent an image so that the classification accuracy could be improved. Then, it uses multiple labels as the output to eliminate the semantic ambiguity of the image. Furthermore, it calculates the entropy to mine the informative subgraphs instead of just mining the frequent subgraphs, which enables selecting the more accurate features for the classification. Lastly, since the current algorithms cannot directly deal with graph-structures, we degenerate the Multi-Graph Multi-Label Learning into the Multi-Instance Multi-Label Learning in order to solve it by MIML-ELM (Improving Multi-Instance Multi-Label Learning by Extreme Learning Machine). The performance study shows that our algorithm outperforms the competitors in terms of both effectiveness and efficiency.

1. Introduction

Due to the advance of smart phones, nowadays people upload a great number of photos to the Internet. Updating photos has become easier, but searching them becomes more difficult. Though the technology of searching images by images has appeared, most people rely on the traditional way to searching an image, which is searching images by typing keywords. For that, we need to add labels for each image, but it cannot be accomplished by human beings due to the great number of unlabeled images. Thus, it is important to use Machine Learning to automatically classify images [1,2] and add correct labels for them.
Multi-Instance Learning is extensively studied in image classification [3]. It uses a kind of data structure called Feature Vector to represent a real image [4], but it is a little bit imprecise, since vectors can only show the pixels of images without the adjacency relations between pixels. Thus, it is natural to consider that Graph may be a better data structure to represent an image [5] because a Graph consists of edges and nodes. Nodes can indicate the texture or color of pixels in an image and the edges can indicate the adjacency relations of nodes. Like the following image in Figure 1a, a Feature Vector can only tell you there are white, blue and green pixels in this image, whereas, after segmenting the image into three Graphs [6] in Figure 1b, Graphs can show the adjacency relations between each pixel. (In the real application, the image will be segmented into more graphs with more nodes.) The latter one represents more real details of an image, which will be more beneficial for the accuracy in the learning part.
Recently, a graph-structure algorithm, namely gMGFL [7,8,9,10], was proposed by Jia Wu et al. Briefly, it works in the following steps. Firstly, there are many different images in the training dataset. gMGFL segments each image into multiple graphs, all which are packed into a graph bag. Secondly, the label is only visible for the graph bag and each graph bag will only be marked with one label. For a specific subject, if an image contains it, the label of the graph bag for this image will be positive; on the contrary, it will be negative. Thirdly, to build an appropriate classifier, it needs to mine some informative subgraphs, which can stand for the traits of the subject in the images (or the traits of the subject not in the images), and use these subgraphs as the features for classifying. It is brilliant, but it still has some drawbacks. Two major problems of gMGFL are listed as below.
Firstly, in the algorithm gMGFL, each image will only be marked with one label, so it can only deal with one subject. Nevertheless, in the real life, it is impossible that an image just contains one subject. It often includes multiple semantic information [11]. For example, in Figure 2, the image contains three different subjects: sea, boat and sky, so it should be marked with three kinds of positive labels (and maybe also marked with some negative labels, like lion or apple). It cannot mark the image with only one label, like sea. Otherwise, it will cause some problems: if the user types a searching keyword boat or sky, this image will not be shown in the result. Unfortunately, gMGFL can only deal with a one-label problem.
Secondly, in the part of mining informative subgraphs, gMGFL considers that if a subgraph is informative, it should be frequent in the dataset. Therefore, gMGFL mines the frequent subgraphs [12] and uses them as the features for classifying, but this is not always accurate. In Figure 3, there are eight graphs and two different classes. The above four are marked with positive labels and belong to the positive class; the four below are marked with negative labels and belong to the negative class. If we only consider the frequent-subgraph mining, the subgraph A B has the frequency of eight because it appears eight times in all graphs. Nevertheless, if we regard it as an informative subgraph (a.k.a. classifying feature), it cannot distinguish the positive class from the negative class, since it is a common feature between two classes. Not only do all positive graphs contain the subgraph A B , but also all negative ones contain it as well. Thus, the subgraph A B is not appropriate to be an informative feature, but, due to its high frequency, gMGFL considers that it is, which will cause imprecise results.
To solve these problems, in this paper, we proposed an advanced graph-structure algorithm named Multi-Graph Multi-Label Learning. This algorithm may also be utilized in calculating the similarity of biological sequences, predicting the function of chemical compounds, analyzing structured texts such as HTML (Hypertext Markup Language) and XML (Extensible Markup Language), etc. The following are our major contributions:
  • Our algorithm is based on a multi-graph and it can also solve multi-label (i.e., multiple subjects) problems, which means it can deal with multiple semantic information. To the best of our knowledge, we are the first one to propose such an algorithm.
  • We introduce a novel subgraph-mining technique called Entropy-Based Subgraph Mining. It calculates the information entropy for each graph [13] and uses this criterion to mine the informative subgraphs, which is more suitable than the one based on frequent-subgraph.
  • In the part of building the classifier, we utilize the algorithm MIML-ELM (we will discuss it briefly in Section 2.3). It uses the Extreme Learning Machine rather than Support Vector Machine to build an image classifier, which is more efficient.
The rest part of this paper is organized as the following. Related works are introduced in Section 2. The algorithm description of Multi-Graph Multi-Label Learning is presented in Section 3. The results of our experiments are provided in Section 4. Our conclusions are in Section 5.

2. Related Work

The research in this paper is related to some previous works of graph-structure classification, Multi-Instance Multi-Label Learning and MIML-ELM. We will briefly review them respectively in Section 2.1, Section 2.2 and Section 2.3.

2.1. Graph-Structure Classification

There are two kinds of algorithms about graph-structure classification: one of them is based on global distance and the other one is based on subgraph feature, and it has been proved that the subgraph-feature approach is better [14]. It converts a set of subgraphs into feature vectors so that most of the current algorithms can be utilized in the graph classification problem. Almost all these kinds of algorithms (such as AGM [15], Gaston [16], gSpan [17,18], gHSIC [19,20]) extract subgraph features by using frequent substructure pattern mining, and the most widespread mining algorithm among them is gSpan.

2.2. Multi-Instance Multi-Label Learning

Multi-Instance Multi-Label Learning is a supervised learning algorithm, which represents real-world objects with bags of instances and labels. The most widespread algorithm is MIML-SVM [21]. It degenerates Multi-Instance Multi-Label to Single-Instance Multi-Label by clustering multiple labels to binary classification tasks with Support Vector Machine. Zhou et al. proposed MIML-SVM in [22] and recently Li et al. improved it in [23].

2.3. MIML-ELM

The full name of MIML-ELM is Improving Multi-Instance Multi-Label Learning by Extreme Learning Machine [24]. Extreme Learning Machine (ELM) is one of the models in Neural Networks and is extensively utilized in Single Hidden-layer Feed-forward Network. Recently, Li et al. proposed an efficient and effective algorithm named MIML-ELM, which utilized the ELM in solving the Multi-Instance Multi-Label problem. Firstly, this algorithm is more effective in the process of degeneration from Multi-Instance Multi-Label to Single-Instance Multi-Label, since it provides a theoretical guarantee to automatically determine the number of clusters. Secondly, this algorithm is more efficient, since it uses ELM instead of Support Vector Machine to improve the two-phase framework.

3. The MGML Algorithm

This section is about the algorithm description of Multi-Graph Multi-Label Learning (MGML). Firstly, we will introduce some relative concepts in Section 3.1. Then, we discuss our proposed approach in Section 3.2, Section 3.3 and Section 3.4. Lastly, an illustrative example of MGML is given in Section 3.5.

3.1. Problem Definition

The following are some basic definitions about our algorithm.
Definition 1.
(Graph Bag):G is a graph denoted as G = ( N , E , L , l ) . N is a set of nodes; E is a set of edges and E N × N ; L is a set of labels for nodes and edges; l is the function mapping labels to nodes and edges and l : N E L . A graph bag B a g = { G 1 , , G j , , G n } contains n graphs, where G j denotes the j-th graph in the bag.
Definition 2.
(Subgraph):subgraph Given G = ( N , E , L , l ) and S u b G k = ( N , E , L , l ) , we say that S u b G k is a subgraph of G, if and only if there exists a function ψ : N N s.t. (1) n N , l ( n ) = l ( ψ ( n ) ) ; (2) ( n 1 , n 2 ) E , ( ψ ( n 1 ) , ψ ( n 2 ) ) E , and l ( n 1 , n 2 ) = l ( ψ ( n 1 ) , ψ ( n 2 ) ) . In addition, we can say that G is a super-graph of S u b G k .
Definition 3.
(Subgraph Feature Representation for Graph):Let S e t S u b G = { S u b G 1 , , S u b G s } be a set of subgraphs mined from a graph set S e t G = { G 1 , , G t } . For each graph G i ( i [ 1 , t ] ) in S e t G , we use a feature vector X i G = [ ( X i S u b G 1 ) G , , ( X i S u b G k ) G , , ( X i S u b G s ) G ] T to represent it. ( X i S u b G k ) G = 1 if and only if S u b G k is a subgraph of G i . Otherwise, ( X i S u b G k ) G = 0 .

3.2. Overall Framework of MGML

The framework of MGML includes two major steps: (1) Mining Informative Subgraphs. Our novel subgraph-mining technique Entropy-based Subgraph Mining will be utilized in this part and it includes the following steps. Firstly, gSpan will be utilized to generate all subgraphs in the dataset. Secondly, entropies of all subgraphs will be calculated and ranked according to our informative-subgraph criterion based on information entropy. We will discuss all details in Section 3.3; (2) Building Classifier. Top-ranked subgraphs will be used as classifying features. Graphs can be represented as instances based on what kinds of classifying features (a.k.a. informative subgraphs) that they contain, so graphs can be represented as multiple instances. Thus, Multi-Graph Multi-Label degenerates to Multi-Instance Multi-Label. After that, we will utilize MIML-ELM, an efficient and effective MIML algorithm to build a classifier. We will discuss all details in Section 3.4.

3.3. Mining Informative Subgraphs

In this section, we will discuss the evaluation of informative subgraphs and the algorithm how to mine them.

3.3.1. Evaluation of Informative Subgraphs

Let us reconsider the example in Figure 3. Although another subgraph B C has the frequency of only three, it appears three times in four positive graphs and does not appear in the negative ones at all, so it can stand for the trait of the positive class and is suitable to be regarded as a classifying feature. Generally, if a subgraph appears frequently in one of the classes but hardly appears in the other class, according to the definition of information entropy, this subgraph has low entropy. Thus, the subgraphs that have low entropy are the informative subgraphs that we need. We will give the formal definition of the informative subgraph in the following. Firstly, we will define it in the single-label problem for the ease of understanding and then expand it to a multi-label problem.
Firstly, we give the definition of information entropy for subgraph in the single-label problem. Assume that there is a set of graphs S e t G = { G 1 , , G m } . Each graph G i ( i [ 1 , m ] ) is only marked with a single label and the label is either positive or negative. S e t S u b G is the complete set of subgraphs mined from S e t G , which is denoted as S e t S u b G = { S u b G 1 , , S u b G n } . For each subgraph S u b G j ( j [ 1 , n ] ) , the set of super-graphs for it is S e t G j = { G 1 j , , G u j } . # p o s is the number of positive graphs in S e t G j and # n e g is the number of negative graphs in S e t G j ( # p o s + # n e g = u ). Since each graph G k j ( k [ 1 , u ] ) in S e t G j is the super-graph of S u b G j , the possibility of S u b G j appearing in a positive (or negative) class equals the percentage of positive (or negative) graphs in S e t G j (based on Definition 2). Thus, the information entropy of subgraph S u b G j is E j = p p o s log 2 ( p p o s ) p n e g log 2 ( p n e g ) . p p o s is the possibility of S u b G j appearing in a positive class and p p o s = # p o s u ; p n e g is the possibility of S u b G j appearing in a negative class and p n e g = # n e g u . The information entropy for the set of subgraphs S e t S u b G is S e t E ( S u b G ) = { E 1 , , E n } .
Furthermore, the following is the definition of information entropy for a subgraph in a multi-label problem. Assume that there is a set of graphs S e t G = { G 1 , , G m } and each graph G i ( i [ 1 , m ] ) is marked with a set of labels S e t L i = { L i , 1 , , L i , t } . S e t S u b G is the complete set of subgraphs mined from S e t G , which is denoted as S e t S u b G = { S u b G 1 , , S u b G n } . For each subgraph S u b G j ( j [ 1 , n ] ) , it has a set of information entropy S e t E j = { E j , 1 , , E j , t } (t different kinds of information entropy for t different labels). We define that, for each subgraph, the information entropy in a multi-label problem is the average entropy of all labels. Thus, for the subgraph S u b G j , the information entropy in a multi-label problem is E j = a v g { E j , 1 , , E j , t } . In the case of multi-label, the information entropy for the set of subgraphs S e t S u b G is denoted as S e t E ( S u b G ) = { E 1 , , E n } .
Lastly, S e t E ( S u b G ) will be ranked increasingly. The top-ranked subgraphs (i.e., the ones with lower entropy) are the informative subgraphs.

3.3.2. Entropy-Based Subgraph Mining

Current algorithms about classifying graph-structure data can be categorized into two groups: one is based on global distance, including graph kernel [25,26], graph embedding [27] and graph transformation [28], which calculates the similarity rate of two graphs; the other one is based on subgraph feature [29], including AGM, Gaston, gSpan and gHSIC, which converts a set of subgraphs to feature vectors. It has been proved that the latter one is better. It converts subgraphs to vectors so that most of the current algorithms can be utilized in the graph-structure classification problem.
To mine informative subgraphs as classifying features, one of the straightforward approaches is to mine the complete set of subgraphs for the graph set and rank these subgraphs with the evaluation function in Section 3.3.1, but this approach will cause a problem: the number of subgraphs grows exponentially when the size of graph set grows, so the enumeration will be tough work. Alternatively, we use a Depth-First-Search (DFS) based on algorithm gSpan to generate all subgraphs, using our evaluation during the process.
The key idea of gSpan is to build a lexicographic order among graphs that need to be mined, and then give each graph a unique label named minimum DFS (Depth-First-Search) code. gSpan uses the strategy of depth-first search to mine the frequent subgraph with the DFS code. Each time it needs to generate a new subgraph, it just recurs the character string (i.e., DFS code), so a subgraph-mining problem can be transformed into a substring-matching problem. Thus, the gSpan performs better than previous similar algorithms.
Generally, gSpan is utilized for mining the frequent subgraphs, but we do not care about the frequency of subgraph. We only use the gSpan to traverse all subgraphs and evaluate the information entropy during the traversal. This is the general idea of Entropy-based Subgraph Mining (ESM). The detailed algorithm of ESM is described in Algorithms 1 and 2.
Algorithm 1: GraphSet_Projection
Entropy 20 00245 i001
Note that Algorithm 1 invokes Algorithm 2 and Algorithm 2 is a recursion function. Firstly, it generates all subgraphs by traversing the graph from one edge (lines 4–9 in Algorithm 1). The search space is shrunk at the end of each turn (line 8 in Algorithm 1). Algorithm 2 traverses all graphs and generates all their subgraphs. It will stop when the code of the subgraph is not a minimum code (line 1 in Algorithm 2). The information entropy of the subgraph is computed and the result set is built by conducting Lines 4–11 in Algorithm 2. Lines 15–19 in Algorithm 2 grows the subgraph and does this function recursively.

3.4. Building Classifier

Since current algorithms of classifiers cannot be utilized directly in the graph-structure, after mining the informative subgraphs, we need to degenerate Multi-Graph Multi-Label (MGML) to Multi-Instance Multi-Label (MIML) based on Definition 3. The general idea is that, assuming there is a set of graphs S e t G = { G 1 , , G m } and a set of informative subgraphs S e t I n f o S u b G = { I n f o S u b G 1 , , I n f o S u b G n } mined from S e t G . For each graph G i ( i [ 1 , m ] ) , it equals a feature vector V i = ( x 1 , , x n ) (a.k.a. instance) and the dimension of it is n (that equals to the number of informative subgraphs). For each informative subgraph I n f o S u b G j ( j [ 1 , n ] ) , if G i is the super-graph of I n f o S u b G j , x j in V i has x j = 1 ; otherwise, x j = 0 . The labeled MIML set is D = { ( B i , Y i ) | i = 1 , , N } , where N is the number of bags in dataset D and B i = { X 1 i , , X j i , , X n i i } is an instance bag with n i instances, Y i { 0 , 1 } M is the label vector of bag B i . Now, the MGML problem is degenerated to the MIML problem.
Algorithm 2:ESM
Entropy 20 00245 i002
Traditionally, Support Vector Machine (SVM) is utilized to solve the MIML, but it has some drawbacks. First, SVM requires the user to input a great number of parameters. Second, using SVM to build a classifier may cause a high computing cost and the performance depends on the specific parameters that the user defined. Thus, we choose to use the Extreme Learning Machine (ELM) to solve the MIML problem. Firstly, ELM develops a theoretical guarantee to determine the number of clusters by AIC [30]. Secondly, a k-medoids cluster process is performed to transform from Multi-Instance Multi-Label into Single-Instance Multi-Label. Then, the Hausdorff distance [31] is used to measure the similarity between two different multi-instance bags. Based on the Hausdorff distance, k-medoids cluster divides the dataset into k parts, the medoids of which are M 1 , , M k . At last, we train the classifier for each label with k-dimensional numerical vectors.

3.5. Example of MGML

In this section, we will give an explanatory example of MGML. In Figure 4a, there are three images and two labels. In image 1, label Lion is positive (+) and label Tree is positive (+). In image 2, label Lion is positive (+) and label Tree is negative (−). In image 3, label Lion is negative (−) and label Tree is negative (−). The following are the brief steps to build a classifier with MGML.
Firstly, segment these images into multiple graphs like Figure 4b. Secondly, utilize Entropy-based Subgraph Mining to mine informative subgraphs. The result is in Figure 4c. Thirdly, transform Multi-Graph to Multi-Instance. The result is in Figure 5a. Lastly, utilize MIML-ELM to build a classifier. The formula (i.e., relation) between informative subgraphs and labels is in Figure 5b. For example, subgraph A D is the trait when label Tree is negative (−).
Assume that there is a new image without labels in Figure 6a. Segment the image into multiple graphs like Figure 6b. If we want to add labels for it with our MGML classifier, we just need to see what kinds of classifying features (i.e., the informative subgraphs in the previous step) it contains. It contains A D and B D , so it should have a negative (−) label Lion and a positive (+) label Tree.

4. Experimental Section

The following experiments are performed on a PC running Linux with an Intel dual-core CPU (2.60 GHz) (Shenyang, China) and 16 GB memory.

4.1. Datasets

Experiments are performed on three image datasets. These datasets have different sizes, including a small size of dataset named MSRC v2 (Microsoft Research Cambridge) [32], a middle size of dataset named Scenes [22,33] and a large size of dataset named Corel5K [34].
The summary of three datasets is given in Table 1.
We use two ways to segment the original datasets: the first one is to segment each image into six graph structures and each graph has eight nodes and 12 edges roughly ( 6 × 8 × 12 ); the second one is to segment each image into six graph structures and each graph has nine nodes and 15 edges roughly ( 6 × 9 × 15 ). In short, graphs in 6 × 9 × 15 set are more complex to mining than the 6 × 8 × 12 one.
In the following experiments, each dataset will be randomly divided to a training dataset and a testing dataset and the ratio of them is about 2:1. The training dataset will be used to build the classifier and the testing dataset will be used to evaluate its performance. All experiments repeatedly run thirty times, and each time the training dataset and the testing dataset will be divided randomly.

4.2. Evaluation Criteria

Assume that there is a test dataset S = { ( X 1 , Y 1 ) , ( X 2 , Y 2 ) , , ( X p , Y p ) } . h ( X i ) denotes a set of correct labels of X i ; h ( X i , y ) denotes the confidence for y to be a correct label of X i ; r a n k h ( X i , y ) denotes the rank of y derived from h ( X i , y ) . The following are four evaluation criteria to measure the performance of our MGML algorithm.
  • RankingLoss = 1 p i = 1 p 1 | | Y i | | | | Y i ¯ | | | { ( y 1 , y 2 ) | h ( X i , y 1 ) h ( X i , y 2 ) , ( y 1 , y 2 ) Y i × Y i ¯ } | , where Y i denotes the complementary set of Y i in Y. It indicates the average fraction of labels that are misordered for a specific object. The smaller the value of RankingLoss is, the better the performance reaches. When it equals to 0, the performance reaches perfect.
  • OneError = 1 p i = 1 p [ arg max y Y h ( X i , y ) Y i ] . It indicates the average time that the top labels in the rank are not the correct ones for a specific object. The smaller the value of OneError is, the better the performance reached. When it equals 0, the performance can reach perfection.
  • Coverage = 1 p max y Y r a n k h ( X i , y ) 1 . It indicates the average number of labels in the rank that need to be included to cover all the correct labels for a specific object. The smaller the value of Coverage is, the better the performance.
  • Average Precision a v g p r e c S ( h ) = 1 p i = 1 p 1 | Y i | y Y i | { y | r a n k h ( X i , y ) r a n k h ( X i , y ) r a n k h ( X i , y ) , y Y i } | r a n k h ( X i , y ) . It indicates the the average fraction of correct labels in all labels Y i . The larger the value of Average Precision is, the better the performance. When it equals 1, the performance can reach perfection.

4.3. Effectiveness

In this section, we will use ①, ② and ③ to mark different algorithms for the differentiation, and we will call our MGML algorithm ① MGML-ELM, which means that “the MGML algorithm using MIML-ELM”.
Currently, there are no other methods of MGML learning that can be compared to our algorithm. Thus, we use the ③ MIML-SVM, one of the state-of-the-art algorithms for MIML learning, as the competitor. In addition, we use the ② MGML-SVM as the baseline algorithm for competitions. The ② MGML-SVM algorithm is generally the same as the ① MGML-ELM. It also needs to degenerate the MGML problem into the MIML problem, but then it uses the SVM instead of ELM in the next step.
The parameter of ① MGML-ELM is the number of hidden layer (hn), which is respectively set to 50, 100, 150, 200; the parameter of ② MGML-SVM and ③ MIML-SVM is the penalty factor of Gaussian kernel (Cost), which is respectively set to 1, 2, 3, 4, 5. The final results on average are in Table 2, Table 3 and Table 4. The bold one means the best performance for every criterion. The ↓ indicates the smaller the better, while the ↑ indicates the larger the better.
For ease of reading, we use ①, ② and ③ in this paragraph to respectively indicate ① MGML-ELMMGML-SVM and ③ MIML-SVM. As seen from the results in Table 2, in the dataset MSRC v2, our algorithm ① performs best when the h n = 100 and the precision reaches 82 % , RankingLoss reaches 0.07 , OneError reaches 0.17 and Coverage reaches 3.93 , while the precision of ② reaches 76 % at best when C o s t = 5 , RankingLoss reaches 0.14 , OneError reaches 0.19 and Coverage reaches 6.19 , and ③ reaches 75 % at best when C o s t = 5 , RankingLoss reaches 0.10 , OneError reaches 0.24 and Coverage reaches 4.88 .
As seen from the results in Table 3, in the dataset Scenes, our ① reaches 81 % at best when the h n = 150 , RankingLoss reaches 0.17 , OneError reaches 0.30 and Coverage reaches 1.78 , while ② reaches 74 % at best when C o s t = 3 , RankingLoss reaches 0.29 , OneError reaches 0.34 and Coverage reaches 1.28 , and ③ reaches 25 % at best when C o s t = 5 , RankingLoss reaches 0.92 , OneError reaches 0.96 and Coverage reaches 3.87 .
As seen from the results in Table 4, in the dataset Corel5K, our ① reaches 23 % at best when the h n = 200 , RankingLoss reaches 0.21 , OneError reaches 0.75 and Coverage reaches 119.31 , while ② reaches 13 % at best when C o s t = 5 , RankingLoss reaches 0.31 , OneError reaches 0.88 and Coverage reaches 144.87 , and ③ reaches 22 % at best when C o s t = 5 , RankingLoss reaches 0.19 , OneError reaches 0.77 and Coverage reaches 110.52 . Thus, we can say that our ① MGML-ELM achieves the best performance in all cases.

4.4. Efficiency

In this section, we test the efficiency of our MGML algorithm. For either MGML or MIML, as long as the input numbers of features for ELM and SVM are the same, the runtime for both of these two algorithms will be equal. Therefore, the only way to test the efficiency is focusing on the part of mining the classifying features. Our MGML uses an innovate technique Entropy-based Subgraph Mining (ESM) to mine the informative subgraphs as the features, while the gMGFL of Jia Wu mines the frequent subgraphs as the features, which is based on the boosting gSpan named gboost [35]. We respectively compared the time of our ESM for mining subgraphs with gboost in three datasets: MSRC v2, Scenes and Corel5K.
We implement ESM and gboost in two kinds of segmented datasets ( 6 × 9 × 15 and 6 × 8 × 12 ) and the minimum frequency is set to 5%, 10%, 15%, and 20%. In ESM, if the frequency of a graph is lower than the minimum frequency we set, we will not calculate the entropy for this graph; while, in gboost, if the frequency of a graph is lower than the minimum frequency, we will not continue to mine the subgraphs for this graph. In short, the lower the minimum frequency is, the more graphs the algorithm needs to mine. The final results in average are in Figure 7a–c. The time results are logarithmic.
As seen from the results in Figure 7a–c, gboost takes hours to mine the huge datasets, but ESM takes only several minutes to generate results. For example, in the dataset MSRC v2 ( 6 × 9 × 15 ), when the minimum frequency is set to 5%, ESM takes two minutes to mine the results while gboost takes 4 h to do that; in the dataset Scenes ( 6 × 9 × 15 ), when the minimum frequency is set to 10%, ESM takes 4 s to mine the results while gboost takes 10 min to do that; in the dataset Corel5K ( 6 × 9 × 15 ), when the minimum frequency is set to 15%, ESM takes 19 s to mine the results while gboost takes 16 min to do that. These figures show that ESM achieves better performance by 100–1000 times in comparison with gboost.

5. Conclusions

In this paper, we have shown how the Multi-Graph Multi-Label Learning (MGML) works. The MGML uses the more precise structures, multiple graphs, instead of instances to represent an image, which can improve the accuracy of classification dramatically in the latter step. In addition, it uses multiple labels as the output to eliminate the ambiguity of description. Furthermore, we use our technique Entropy-based Subgraph Mining to mine the informative subgraphs, rather than simply regard frequent subgraphs as informative subgraphs. Then, we show how to degenerate MGML to MIML. At last, we use the MIML-ELM as the base classifier. Extensive experimental results prove that MGML achieves a good performance in three image datasets with different sizes. What we are interested in for the future steps is to improve the performance in the dataset that has sparse labels by using other algorithms as the classifier.

Acknowledgments

This work was supported by the National Natural Science Foundation Program of China (61772124 and 61702381), the State Key Program of National Natural Science of China (61332014), the National Training Program of Innovation and Entrepreneurship for Undergraduates (201710145082), the Fundamental Research Funds for the Central Universities under Grant 150402002, and the Peak Discipline Construction of Computer Science and Technology under Grant 02190021821001.

Author Contributions

Zixuan Zhu and Yuhai Zhao conceived and designed the experiments; Zixuan Zhu and Yuhai Zhao wrote and revised the paper; Yuhai Zhao made substantial contributions to the conception of the work; Zixuan Zhu acquired and analyzed the data; and Zixuan Zhu performed the experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, W.; Vong, C.; Yang, Y.; Wong, P. Encrypted image classification based on multilayer extreme learning machine. Multidimens. Syst. Signal Process. 2017, 28, 851–865. [Google Scholar] [CrossRef]
  2. Yan, K.; Li, Z.; Zhang, C. A New multi-instance multi-label learning approach for image and text classification. Multimed. Tools Appl. 2016, 75, 7875–7890. [Google Scholar] [CrossRef]
  3. Tang, P.; Wang, X.; Feng, B.; Liu, W. Learning Multi-Instance Deep Discriminative Patterns for Image Classification. IEEE Trans. Image Process. 2017, 26, 3385–3396. [Google Scholar] [CrossRef] [PubMed]
  4. Chaiyakhan, K.; Kerdprasop, N.; Kerdprasop, K. Feature Selection Techniques for Breast Cancer Image Classification with Support Vector Machine. Lect. Notes Eng. Comput. Sci. 2016, 2221, 327–332. [Google Scholar]
  5. Bashier, H.; Hoe, L.; Hui, L.; Azli, M.; Han, P.; Kwee, W.; Sayeed, M. Texture classification via extended local graph structure. Optik Int. J. Light Electron Opt. 2016, 127, 638–643. [Google Scholar] [CrossRef]
  6. Wang, W.; Li, Z.; Yue, J.; Li, D. Image segmentation incorporating double-mask via graph cuts. Comput. Electr. Eng. 2016, 54, 246–254. [Google Scholar] [CrossRef]
  7. Wu, J.; Pan, S.; Zhu, X.; Zhang, C.; Wu, X. Positive and Unlabeled Multi-Graph Learning. IEEE Trans. Cybern. 2017, 47, 818–829. [Google Scholar] [CrossRef] [PubMed]
  8. Wu, J.; Hong, J.; Pan, S.; Zhu, X.; Cai, Z.; Zhang, C. Multi-graph-view subgraph mining for graph classification. Knowl. Inf. Syst. 2016, 48, 29–54. [Google Scholar] [CrossRef]
  9. Wu, J.; Zhu, X.; Zhang, C.; Yu, P.S. Bag Constrained Structure Pattern Mining for Multi-Graph Classification. IEEE Trans. Knowl. Data Eng. 2014, 26, 2382–2396. [Google Scholar] [CrossRef]
  10. Wu, J.; Zhu, X.; Zhang, C.; Cai, Z. Multi-instance Multi-graph Dual Embedding Learning. In Proceedings of the 2013 IEEE 13th International Conference on Data Mining, Dallas, TX, USA, 7–10 December 2013; pp. 827–836. [Google Scholar]
  11. Wei, Y.; Xia, W.; Lin, M.; Huang, J.; Ni, B.; Dong, J.; Zhao, Y.; Yan, S. HCP: A Flexible CNN Framework for Multi-Label Image Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1901–1907. [Google Scholar] [CrossRef] [PubMed]
  12. Nair, J.; Thomas, S.; Thampi, S.; El-Alfy, E. Improvised Apriori with frequent subgraph tree for extracting frequent subgraphs. J. Intell. Fuzzy Syst. 2017, 32, 3209–3219. [Google Scholar] [CrossRef]
  13. Bai, L.; Hancock, E. Fast depth-based subgraph kernels for unattributed graphs. Pattern Recognit. 2016, 50, 233–245. [Google Scholar] [CrossRef]
  14. Ketkar, N.; Holder, L.; Cook, D. Empirical Comparison of Graph Classification Algorithms. In Proceedings of the IEEE Symposium Computational Intelligence and Data Mining (CIDM), Nashville, TN, USA, 30 March–2 April 2009; pp. 259–266. [Google Scholar]
  15. Inokuchi, A.; Washio, T.; Motoda, H. An Apriori-Based Algorithm for Mining Frequent Substructures from Graph Data. In Proceedings of the Fourth European Conference on Principles of Data Mining and Knowledge Discovery (PKDD), Lyon, France, 13–16 September 2000; pp. 13–23. [Google Scholar]
  16. Nijssen, S.; Kok, J. A Quickstart in Frequent Structure Mining can Make a Difference. In Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), Seattle, WA, USA, 22–25 August 2004; pp. 647–652. [Google Scholar]
  17. Saigo, H.; Nowozin, S.; Kadowaki, T.; Kudo, T.; Tsuda, K. gBoost: A Mathematical Programming Approach to Graph Classification and Regression. Mach. Learn. 2009, 75, 69–89. [Google Scholar] [CrossRef]
  18. Yan, X.; Han, J. gSpan: Graph-Based Substructure Pattern Mining. In Proceedings of the IEEE International Conference on Data Mining (ICDM), Maebashi, Japan, 9–12 December 2002; pp. 721–724. [Google Scholar]
  19. Zhao, Y.; Kong, X.; Yu, P.S. Positive and Unlabeled Learning for Graph Classification. In Proceedings of the IEEE 11th International Conference on Data Mining (ICDM), Vancouver, BC, Canada, 11–14 December 2011; pp. 962–971. [Google Scholar]
  20. Kong, X.; Yu, P. Multi-Label Feature Selection for Graph Classification. In Proceedings of the 10th International Conference on Data Mining (ICDM), Sydney, Australia, 13–17 December 2010; pp. 274–283. [Google Scholar]
  21. Wu, J.; Huang, S.; Zhou, Z. Genome-wide protein function prediction through multi-instance multi-label learning. IEEE/ACM Trans. Comput. Biol. Bioinform. 2014, 11, 891–902. [Google Scholar] [CrossRef] [PubMed]
  22. Zhou, Z.; Zhang, M. Multi-instance multi-label learning with application to scene classification. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2007; pp. 1609–1616. [Google Scholar]
  23. Li, Y.; Ji, S.; Kumar, S.; Ye, J.; Zhou, Z. Drosophila gene expression pattern annotation through multi-instance multi-label learning. IEEE/ACM Trans. Comput. Biol. Bioinform. 2012, 9, 98–112. [Google Scholar] [PubMed]
  24. Yin, Y.; Zhao, Y.; Li, C.; Zhang, B. Improving Multi-Instance Multi-Label Learning by Extreme Learning Machine. Appl. Sci. 2016, 6, 160. [Google Scholar] [CrossRef]
  25. Martino, G.; Navarin, N.; Sperduti, A. Graph Kernels exploiting Weisfeiler-Lehman Graph Isomorphism Test Extensions. Neural Inf. Process. 2014, 8835, 93–100. [Google Scholar]
  26. Seeland, M.; Kramer, A.K.S. A Structural Cluster Kernel for Learning on Graphs. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), Beijing, China, 12–16 August 2012; pp. 516–524. [Google Scholar]
  27. Riesen, K.; Bunke, H. Graph Classification by Means of Lipschitz Embedding. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2009, 39, 1472–1483. [Google Scholar] [CrossRef] [PubMed]
  28. Hidaka, S.; Tisi, M.; Cabot, J.; Hu, Z. Feature-based classification of bidirectional transformation approaches. Softw. Syst. Model. 2016, 15, 907–928. [Google Scholar] [CrossRef]
  29. Deshpande, M.; Kuramochi, M.; Wale, N.; Karypis, G. Frequent Substructure-Based Approaches for Classifying Chemical Compounds. IEEE Trans. Knowl. Data Eng. 2005, 17, 1036–1050. [Google Scholar] [CrossRef]
  30. Maruping, L.; Venkatesh, V.; Brown, S. Going beyond intention: Integrating behavioral expectation into the unified theory of acceptance and use of technology. J. Assoc. Inf. Sci. Technol. 2017, 68, 623–637. [Google Scholar] [CrossRef]
  31. Balankin, A.; Mena, B.; Cruz, M. Topological Hausdorff dimension and geodesic metric of critical percolation cluster in two dimensions. Phys. Lett. A 2017, 381, 2665–2672. [Google Scholar] [CrossRef]
  32. Winn, J.; Criminisi, A.; Minka, T. Object categorization by learned universal visual dictionary. In Proceedings of the Tenth IEEE International Conference on Computer Vision, Beijing, China, 17–21 October 2005; Volume 2, pp. 1800–1807. [Google Scholar]
  33. Zhou, Z.; Zhang, M.; Huang, S.; Li, Y. Multi-instance multilabel learning. Artif. Intell. 2012, 176, 2291–2320. [Google Scholar] [CrossRef]
  34. Duygulu, P.; Barnard, K.; Freitas, J.; Forsyth, D. Object recognition as machine translation: Learning a lexicon for a fixed image vocabulary. In Proceedings of the 7th European Conference on Computer Vision, Copenhagen, Denmark, 28–31 May 2002; pp. 97–112. [Google Scholar]
  35. Nowozin, S.; Tsuda, K.; Uno, T.; Kudo, T.; Bakir, G. Weighted Substructure Mining for Image Analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
Figure 1. An example figure with structure Graph. (a) Original image; (b) Represented in Graph.
Figure 1. An example figure with structure Graph. (a) Original image; (b) Represented in Graph.
Entropy 20 00245 g001
Figure 2. An example figure with Multi-Label. (a) Original image; (b) Segmented and labeled.
Figure 2. An example figure with Multi-Label. (a) Original image; (b) Segmented and labeled.
Entropy 20 00245 g002
Figure 3. A graph dataset with class label.
Figure 3. A graph dataset with class label.
Entropy 20 00245 g003
Figure 4. An example of MGML (1). (a) Original images and labels; (b) Segmented to multiple graphs; (c) Informative subgraphs.
Figure 4. An example of MGML (1). (a) Original images and labels; (b) Segmented to multiple graphs; (c) Informative subgraphs.
Entropy 20 00245 g004
Figure 5. An example of MGML (2). (a) Multiple instances; (b) Relation between informative subgraphs and labels.
Figure 5. An example of MGML (2). (a) Multiple instances; (b) Relation between informative subgraphs and labels.
Entropy 20 00245 g005
Figure 6. An example of MGML (3). (a) Unlabeled image; (b) Segmented to multiple graphs.
Figure 6. An example of MGML (3). (a) Unlabeled image; (b) Segmented to multiple graphs.
Entropy 20 00245 g006
Figure 7. Results of efficiency experiments. (a) MSRC v2 Dataset; (b) Scenes Dataset; (c) Corel5K Dataset.
Figure 7. Results of efficiency experiments. (a) MSRC v2 Dataset; (b) Scenes Dataset; (c) Corel5K Dataset.
Entropy 20 00245 g007
Table 1. The summary of datasets.
Table 1. The summary of datasets.
Dataset# of Bags# of LabelsLabels Per Bag
MSRC v2591232.5
Scenes200051.2
Corel5K50002603.5
Table 2. MSRC v2 Dataset.
Table 2. MSRC v2 Dataset.
MSRC v2Evaluation Criterion
RankingLossOneErrorCoverageAverage Precision
MGML-ELMhn = 500.0700790.1830393.928240.809013
hn = 1000.0713670.1725893.9289340.820388
hn = 1500.0751820.197973.9898480.804771
hn = 2000.071810.1878173.868020.808192
MGML-SVMCost = 10.1546640.197977.355330.754622
Cost = 20.1711890.2290667.1228440.761665
Cost = 30.1833570.2335037.6345180.735131
Cost = 40.1402840.2195577.6283520.735253
Cost = 50.1373610.1878176.1878170.762115
MIML-SVMCost = 10.1055810.2950735.2674760.710714
Cost = 20.1042090.2922045.2182230.715079
Cost = 30.1009980.2539955.0445840.721987
Cost = 40.0975870.2477754.8904710.73787
Cost = 50.09550.2406824.8808970.745322
Table 3. Scenes Dataset.
Table 3. Scenes Dataset.
SceneEvaluation Criterion
RankingLossOneErrorCoverageAverage Precision
MGML-ELMhn = 500.169270.3181.7710.798919
hn = 1000.1728330.331.810.798367
hn = 1500.1650.3041.780.811867
hn = 2000.1606670.3121.7620.8102
MGML-SVMCost = 10.2996670.8061.3240.555683
Cost = 20.2988350.4341.3240.694689
Cost = 30.29350.341.2840.7401
Cost = 40.2529330.361.3120.630968
Cost = 50.2371670.4581.0620.71515
MIML-SVMCost = 10.9102050.9508153.8106870.242251
Cost = 20.910730.951783.8446750.242479
Cost = 30.913480.9561723.8649880.245989
Cost = 40.9143780.9574713.866760.246726
Cost = 50.9179390.9583223.8679070.249013
Table 4. Corel5K Dataset.
Table 4. Corel5K Dataset.
Corel5KEvaluation Criterion
RankingLossOneErrorCoverageAverage Precision
MGML-ELMhn = 500.2024930.750168113.89680.224968
hn = 1000.1971030.743487113.3547090.224146
hn = 1500.215840.755511120.5490980.219783
hn = 2000.2054240.751503119.3066130.225752
MGML-SVMCost = 10.301240.857229139.00050.120073
Cost = 20.3012640.86724140.97560.121792
Cost = 30.3019060.868129141.80670.122804
Cost = 40.3048380.870358143.31420.123633
Cost = 50.3078720.880693144.86870.128766
MIML-SVMCost = 10.1918670.768118110.42070.217195
Cost = 20.1918990.768204110.43220.217209
Cost = 30.1919220.768299110.46570.217219
Cost = 40.1919780.768416110.47190.217285
Cost = 50.1919970.768899110.51870.217299

Share and Cite

MDPI and ACS Style

Zhu, Z.; Zhao, Y. Multi-Graph Multi-Label Learning Based on Entropy. Entropy 2018, 20, 245. https://doi.org/10.3390/e20040245

AMA Style

Zhu Z, Zhao Y. Multi-Graph Multi-Label Learning Based on Entropy. Entropy. 2018; 20(4):245. https://doi.org/10.3390/e20040245

Chicago/Turabian Style

Zhu, Zixuan, and Yuhai Zhao. 2018. "Multi-Graph Multi-Label Learning Based on Entropy" Entropy 20, no. 4: 245. https://doi.org/10.3390/e20040245

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop