[go: up one dir, main page]

Next Article in Journal
Gearbox Fault Identification Model Using an Adaptive Noise Canceling Technique, Heterogeneous Feature Extraction, and Distance Ratio Principal Component Analysis
Previous Article in Journal
Deep Learning-Based Dynamic Computation Task Offloading for Mobile Edge Computing Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MobilePrune: Neural Network Compression via 0 Sparse Group Lasso on the Mobile System

1
Department of Computer Science, Purdue University, West Lafayette, IN 47907, USA
2
Department of Computer Science, Indiana University at Bloomington, Bloomington, IN 47405, USA
3
Department of Computer Graphics, Purdue University, West Lafayette, IN 47907, USA
4
Institute for Interdisciplinary Information Sciences, Tsinghua University, Beijing 100190, China
5
Institute for Artificial Intelligence, Peking University, Beijing 100871, China
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(11), 4081; https://doi.org/10.3390/s22114081
Submission received: 18 March 2022 / Revised: 19 May 2022 / Accepted: 25 May 2022 / Published: 27 May 2022
(This article belongs to the Section Physical Sensors)

Abstract

:
It is hard to directly deploy deep learning models on today’s smartphones due to the substantial computational costs introduced by millions of parameters. To compress the model, we develop an 0 -based sparse group lasso model called MobilePrune which can generate extremely compact neural network models for both desktop and mobile platforms. We adopt group lasso penalty to enforce sparsity at the group level to benefit General Matrix Multiply (GEMM) and develop the very first algorithm that can optimize the 0 norm in an exact manner and achieve the global convergence guarantee in the deep learning context. MobilePrune also allows complicated group structures to be applied on the group penalty (i.e., trees and overlapping groups) to suit DNN models with more complex architectures. Empirically, we observe the substantial reduction of compression ratio and computational costs for various popular deep learning models on multiple benchmark datasets compared to the state-of-the-art methods. More importantly, the compression models are deployed on the android system to confirm that our approach is able to achieve less response delay and battery consumption on mobile phones.

1. Introduction

Deep neural networks (DNNs) have achieved tremendous success in many real-world applications. However, the computational cost of DNN models significantly restricts their deployment on platforms with limited computational resources, such as mobile devices. To address this challenge, numerous model compression algorithms have been proposed to reduce the sizes of DNN models. The most popular solution is to prune the weights with small magnitudes by adding 0 or 1 penalties [1,2,3,4]. The non-zero weights selected by these methods are randomly distributed and do not reduce the memory consumption due to the matrix operations widely adopted in nowadays deep learning architectures as shown in Figure 1(b.2). The implementation of such a non-structured sparse matrix in cuDNN [5], which is the Basic Linear Algebra Subroutines library used by deep learning models, has similar memory consumption as the original matrix without pruning, as shown in Figure 1(b.2),(c.2). To overcome the problem, structured pruning models [6,7,8,9] are proposed to enforce group sparsity by pruning a group of pre-defined variables together. By tying the weights connecting to the same neuron together, these approaches are able to prune a number of hidden neurons to reduce the sizes of weight matrices and benefit General Matrix Multiply (GEMM) used in cuDNN as shown in Figure 1(b.3),(c.3). However, one of the main problems of the structured methods is that they do not consider and take advantage of the hardware accelerator architectures.
In this paper, we observe three key factors that could lead to an extremely compact deep network. First, we observe that most modern deep learning architectures rely on the General Matrix Multiply (GEMM) functions implemented in the cuDNN package. We, therefore, propose a new network compression algorithm to harmonize the group selections in the structured penalty and the implementation of GEMM in cuDNN as well as the hardware accelerator using sparse systolic tensor array [10,11]. Figure 1 demonstrates the basic rationale of our observation. In comparison to the pruned model in Figure 1(c.3),(c.4) needs additional sparsity within the remaining groups, which could be utilized by the sparse systolic tensor array for hardware acceleration.
The second observation is that recent studies [12,13] have demonstrated that 0 norm is the best sparsity-inducing penalty compared to lasso 1 [14], elastic net [15], SCAD [16], and MCP [17] models. Remarkably, even current 0 optimization techniques can not achieve the global optimal solution, these 0 -based methods with sub-optimal solutions still significantly outperform other sparsity norms, which can be solved to global optima [13,18]. Hence, the community has shown great interest in using 0 norm to compress the large-scale DNN models [8,19,20,21,22,23,24]. However, some 0 -based DNN pruning approaches [8,25,26] rely on different relaxation strategies to conquer the non-convex and non-differentiable challenges, which does not fully exploit the strength of 0 regularization. The other methods [19,20,21,22,23] rely on the alternating direction method of multipliers (ADMM) whose global convergence has not been proved so far when applied to 0 norm optimization [27,28,29].
Third, most of these algorithms are designed originally for mobile platforms, as the computational resources are relatively rich for desktop applications. However, few of them have been deployed on real mobile systems to test the running time and energy consumption to verify their assumptions.
Therefore, we develop the very first algorithm in this paper, named MobilePrune, which is able to solve the optimization of 0 sparse group lasso regularization in an exact manner. The main technical contribution is that we solve the proximal operators for 0 sparse group lasso, where groups in the group lasso term could have overlapping group structure and tree structure [30]. From the theoretical point of view, we prove our algorithm always converges to the critical point (local optimal point) under mild conditions. In addition, we conduct extensive experiments on multiple public datasets and find MobilePrune achieves superior performance at sparsifying networks with both fully connected and convolutional layers. More importantly, we deploy our system on the real android system on several mobile devices and test the performance of the algorithm on multiple Human Activity Recognition (HAR) tasks. The results show that MobilePrune achieves much lower energy consumption and higher pruning rate while still retaining high prediction accuracy. Besides a powerful network compression algorithm, this work also provides a valuable platform and mobile dataset for further work to evaluate their methods in a very real scenario.
The rest of this paper is organized as follows. Section 2 provides the relevant background and related work. In Section 3, we give a brief overview of the proposed MobilePrune methods. In Section 4, we discuss the detailed information of the proposed methods and algorithms. In Section 5, we describe how the experiments are set up and evaluate the proposed methods from different perspectives. Section 6 discusses the future work and summarizes the paper.

2. Related Work

2.1. Sparsity for Deep Learning Models

Many pruning algorithms for deep learning models achieve slim neural networks by introducing sparse-inducing norms to the models. 1 regularization [31,32,33] and 0 regularization [8] were applied to induce sparsity on each individual weight. However, such individual sparsity has arbitrary structures, which cannot be utilized by software and hardware. Wei et al. [24] applied group sparsity to prune filters or channels, which can reduce the matrix size used in GEMM in cuDNN. Because the pruned models are compatible with cuDNN, they achieved large speedups. There are methods [34,35] aiming to find sparse models at both individual and group levels, which is similar to our goal. However, they all used 1 norm to induce individual sparsity in addition to group sparsity. We have performed a comprehensive comparison and demonstrated that our MobilePrune method is the best in inducing sparsity at both individual and group levels for pruning deep learning models in Section 5.

2.2. Learning Algorithms for 0 Norm

Recent studies [12,13] demonstrate that 0 norm is the best sparsity-inducing penalty comparing to lasso 1 [14], elastic net [15], SCAD [16], and MCP [17] models. Remarkably, even current 0 optimization techniques cannot achieve the global optimal solution, these 0 -based methods with sub-optimal solutions still significantly outperform 1 and elastic net models, which can be solved to global optima [13,18]. Therefore, the machine learning community has shown great interest in using 0 norm to compress the large-scale DNN models [8,19,20,21,22,23,24]. However, some 0 -based DNN pruning approaches [8,25] rely on different relaxation strategies to conquer the non-convex and non-differentiable challenges, which does not fully exploit the strength of 0 regularization. The other methods [19,20,21,22,23] rely on the alternating direction method of multipliers (ADMM) whose global convergence has not proved so far when applied to 0 norm optimization [27,28,29].

2.3. Software & Hardware Compatibility

In this paper, we aim to design an algorithm to make the pruned DNN models compatible with cuDNN library [5] and hardware accelerator architecture that uses the sparse systolic tensor array [11,36]. cuDNN is the GPU-accelerated library used in deep learning models [5]. As shown in Figure 1a, convolution used in the convolutional neural network is lowered to matrix multiplication. Therefore, the size of the filter matrix can be reduced when inducing group sparsity column-wise as shown in Figure 1(b.3),(b.4), which can reduce the memory of the DNN models to achieve practical performance improvement. The systolic tensor array is an efficient hardware accelerator for structured sparse matrix as shown in Figure 1(c.4). Specifically, each column in Figure 1(c.4) is sparse. To achieve a pruned DNN model that is compatible with cuDNN and the systolic tensor array, sparsity needs to be induced on both the group level and within-group level. We will show how we achieve this in the following sections.

3. Overview

The central idea of MobilePrune is to compress the deep learning model in a way that is compatible with the architecture of data organization in the memory by combining 0 regularization and group lasso regularization. The group lasso regularization helps to keep important groups of weights that benefit cuDNN, while 0 regularization helps to achieve additional sparsity within those important groups that are needed for hardware acceleration. Figure 2 provides an overview of the proposed MobilePrune method. As illustrated in Figure 2a–c, the group lasso penalty will remove all the weights together with the ith neuron if it is less important for the prediction and if a group is selected, the 0 penalty further removes the weights with small magnitudes within the group. Note that zeroing out the weights connected to the ith neuron results in removing the ith neuron and all the associated weights entirely. We will discuss more detail information in next section.

4. Methods

Our main objective is to obtain a sparse deep neural network with a significantly less number of parameters at both individual and group levels by using the proposed novel combined regularizer: 0 sparse group lasso.

4.1. 0 Sparse Group Lasso

We aim to prune a generic (deep) neural network, which includes fully connected feed-forward networks (FCN) and convolutional neural networks (CNN). Assuming the generic neural network has N neurons in FCN and M channels in CNN. Let W i denote the outgoing weights of the ith neuron in FCN and T j represent the 3D tensor of all filters in the jth channel, which can come from different layers. The training objective for the neural network is given as follows:
min W , T : L ( W , T ; D ) + Ω λ η W + Γ β , γ α T
where W = { W 1 , , W N } , T = { T 1 , , T M } , D = { x i , y i } i = 1 P is a training dataset with P instances, L is an arbitrary loss function parameteized by W and T, Ω λ η ( W ) and Γ β , γ α T represent the 0 sparse group lasso penalties for neurons and channels, respectively. Specifically, Ω λ η ( W ) is defined as
Ω λ η ( W ) = i = 1 N ( η W i 0 + λ W i g )
where η 0 and λ 0 are regularization parameters. Let n ( i ) represent the set of outgoing edge weights of neuron i. Then, W i 0 = j n ( i ) W j i 0 ( W j i is the jth outgoing edge weight of neuron i; W j i 0 = 0 when W j i = 0 and W j i 0 = 1 otherwise) computes the number of the non-zero edges in W i and W i g = j ( W j n ( i ) i ) 2 aggregates the weights associated with the ith neuron as a group. The core spirit of Equation (2) is illustrated in Figure 2a–c, the group lasso penalty W i g tends to remove all the weights together with the ith neuron if it is less important. If a group is selected, the 0 penalty further removes the weights with small magnitudes within the group. Group sparsity W i g can help remove neurons in the neural network, which reduces the size of the neural network and further improve efficiency. Individual sparsity W i 0 helps to achieve additional sparsity within the remaining neurons. Such structured sparsity can be used by the systolic tensor array [10,11]. The other regularization term Γ β , γ α ( T ) is defined as following,
Γ β , γ α ( T ) = j = 1 M ( α T j 0 + β T j g + γ h , w T : , h , w j g ) ,
where α , β , and γ are non-negative regularization parameters. Equation (3) defines a hierarchical-structured sparse penalty in which structure is guided by the memory organization of GEMM using cuDNN [5]. As demonstrated in Figure 2d, the pruning strategy encoded in Equation (3) explicitly takes advantage of the GEMM used in cuDNN. T j g enforces the group sparsity of all the filters applied to the jth channel and T : , h , w j g enforces the group sparsity across the filters on the same channel and T j 0 = f h w T f , h , w 0 prunes the small weights within the remaining channels and filters. Equation (3) can help to achieve an extremely compact model as Figure 1(c.4). Therefore, the computation can be accelerated at both software and hardware levels.

4.2. Exact Optimization by PALM

In this subsection, we first briefly review the general PALM (Proximal Alternating Linearized Minimization) framework used in our MobilePrune algorithm. Then, we introduce how we modify PALM to efficiently optimize the 0 sparse group lasso for neural network compression. PALM is designed to optimize a general optimization problem formulated as:
min W , T : F ( W , T ) + Φ 1 ( W ) + Φ 2 ( T ) .
where F ( W , T ) is a smooth function and Φ 1 ( W ) and Φ 2 ( T ) do not need to be convex or smooth, but are required to be lower semi-continuous. The PALM algorithm applies proximal forward–backward algorithm [37] to optimize Equation (4) with respect to W and T in an alternative manner. Specifically, at iteration k + 1 , the temporal values of W ( k + 1 ) and T ( k + 1 ) for the proximal forward–backward mapping are derived by solving the following sub-problems,
W ( k + 1 ) min W : c k 2 W U i ( k ) F 2 + Φ 1 ( W ) ,
T ( k + 1 ) min T : d k 2 T V i ( k ) F 2 + Φ 2 ( T ) ,
where U i ( k + 1 ) = W i ( k ) 1 c k W F ( W ( k ) , T ( k ) ) and V i ( k + 1 ) = T i ( k ) 1 d k T F ( W ( k + 1 ) , T ( k ) ) . Additionally, c k and d k are positive constants. This optimization process has been proven to converge to a critical point when functions F, Φ 1 , and Φ 2 are bounded [37]. We further extend the convergence proof in [37] and prove that the global convergence of PALM holds for training deep learning models under mild conditions. The detailed proof can be found in Appendix A.
To optimize Equation (1), we define two proximal operators for the two penalty terms as the following,
π λ η ( y ) arg min x 1 2 x y 2 2 + Ω λ η ( x ) ,
θ β , γ α ( y ) arg min x 1 2 x y 2 2 + Γ β , γ α ( x ) .
Here, functions Ω λ η ( · ) and Γ β , λ α ( · ) take vectors as inputs, which are equivalent to Equations (2) and (3) once we vectorize W i and T j . The overall optimization process of MobilePrune is described in Algorithm 1. Once we can efficiently compute the optimal solution of π λ η ( · ) and θ β , λ α ( · ) , the computational burden mainly concentrates on the partial derivative calculation of functions H ( · ) and F ( · ) , which is the same as training a normal DNN model.
Algorithm 1 The framework of MobilePrune Algorithm.
  • Initialize ( W i ) 0 , i , ( T j ) 0 , j and L r .
  • for k = 0 , 2 , do
  •    for  i = 1 , 2 , , N  do
  •       H k i = W k i L W k i , W k j i , T k .
  •       W k + 1 i π λ / L r η / L r W k i 1 L r H k i by solving Equation (7).
  •    end for
  •    for  j = 1 , 2 , , M  do
  •       F k j = T k j L ( W k + 1 , T k j , T k l j ) .
  •       T k + 1 j θ β / L r , γ / L r α / L r T k j 1 L r F k j by solving Equation (8).
  •    end for
  • end for

4.3. Efficient Computation of Proximal Operators

To the best of our knowledge, π λ η ( y ) and θ β , γ α ( y ) defined in Equations (7) and (8) are novel proximal operators that have not been attempted before. Solving π λ η ( y ) and θ β , γ α ( y ) is the key to apply MobilePrune in Algorithm 1. Therefore, this subsection elaborates the algorithmic contributions we made to efficiently calculate π λ η ( y ) and θ β , γ α ( y ) .

4.3.1. Proximal Operator π λ η ( · )

The difficulty of solving π λ η ( y ) in Equation (9) is that both x g and x 0 are not differentiable when the vector x = 0 . Furthermore, x 0 calculates the number of non-zeros in the vector x R n and there are C ( n , 0 ) + C ( n , 1 ) + + C ( n , n ) = 2 n ( C ( n , k ) computes the number of non-zero patterns in x where k elements in x are not zeros) possible combinations, which indicates the brute-force method needs 2 n computations to find the global optimal solution. However, here we prove that the π λ η ( y ) in Equation (9) can be efficiently solved by a O ( n log ( n ) ) algorithm in a closed from. We illustrate the algorithm in Algorithm 2 and prove its correctness in Theorem 1. To the best of our knowledge, it is the first efficient algorithm that can calculate this novel proximal operator.
Theorem 1.
The proximal operator π λ η ( y ) can be written as
π λ η ( y ) arg min x f ( x ) : = 1 2 x y 2 2 + λ x g + η x 0
The optimal solution of this proximal operator can be computed by Algorithm 2.
Algorithm 2 Efficient calculation of π λ η ( y )
  • Input: A sorted vector y, such that | y 1 | | y 2 |
  • Output: x *
  • for i = 0 , , n do
  •    if  y i g λ  then
  •       U i = 1 2 y 2 2
  •    else
  •       U i = 1 2 ( y i g λ ) 2 + i η + 1 2 y 2 2
  •    end if
  • end for
  • Compute k = arg min j U j
  • if U k 1 2 y 2 2 then
  •     x * = 0
  • else
  •     x * = ( y k g λ ) y k y k g
  • end if
Proof. 
Without loss of generality, we assume y = y 1 , y 2 , , y n T R n is an ordered vector, where | y 1 | | y 2 | | y n | . Then, we define y k = [ y 1 , y 2 , , y k , 0 , , 0 ] , where the top k elements with largest absolute values are kept and all the rest elements are set to zeroes. We define another set Φ k = { x | x 0 = k , x R n } to represent all n-dimensional vectors with exact k non-zero elements. For any x = [ x 1 , x 2 , , x n ] T Φ k , we further define a mask function e : e i = 1 { x i = 0 } to reveal the non-zero locations of x.
Since we do not know how many non-zero elements remain in the optimal solution of Equation (9), we need to enumerate all possible k and solve n + 1 sub-problems for all x k Φ k with k = 0 , , n . For each k, the sub-problem is defined as
min x k Φ k f ( x k ) : = 1 2 x k y 2 2 + Ω λ η ( x k ) min x k Φ k 1 2 x k y k 2 2 + 1 2 y k ¯ 2 2 + Ω λ η ( x k )
Based on Lemma A4 in Appendix A.2, we observe that if y k g λ , then x k * = 0 and f ( x k * ) = 1 2 y 2 2 . If  y k g > λ , then x k * = ( y k g λ ) y k y k g and the value of the objective function can be computed as
f ( x k * ) = 1 2 x k * y k 2 2 + 1 2 y k ¯ 2 2 + Ω λ η ( x k * ) = 1 2 ( y k g λ ) 2 + η k + 1 2 y 2 2 .
Equation (11) tells us f ( x k * ) is a function of y k . The task of calculating the minimum value of function f ( x k * ) is transformed into solving another optimization 1 2 ( y k g λ ) 2 + η k + 1 2 y 2 2 with respect to y k , which is equivalent to ask which of the k components of y can achieve the minimum value of f. Since we assume y k g > λ , the optimal solution is clearly to select the top k components with the largest value from y. Therefore we have y k = arg min y k 1 2 ( y k g λ ) 2 + η k + 1 2 y 2 2 and the corresponding x k * = ( y k g λ ) y k y k g . Furthermore, the objective function value is 1 2 ( y k g λ ) 2 + η k + 1 2 y 2 2 . Hence, problem (10) has a closed-form solution.    □
As shown in Algorithm 2, the heaviest computation is to sort the input vector y, therefore, the time complexity for solving Equation (9) is O ( n log ( n ) ) .

4.3.2. Proximal Operator θ β , γ α ( y )

Similar as π λ η ( y ) , θ β , γ α ( y ) is the solution of the following optimization problem:
min x : κ ( x ) : = 1 2 x y 2 2 + α x 0 + β x g + γ i = 1 d x G i g s . t . i = 1 d G i = { 1 , 2 , , n } , G i G j = , i , j
where we assume x , y R n . G i { 1 , , n } , i is the index of a group and d represents the number of groups. Note that the grouping structures specified in Equation (12) is a special case of the grouped tree structures [30], where x g is the group lasso for the root of the tree and all the x G i g are the group lasso terms of its children. Notice that groups from the same depth on the tree do not overlap and furthermore x 0 = i = 1 d x G i 0 . To simplify the notation, assuming x 0 0 we define h ( x ) = 1 2 x y 2 2 + β x g that is a convex and differentiable and rewrite the problem as
min x : h ( x ) + i = 1 d Ω γ α ( x G i ) s . t . x 0 , i = 1 d G i = { 1 , 2 , , n } , G i G j = , i , j { 1 , , d } ,
where i = 1 d Ω γ α ( x G i ) = i = 1 d α x G i 0 + γ x G i g . We can use the proximal method [38] to find a solution x of Equation (13). In the proximal method, we need to estimate the Lipschitz constant L ( x ) = 1 + β x g and the partial derivative x G i h ( x ) = ( 1 + β x g ) x G i y G i . In addition, we need to use Algorithm 2 to solve π γ α ( · ) for each group x G i . After obtaining x , we can find the solution of Equation (12) by comparing κ ( 0 ) with κ ( x ) . If κ ( 0 ) κ ( x ) , then the local optimal solution of Equation (12) is x * = 0 , otherwise, x * = x . We elaborate the algorithm for the proximal operator θ β , γ α ( y ) in the Algorithm 3. The major computation cost is the proximal method, therefore, the convergence rate of Algorithm 3 is O ( 1 / k ) [38].
Algorithm 3 Efficient calculation of θ β , γ α ( y )
  • Input: L 0 and x 0
  • Output: x *
  • for l = 0 , 1 , 2 , , k do
  •    Let L l = 1 + β x l g and u = x G i l 1 L l x G i h ( x l ) , then compute x G i l π γ / L l α / L l ( u ) , i by applying Algorithm 2.
  • end for
  • if κ ( 0 ) κ ( x k ) then
  •     x * = 0
  • else
  •     x * = x k
  • end if

5. Experimental Setup and Results

5.1. Performance on Image Benchmarks

In this subsection, we compared our proposed MobilePrune approach with other state-of-the-art pruning methods in terms of prune rate, computational costs, and test accuracy. We mainly compared our methods with structured pruning methods because DNN models pruned by non-structure pruning methods could not obtain practical speedup as shown in Figure 1. Notably, we only compared the results that can be reproduced by the source codes provided by the competing methods. First, we briefly summarized their methodology. PF [32] and NN slimming [33] were simple magnitude-based pruning methods based on l 1 norm. BC [9], SBP [7], and VIBNet [39] cast the DNN pruning into probabilistic Bayesian models. C-OBD [40], C-OBS [2], Kron-OBD [40,41], Kron-OBS [2,41], and EigenDamage [42] are Hessian matrix-based methods. 0 norm penalized method [8] and group lasso penalized method [24] are also well-known methods.
In our experiments, we use NVIDIA Corporation as the GPU and the number of cores of the CPU is 12. All the baseline models were trained from scratch via stochastic gradient decent(SGD) with a momentum of 0.9. We trained the networks for 150 epochs on MNIST and 400 epochs on CIFAR-10 and Tiny-ImageNet with an initial learning rate of 0.1 and weight decay of 5 × 10−4. The learning rate is decayed by a factor of 10 at 50, 100 on MNIST and at 100, 200 on CIFAR-10 and Tiny-ImageNet, respectively. The details of hyper-parameters for all experiments are summarized in Appendix B. We also provide the computational efficiency of our methods in Appendix C.

5.1.1. MNIST Dataset

We first applied MobilePrune to prune the LeNet-300-100 and LeNet-5 [7,8,9] models on the MNIST dataset [43]. LeNet-300-100 is a fully-connected neural network model with three layers and 267 K parameters. LeNet-5 is comprised of two [20, 50] convolutional layers and two [800, 500] fully-connected layers with 431K parameters. Here, we compared with the state-of-the-art structured network compression algorithms [7,8,9] in terms of pruned accuracy, remaining parameters, pruned architecture, and FLOPs of the pruned models.
As shown in the top half of the MNIST dataset of Table 1, our model achieves the least number of neurons after pruning the LeNet-300-100 model and the lowest drop of the prediction accuracy 0.01% compared to other methods. More importantly, our pruned model achieves the lowest FLOPs. Note that the architecture of our pruned model is as compact as L0-sep [8], but is extremely sparse with only 5252 weights left. This additional sparsity would be critical when applying hardware acceleration [10,11] to our pruned model.
In addition, we compared with SSL on pruning the first two convolutional layers as done in [24] in Table A2. SSL has the same group lasso penalty term as ours but without 0 norm regularization. As shown, our method decreases the sizes of the filters from 25 and 500 to 16 and 65, respectively, which dramatically lowers the FLOPs. In addition, the non-zero parameters in those remaining filters is very sparse in our model.
The bottom half of the MNIST dataset in Table 1 shows the performance comparison on pruning the LeNet-5 model. The LeNet-5 model pruned by our method achieves the lowest FLOPs (113.50 K) with the smallest predicting accuracy drop 0.01 % . Moreover, our pruned model also has the smallest number of weights (around 2310). In addition, we compared with SSL on pruning the first two convolutional layers as done in [24]. SSL has the same group lasso penalty term as ours, but without 0 norm regularization. More details about SSL can be found in Appendix C.2. As shown, our method decreases the sizes of the filters from 25 and 500 to 16 and 65, respectively, which dramatically lowers the FLOPs. In addition, the non-zero parameters in those remaining filters are very sparse in our model.

5.1.2. CIFAR-10 Dataset

We further evaluated our method on pruning more sophisticated DNN architectures, VGG-like [44] and ResNet-32 [42,45] and widen the network by a factor of 4, on CIFAR-10 [46]. Similarly, we compared with the state-of-the-art structured pruning methods [2,7,32,39,40,41,42] in terms of various metrics. As shown in the middle of Table 1, the pruned VGG-like model obtained by our method achieves the lowest FLOPs with the smallest test accuracy drop. Similar to previous results, our pruned model is able to keep the smallest number of weights in comparison to other methods, the key for hardware acceleration [10,11]. As presented in Table 1, the pruned ResNet-32 model achieved by our method outperforms other pruned models in terms of pruned test accuracy and FLOPs. In addition, in terms of the remaining weights, our pruned model is at the same sparsity level as C-OBD [40] while our pruned accuracy outperforms C-OBD by a large margin.

5.1.3. Tiny-ImageNet Dataset

Besides the experiments on MNIST and CIFAR-10 datasets, we further evaluated the performance of our method on a more complex dataset, Tiny-ImageNet [47], using VGG-19 [48]. Tiny-ImageNet is a subset of the full ImageNet, which consists of 100,000 images for validation. There are 200 classes in Tiny-ImageNet. We compared our method with some state-of-the-art methods [2,33,40,42] in Table 1. As shown in Table 1, the test accuracy of the pruned model derived from our method outperforms all the other methods by a significant margin, about 10%, except EigenDamage. Our proposed method obtains the same-level test accuracy as EigenDamage. However, our method achieves a much sparser DNN model with 1.16 million fewer weights than EigenDamage. Meanwhile, our pruned model achieves lower FLOPs.

5.2. Performance on Human Activity Recognition Benchmarks

To demonstrate the efficacy and effectiveness of our proposed MobilePrune method, we perform a series of compassion studies with other state-of-the-art pruning methods such as 0 norm, 1 norm, 2 norm, group lasso, and 1 sparse group lasso for all three datasets—WISDM [49,50], UCI-HAR [51,52], and PAMAP2 [53,54,55]. We evaluate the pruning accuracy and pruning rate of weights (parameters) and nodes for our proposed MobilePrune approach and all other state-of-the-art pruning methods using the same learning rate (0.0001) and the same number of epochs (150) for all three datasets. The pruning thresholds are 0.015, 0.005, 0.01 for the pruning methods in the WISDM, HCI-HAR, and PAMAP2 datasets, respectively. In addition, we evaluate the computational cost and battery consumption for our proposed method with all other state-of-the-art pruning methods as well. The details of the dataset descriptions and the hyper-parameters for all experiments are summarized in Appendix D.

5.2.1. Performance on the Desktop

We use Google Colab [56] to build a PyTorch backend on the above datasets. The GPU for Google Colab is NVIDIA Tesla K80 and the number of cores of the CPU for Google Colab is 2. As shown in Table 2, if we only use 0 norm penalty or 2 norm penalty, there is no effect on neurons or channels pruning as expected for all three datasets. Similarly, if we only employ group lasso penalty, the pruned model still has more weights or nodes left. For the UCI-HAR dataset, 1 norm penalty and 1 sparse group lasso penalty cannot sparse down the model while for the other two datasets, these two penalties could achieve better sparsity, but cannot be better than MobilePrune approach. There exists a trade-off between the pruned accuracy and the pruning rate. As can be seen in Table 2, our MobilePrune method still has high pruned accuracy even if there are not too many parameters and nodes left. In addition, we compare our MobilePrune method with 1 sparse group lasso penalty. The 0 sparse group lasso model still significantly outperforms the 1 sparse group lasso model in weights and nodes pruning, which demonstrates its superiority in pruning CNN models.
We also calculate the response delay and time saving percentage for all the above methods on the desktop platform. Response delay is the time needed for the desktop to run the pre-trained model after the raw input signal is ready. Here in Table 2, the response delay results are obtained after running 200 input samples. As can be seen in Table 2, MobilePrune could save up to 66.00%, 57.43%, 90.20% on response delay on WISDM, HCI-HAR, and PAMAP2 datasets, respectively.
Overall, if we apply MobilePrune method, the pruned CNN models can still achieve the best sparsity in terms of both neurons (or channel) and weights without loss of performance. Additionally, the results in Table 2 show that our MobilePrune method could achieve 28.03%, 46.83%, 3.72% on weight (parameter) sparsity, and 52.52%, 68.75%, and 10.74% on node sparsity for the WISDM, UCI-HAR, and PAMAP2 datasets, respectively.

5.2.2. Performance of Mobile Phones

We evaluate the computational cost and battery consumption for our proposed MobilePrune approach with all other state-of-the-art pruning methods. In order to obtain the final results about how these models perform on today’s smartphone, we implement an Android Application using Android Studio on Huawei P20 and OnePlus 8 Pro. PyTorch Android API [57] is used here for running trained models on Android devices. Currently, the Android devices only support running machine learning models by using CPU only. For the Huawei P20, the CPU is Cortex-A73. For the OnePlus 8 Pro, it is using Octa-Core as its CPU. We also use the Batterystats tool and the Battery Historian script [58] to test the battery consumption.
Table 3 shows the response delay results and battery usage for our proposed method and all other state-to-the-arts pruning methods. Response delay is the time needed for the smartphone’s system to run the pre-trained model after the raw input signal is ready. Here, in Table 3, the response delay results are obtained after running 200 input samples and the battery consumption results are obtained after running 2000 input samples for each penalty in all three datasets. For the HCI-HAR dataset, our MobilePrune approach could save up to 40.14%, 22.22% on response delay and 34.52%, 19.44% on battery usage for Huawei P20 and OnePlus Pro 8, respectively, while the other pruning methods stay almost the same compared to the uncompressed version. For the WISDM and PAMAP2 datasets, 0 norm penalty, 2 norm penalty, and group lasso penalty cannot sparse down the model, and therefore they cannot provide any savings in both response delay and battery consumption. 1 norm and 1 sparse group lasso methods could provide better time saving and battery consumption saving compared to those three penalties, but they still cannot perform better than the MobilePrune method, which saves 61.94% and 88.15% in response time, and 37.50% and 36.71% in battery consumption for WISDM and PAMAP2 dataset, respectively, on Huawei P20. Additionally, it also saves 52.08% and 77.66% in response time, and 32.35% and 37.93% in battery consumption for WISDM and PAMAP2 dataset, respectively, on OnePlus 8 Pro. Overall, results in Table 3 demonstrate MobilePrune’s superiority in pruning HAR CNN models for battery usage and computational cost on today’s smartphone.

5.3. Ablation Studies

To demonstrate the efficacy and effectiveness of the 0 sparse group lasso penalty, we performed a series of ablation studies on various DNN models. As shown in Table 4, if we only use 0 norm penalty, there is no effect on a neuron or channel pruning as expected. Similarly, if we only employ the group lasso penalty, the pruned model still has more weights left. However, if we apply 0 sparse group lasso, we can achieve pruned DNN models that are sparse in terms of both neurons (or channel) and weights. In addition, we compare our 0 sparse group lasso model with 1 sparse group lasso [59] on pruning DNN models. Table 4 shows their comparison on pruning various DNN models. More details can be found in Appendix A.3 and Appendix C. As shown in Table 4 and the results in supplementary, the 0 sparse group lasso model significantly outperforms the 1 sparse group lasso model in all aspects, which demonstrates its superiority in pruning DNN models.

6. Conclusions

In this work, we proposed a new DNN pruning method MobilePrune, which is able to generate compact DNN models that are compatible with both cuDNN and hardware acceleration. MobilePrune compress DNN models at both group and individual levels by using the novel 0 sparse group lasso regularization. We further developed a global convergent optimization algorithm MobilePrune based on PALM to directly train the proposed compression models without any relaxation or approximation. Furthermore, we developed several efficient algorithms to solve the proximal operators associated with 0 sparse group lasso with different grouping strategies, which is the key computation of our MobilePrune. We have performed empirical evaluations on several public benchmarks. Experimental results show that the proposed compression model outperforms existing state-of-the-art algorithms in terms of computational costs and prediction accuracy. MobilePrune has a great potential to design slim DNN models that can be deployed on dedicated hardware that uses a sparse systolic tensor array. More importantly, we deploy our system on the real android system on both Huawei P20 and OnePlus 8 Pro, and the performance of the algorithm on multiple Human Activity Recognition (HAR) tasks. The results show that MobilePrune achieves much lower energy consumption and higher pruning rate while still retaining high prediction accuracy.
There are other options to further compress the neural network models such as Neural Logic Circuits and Binary Neural Networks, which all use binary variables to represent inputs and hidden neurons. These two models are orthogonal to our methods, which means our pruning model could be adopted on Neural Logic Circuits, Binary Neural Networks and other neural network architectures designed for mobile systems. We will explore which mobile neural network could be better integrated with our network compression model in the future.

Author Contributions

Conceptualization: Y.S., Y.W. and J.M.; Methodology: K.Z, Y.W. and J.M.; Software: Y.S.; Validation: Y.S. and K.Z.; Writing—original draft preparation: Y.S., Z.C., Z.P., Y.W. and J.M.; Writing—review and editing: Y.S., X.P. and K.Z.; Supervision: P.L., Y.W. and J.M.; Project administration: Y.W. and J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. The Convergence Analysis of Applying PALM Algorithm to Deep Learning Models

For simplicity, we prove the convergence of the PALM algorithm on a neural network that only has fully connected layers, and the regularization is added on the weight matrix of each layer. The proof can be easily extended to DNN models with regularization added on each neuron and DNN with convolutional layers.
Given a feed-forward neural network with N 1 hidden layers, there are d i neurons in the i-th layer. Let d 0 and d N represent the number of neurons in the input and output layers, respectively. Therefore, the input data can be presented by X : = { x 1 , , x n } R d 0 × n and the output data can be denoted as Y : = { y 1 , , y n } R d N × n . Let W i R d i × d i 1 be the weight matrix between the ( i 1 )-th layer and the i-th layer. Here, in order to simplify the notation, we let W i absorb the bias of the i-th layer. We denote the collection of W i as W = { W i } i = 1 N . The DNN model training problem can be formulated as
min W : R ( W ; X , Y ) + i r i ( W i ) ,
where R ( W ; X , Y ) = 1 n n ( Φ ( W ; x i ) , y i ) , and ( · ) is the loss function. Φ ( W ; x i ) = σ N ( W N σ N 1 ( . . . W 2 σ 1 ( W 1 x i ) ) ) is the DNN model with N layers of model parameter W and σ i is the activation function for neurons in the i-th layer. r i is the regularization function applied to W i . We make the assumptions for our DNN model as follows:
Assumption A1.
Suppose that the DNN model satisfies the following assumptions:
1. 
The regularization functions r i , i = 1 , , N are lower semi-continuous.
2. 
The derivatives of the loss function ℓ and all activation functions σ i , i = 1 , , N are bounded and Lipschitz continuous.
3. 
The loss function ℓ, activation function σ i , i , and the regularization function r i , i are either real analytic or semi-algebraic [37], and continuous on their domains.
Remark A1.
The DNN model that satisfies the assumptions made in Assumption A1 could have squared, logistic, hinge, cross-entropy, or soft-max loss function and smooth activation functions, such as sigmoid, or hyperbolic tangent, and 1 norm, 2 norm, or 0 norm regularization term, and the assumption requires the activation functions to be smooth. Then, activation function, such as rectified linear unit(ReLU), does not satisfy the requirement. However, we can use the Softplus activation function or Swish activation function to replace ReLU.
Since R ( W ; X , Y ) depends on weight matrices W i , i = 1 , , N between the DNN layers, we rewrite (A1) as the following format in terms of only the independent variables
min W 1 , , W N : Ψ ( W ) : = H W 1 , , W N + i r i ( W i ) ,
where H is exactly the same as R in (A1), but appears in a different form. We propose Algorithm A1 to solve (A2) and prove the convergence via the following theorem. In Algorithm A1, we use the proximal operator prox t σ ( x ) at each step, which is defined as
prox t σ ( x ) = arg min u t 2 u x + σ ( u ) .
Theorem A1.
Suppose Assumption A1 holds. As in [37], the sequence W k = ( W 1 k , , W N k ) generated by Algorithm A1 converges to the critical point of (A2) if the following conditions hold:
1.
Ψ ( W ) is a Kurdyka-Lojasiewicz (KL) function.
2.
The partial gradient W i H , i is Lipschitz continuous and there exist positive constants l i ̲ , l i ¯ such that c i k ( l i ̲ , l i ¯ ) , k = 1 , 2 , .
3.
W H ( W 1 , , W N ) has Lipschitz constant on any bounded set.
Proof. 
For the first condition, utilizing Proposition 1 and Lemma 3–6 in [60], we can easily prove that under our assumption Ψ ( W ) is a KL function. For Equation (1) in the main text, it is also a KL function because 0 and group lasso are semi-algebraic [60,61]. We will add the simple proof in the Appendix.
According to our model assumption and Remark 1 in [62], we can show that Ψ ( W ) is a KL function. For the second condition, based on our model Assumption A1 and Lemma A1 provided in the following, we know that W i H , i is Lipschitz continuous. In addition, in Algorithm A1, we use backtracking strategy to estimate L i at each iteration, therefore, there exist l i ̲ = L i 0 and l i ¯ = L ¯ i , such that c i k ( L i 0 , L ¯ i ) , k = 1 , 2 , .
For the last condition, based on Lemma A3 provided in the following, we have W H ( W 1 , ..., W N ) is Lipschitz continuous for any bounded set.    □
Algorithm A1 PALM Algorithm for Deep Learning Models
  • Initialize η > 1 , L 1 0 , , L N 0 , and  W 1 0 , , W N 0 .
  • for k = 1 , 2 , do
  •    for  i = 1 to N do
  •      Find the smallest i k such that with L i = η i k L i k 1
  •      
    W i H ( W i k 1 ) W i H ( W i k ) L i W i k 1 W i k
  •      Set c i k = η i k L i k 1 and compute
  •      
    W i k = p r o x c i k r i W i k 1 1 c i k W i H ( W i k 1 )
  •    end for
  • end for
Lemma A1.
According to Assumption A1, the derivatives of the loss function and all the activation functions used in function R in (A1) are bounded and Lipschitz continuous, then W i H = W i R is also bounded and Lipschitz continuous.
Proof. 
The partial derivative of W i can be written as
W i H = W i R = Φ · Φ σ N σ i + 1 σ i · σ i W i .
From Assumption A1, we know both the derivatives of the loss function Φ and the derivatives of any activation function σ i + 1 σ i , i are bounded and Lipschitz continuous. Based on Lemma A2 and (A6), we know the multiplication of bounded and Lipschitz continuous functions is still bounded and Lipschitz continuous. Therefore, W i H is bounded and Lipschitz continuous.    □
Lemma A2.
If f i : R N R , i is bounded | f i ( X ) | < M i , i and Lipschitz continuous f i ( X ) f i ( Y ) L i X Y , then their multiplication f 1 ( X ) f 2 ( X ) . . . f n ( X ) is still Lipschitz continuous.
Proof. 
We first prove that the multiplication of two bounded Lipschitz continuous functions is still Lipschitz continuous as follows.
f 1 ( X ) f 2 ( X ) f 1 ( Y ) f 2 ( Y ) = f 1 ( X ) f 2 ( X ) f 1 ( X ) f 2 ( Y ) + f 1 ( X ) f 2 ( Y ) f 1 ( Y ) f 2 ( Y ) f 1 ( X ) f 2 ( X ) f 1 ( X ) f 2 ( Y ) + f 1 ( X ) f 2 ( Y ) f 1 ( Y ) f 2 ( Y ) | f 1 ( X ) | L 2 X Y + | f 2 ( Y ) | L 1 X Y ( M 1 L 2 + M 2 L 1 ) X Y .
We can then extend the above to the multiplication of multiple functions and prove the lemma.    □
Lemma A3.
If W i H , i 1 , , N is bounded and Lipschitz continuous, W H is also Lipschitz continuous.
Proof. 
Lemma A1 has shown that W i H , i 1 , , N is bounded and Lipschitz continuous. Here we want to show that W H is Lipschitz continuous on any bounded set. Define vec ( W ) = [ vec ( W 1 ) T , , vec ( W N ) T ] T as a vector where vec ( ) vectorizes a matrix by stacking its columns. Specifically, there exists a constant M > 0 , such that for any W , W ^
W H W ^ H 2 = vec ( W ) H vec ( W ^ ) H 2 = [ ( vec ( W 1 ) H ) T , , ( vec ( W N ) H ) T ] [ ( vec ( W ^ 1 ) H ) T , , ( vec ( W ^ N ) H ) T ] 2 = i ( vec ( W i ) H ) T ( vec ( W ^ i ) H ) T 2 i L i 2 vec ( W i ) vec ( W ^ i ) 2 L max 2 i vec ( W i ) vec ( W ^ i ) 2 = L max 2 W W ^ 2 ,
where L max is the largest Lipschitz constant among L 1 , , L N . Then, we define M = L max and we have
W H W ^ H M W W ^ ,
as desired.    □
Theorem A2.
The sequence generated by Algorithm A1 ( W 1 k , , W N k ) converges to a critical point ( W 1 * , , W N * ) of Equation (A2) at least in the sub-linear convergence rate, i.e. there exists some ω > 0 , such that
( W 1 k , , W N k ) ( W 1 * , , W N * ) ω k 1 θ 2 θ 1 ,
where θ ( 1 2 , 1 ) .
Proof. 
Based on Theorem 2 in [63] and Theorem 6.4 in [61], we can prove the above theorem.    □

Appendix A.2. The Lemma Used in the Proof of Theorem 1

Lemma A4.
The proximal operator associated with the Euclidean norm · has a closed form solution:
arg min x 1 2 x y 2 + λ x = max ( y λ , 0 ) y y ,
where x R n and y R n are both n-dimensional vectors.
This is a known result and has been used in previous study [64].

Appendix A.3. The Algorithm for ℓ 1 Sparse Group Lasso

In Section 5.3, we conduct experiments regarding 1 norm sparse group Lasso for comparison with our proposed 0 sparse group lasso. Here, we elaborate the 1 norm sparse group Lasso algorithm. We first consider the following proximal operator associated with 1 overlapping group Lasso regularization:
π λ 2 λ 1 ( v ) = arg min x R n g λ 2 λ 1 ( x ) 1 2 x v 2 + λ 1 x 1 + λ 2 i = 1 k x G i ,
where the regularization coefficients, λ 1 and λ 2 , are non-negative values, i = 1 , 2 , , k , G i { 1 , 2 , , n } denotes the indices corresponding to the i-th group. According to Theorem 1 in [59], π λ 2 λ 1 ( · ) can be derived from the following π 0 λ 1 ( · ) and we present this conclusion in Lemma A5.
Lemma A5.
Let u = sgn ( v ) max | v | λ 1 , 0 , and
π λ 2 0 ( u ) = arg min x R n h λ 2 ( x ) 1 2 x u 2 + λ 2 i = 1 g x G i .
Then, π λ 2 λ 1 ( v ) = π λ 2 0 ( u ) holds.
According to Lemma A4, it is easy to verify that given x G i x G j = , the optimal x G i minimizing h λ 2 ( x ) is given by
x G i = max ( u G i λ 2 , 0 ) u G i u G i ,
where i = 1 , 2 , , k . For the fully connected layers, we define the 1 overlapping group Lasso regularizer as
Ψ ( W ) = q = 1 Q ψ λ 2 λ 1 ( W : , q ) = q = 1 Q λ 1 W : , q 1 + λ 2 W : , q = λ 1 W 1 + q = 1 Q λ 2 W : , q ,
where W : , q represents the output weights of the q-th neuron of W. According to Lemma A5, let W ^ = sgn ( W ) max | W | λ 1 , 0 , and then ψ λ 2 λ 1 ( W : , q ) can be reduced to ψ λ 2 0 ( W ^ : , q ) , which can be solved via (A14) as follows.
W : , q = max ( W ^ : , q λ 2 , 0 ) W ^ : , q W ^ : , q
For the convolutional layers, we denote the 1 overlapping group Lasso regularizer as
Φ ( T ) = c = 1 N c ϕ β b β a ( T : , c , : , : ) = c = 1 N c β a T : , c , : , : 1 + β b h = 1 N h w = 1 N w T : , c , h , w = β a T 1 + c = 1 N c h = 1 N h w = 1 N w β b T : , c , h , w ,
Likewise, let T ^ : , c , : , : = sgn ( T : , c , : , : ) max | T : , c , : , : | β a , 0 , and then ϕ β b β a ( T : , c , : , : ) can be reduced to ϕ β b 0 ( T ^ : , c , : , : ) , which can be solved via (A14) as follows.
T : , c , h , w = max ( T ^ : , c , h , w β b , 0 ) T ^ : , c , h , w T ^ : , c , h , w
By incorporating (A15) and (A17) into our model, the 1 norm sparse group Lasso regularized problem can be formulated as
min W : L ( W ; X , Y ) + i = 1 N f q = 1 Q ψ λ 2 λ 1 ( W : , q ( i ) ) + j = 1 N l c = 1 N c ϕ β b β a T : , c , : , : ( j ) .
We elaborate the DNN_PALM algorithm associated with 1 norm sparse group lasso in Algorithm A2, in which the partial derivatives are denoted as H ( W : , q ( i ) ) = W : , q ( i ) L ( W : , q ( i ) ) and F ( T : , c , h , w j ) = T : , c , h , w j L ( T : , c , h , w j ) for fully connected layers and convolutional layers, respectively.
Algorithm A2 DNN_PALM Algorithm for 1 norm Group Lasso
  • Initialize μ > 1 , L i 0 > 0 , ( W ( i ) ) 0 , i and L j 0 > 0 , ( T ( j ) ) 0 , j .
  • for k = 1 , 2 , do
  •    for  i = 1 to N f  do
  •      Let ω k 1 = ( W : , q ( i ) ) k 1 and ω k = ( W : , q ( i ) ) k , q .
  •      Find the smallest L = μ b L i k 1 , b N , such that H ( ω k 1 ) H ( ω k ) L ω k 1 ω k , where ω k is computed via (A16).
  •    end for
  •    for  j = 1 to N l  do
  •      for  c = 1 to N c  do
  •         Let γ k 1 = ( T : , c , h , w j ) k 1 and γ k = ( T : , c , h , w j ) k , h , w .
  •         Find the smallest L = μ b L j k 1 , b N , such that F ( γ k 1 ) F ( γ k ) L γ k 1 γ k , where γ k is computed via (A18).
  •      end for
  •    end for
  • end for

Appendix B

Appendix B.1. Iterative Method

Iterative pruning [4] is another effective method for obtaining a sparse network while maintaining high accuracy. As iterative pruning is orthogonal to our method, we can couple the two methods to obtain even better performance per number of parameters used; specifically, we replace the usual weight decay regularizer used in [4] with our 0 sparse group lasso regularizer. In practice, we find that, empirically, the iterative method is able to achieve better performance. All results reported in the paper are from the iterative method.

Appendix B.2. Hyper-Parameter Settings

In our experiments, all the baseline models were trained from scratch via stochastic gradient decent(SGD) with a momentum of 0.9. We trained the networks for 150 epochs on MNIST and 400 epochs on CIFAR-10 and Tiny-ImageNet with an initial learning rate of 0.1 and weight decay of 5e-4. The learning rate is decayed by a factor of 10 at 50, 100 on MNIST and at 100, 200 on CIFAR-10 and Tiny-ImageNet, respectively.
The experimental settings regarding hyper-parameters for all DNN models we used in the paper are summarized in Table A1. We employ iterative pruning strategy to prune all the models. Namely, the pruning process and the retraining process are performed alternately.
Table A1. List of hyper-parameters and their values(“-” denotes “not applicable”).
Table A1. List of hyper-parameters and their values(“-” denotes “not applicable”).
Hyper-ParameterLeNet300LeNet5VGG-LikeResNet-32VGG-19Description
learning rate1 × 10−31 × 10−31 × 10−31 × 10−31 × 10−3The learning rate used in retraining process
gradient momentum0.90.90.90.90.9The gradient momentum used in retraining process
weight decay1 × 10−41 × 10−55 × 10−41 × 10−41 × 10−4The weight decay factor used in retraining process
minibatch size1 × 1026 × 1021 × 1033 × 1024 × 102The number of training samples over which each SGD update is computed during the retraining process
0 norm factor4 × 10−42 × 10−41 × 10−61 × 10−81 × 10−10The shrinkage coefficient for 0 norm regularization
channel factor-1 × 10−31 × 10−3–1 × 10−2 15 × 10−25 × 10−2The shrinkage coefficient of channels for group Lasso
neuron factor2 × 10−42 × 10−41 × 10−401 × 10−2The shrinkage coefficient of neurons for group Lasso
filter size factor-1 × 10−31 × 10−41 × 10−41 × 10−4The shrinkage coefficient of filter shapes for group Lasso
pruning frequency (epochs/minibatches)1010121No. of epochs(LeNet)/minibatches(VGGNet/ResNet) for pruning before retraining
retraining epochs3030203015The number of retraining epochs after pruning
iterations7410263266The number of iterations for obtaining the final results
1 On VGG-like, the channel factor is adaptive and it is increased by 0.001 if its cross-entropy loss is not greater than the loss before performing pruning for the current mini-batch. Its range is [0.001,0.01].

Appendix C

Appendix C.1. Computational Efficiency

We want to mention that our main focus here is to compress DNN models via 0 sparse group lasso. Our main contribution is to solve the corresponding optimizations. Speed is not our primary focus. However, we still provide the run time of our methods as a reference. We compared the run time of our MobilePrune with the baseline methods (original methods without pruning involved), and we found the ratio of the run time of MobilePrune to the run time of the baseline method is around 5 on average on the same machine.

Appendix C.2. Results about the SSL

Table A2. Results about learning filter shapes in LeNet-5. (We highlight our MobilePrune results).
Table A2. Results about learning filter shapes in LeNet-5. (We highlight our MobilePrune results).
MethodBase/Pruned Accuracy (%)Filter SizeRemaining FiltersRemaining ParametersFLOPs (K)
Baseline-25–50020–50500–25,0002464
SSL [24]99.10/99.007–141–50-63.82
MobilePrune99.12/99.0314–94–1646–2651.21

Appendix C.3. Additional Ablation Studies

In this section, we perform ablation studies to compare DNN models regularized by the proposed 0 sparse group lasso and other DNN models that are regularized by its individual components. Specifically, we compare DNN models regularized by the proposed 0 sparse group lasso with DNN models regularized by 0 norm penalty (set group Lasso penalty to 0) and DNN models regularized by group Lasso penalty (set 0 norm penalty to 0), respectively. For fair comparison, for all regularized DNN models, we use the same hyper-parameter setting.
From Table A3 to Table A6, we observe that 0 norm penalty has no effect on structured pruning as expected and group Lasso penalty can effectively remove redundant structure components. Furthermore, the combination of 0 norm and group Lasso (our proposed 0 sparse group lasso penalty) can yield sparser models at both structure level and individual weight level. Notably, 0 norm can help group Lasso to remove more redundant structure components. Therefore, better acceleration in terms of FLOPs can be obtained by applying our proposed 0 sparse group lasso penalty. We want to mention that when we compute FLOPs, we do not take the individual weight sparsity into account. However, based on [11,65], lower FLOPs could be achieved by unitizing the sparsity on weight level on dedicated architectures.
Table A3. Alation studies on LetNet-5 (Architecture: 20-50-800-500).
Table A3. Alation studies on LetNet-5 (Architecture: 20-50-800-500).
PenaltyBase/Pruned Accuracy (%)Original/Remaining Parameters (K)Pruned ArchitectureFilter SizeFLOPs (K)Sparsity (%)
0 norm99.12/99.20431/321.0020-50-800-50025–5002293.074.48
Group Lasso99.12/99.11431/8.814-19-301-2925–99187.002.04
1 Group Lasso99.12/99.03431/9.984-17-271-8223–99183.832.32
0 sparse group lasso99.12/99.11431/2.315-14-151-5716–65113.501.97
Table A4. Alation studies on VGG-like.
Table A4. Alation studies on VGG-like.
PenaltyBase/Pruned Accuracy (%)Original/Remaining Parameters (Mil)Pruned ArchitectureFLOPs (Mil)
0 norm92.96/93.4015/3.3918-43-92-99-229-240-246-507-504-486-241-114-428-168210.94
Group Lasso92.96/92.4715/0.8417-43-89-99-213-162-93-42-32-28-8-5-429-16878.07
1 Group Lasso92.96/92.9015/0.6117-43-92-99-229-240-246-323-148-111-41-39-159-161134.35
0 sparse group lasso92.96/92.9415/0.6017-43-87-99-201-185-80-37-27-25-9-4-368-16777.83
Table A5. Alation studies on ResNet-32.
Table A5. Alation studies on ResNet-32.
PenaltyBase/Pruned Accuracy (%)Original/Remaining Parameters (Mil)FLOPs (Mil)Sparsity (%)
0 norm95.29/95.687.42/6.74993.1190.84
Group Lasso95.29/95.307.42/3.43393.0945.95
1 Group Lasso95.29/95.047.42/5.66735.1276.28
0 sparse group lasso95.29/95.477.42/2.93371.3039.49
Table A6. Alation studies on VGG19.
Table A6. Alation studies on VGG19.
PenaltyTest Accuracy (%)Remaining Parameters (Mil)Pruned ArchitectureFLOPs (Mil)
Baseline61.5620.1264-64-128-128-256-256-256-256-512-512-512-512-512-512-512-5121592.53
0 norm61.9919.2945-64-114-128-256-256-256-256-512-511-512-509-512-512-512-5121519.23
Group Lasso53.255.9323-61-80-128-122-114-164-253-255-322-412-462-23-93-129-512683.99
1 Group Lasso53.970.2129-64-109-128-254-246-254-256-510-509-509-509-512-512-484-5121282.82
0 sparse group lasso56.274.0519-48-57-102-79-83-100-179-219-273-317-341-256-158-116-512407.37

Appendix C.4. Additional Comparison between ℓ 0 Sparse Group Lasso and ℓ 1 Norm Sparse Group Lasso

In addition, we compare the proposed 0 sparse group lasso with 1 norm group Lasso. The algorithm for DNN models with 1 norm group Lasso penalty is introduced in Algorithm A2. For hyper-parameter setting, we use the same parameters as 0 sparse group lasso penalty (as shown in Table A1) except the parameter for 1 norm. We search the parameter of 1 norm in [ 0.0001 , 0.01 ] and report the best results in terms of the pruned test accuracy.
From Table A3 to Table A6, we find that 0 sparse group lasso penalized models outperform 1 sparse group Lasso penalized models in terms of test accuracy and FLOPs. For VGG19 model (Table A6), 1 sparse group Lasso penalized model can achieve the fewest number of parameters, but the pruned test accuracy and FLOPs are much worse than the 0 sparse group lasso penalized model.

Appendix C.5. The Effect of the Coefficient of ℓ 0 Norm Regularizer

In our proposed 0 sparse group lasso, 0 norm regularizer plays an important role of facilitating pruning networks effectively and efficiently, which has been shown through the results in ablation studies with and without 0 norm penalty. We further explore the effect of the strength of the 0 norm coefficient on the pruning performance. We vary the shrinkage strength for 0 norm penalty by a factor of 10 while keeping the other settings fixed.
As can be seen from Table A7 and Table A8, the larger the 0 norm coefficient is, the more parameters are pruned as expected. Additionally, there is a trade-off between the shrinkage coefficients for 0 norm penalty and group Lasso penalty, which depends on the practical demand.
Table A7. The effect of the coefficient of 0 norm penalty on VGG-like.
Table A7. The effect of the coefficient of 0 norm penalty on VGG-like.
0 Penalty CoefficientBase/Pruned Accuracy (%)Original/Remaining Parameters (Mil)Pruned ArchitectureFLOPs (Mil)
1 × 10−492.96/89.7715/0.0617-43-83-99-161-105-57-28-24-15-11-4-104-15756.43
1 × 10−592.96/92.1915/0.3016-43-85-99-171-155-75-33-23-18-10-3-264-16766.82
1 × 10−692.96/92.9415/0.6017-43-87-99-201-185-80-37-27-25-9-4-368-16777.83
1 × 10−792.96/92.5415/0.7417-43-87-99-213-188-91-40-26-27-9-4-400-16881.64
Table A8. The effect of the coefficient of 0 norm penalty on ResNet-32.
Table A8. The effect of the coefficient of 0 norm penalty on ResNet-32.
0 Penalty CoefficientBase/Pruned Accuracy (%)Original/Remaining Parameters (Mil)FLOPs (Mil)Sparsity
1 × 10−695.29/95.117.42/2.06330.9027.76
1 × 10−795.29/95.337.42/2.72369.3636.66
1 × 10−895.29/95.477.42/2.9377.8339.49
1 × 10−995.29/95.447.42/3.02372.9840.70

Appendix D

Appendix D.1. Har Dataset Description

Appendix D.1.1. Wisdm Dataset

The WISDM dataset [49], publicly available in [50], includes six activities (walking, jogging, walking upstairs, walking downstairs, sitting, and standing) that contain 3D ( x , y , z ) raw signals collected from the smartphone’s accelerometer at a sampling rate of 20Hz. The total number of participants involved in the experiment is 36. These participants performed certain daily activities with an Android phone in their front leg pockets. This dataset has a total of 1,098,209 samples and each sample consists of a timestamp, a user ID, an activity ID, and the acceleration ( x , y , z ) raw data. Here for this dataset, 3 features are used—the gravitational acceleration ( x , y , z ) toward the center of the Earth. A sliding window approach with a window size of 80 readings (4 seconds) is used for segmenting the sequences with a 50% overlapping.

Appendix D.1.2. UCI-HAR Dataset

The UCI-HAR [51] dataset, publicly available in [52], includes six activities (walking, walking upstairs, walking downstairs, sitting, standing, and jogging) that contains 3D ( x , y , z ) raw signals extracted from the accelerometer and gyroscope of a smartphone at a constant rate of 50 Hz strapped to the waist of a subject. These raw signals were applying a noise filter to remove the noise first and then sampled in fixed-width sliding windows of 2.56 s (128 readings). This dataset was collected from a group of 30 volunteers. And all volunteers were instructed to follow an activity protocol and wore a Samsung Galaxy S II smartphone on their waist. The dataset includes a total of 10,299 samples including 7352 training samples (71.39%) and 2947 testing samples (28.61%). The dimension for each sample is 128 readings × number of features with a 50% overlapping. Here for this dataset, 9 features are used—the acceleration signals ( x , y , z ) collected by the smartphone accelerometer in standard gravity units, the body acceleration signals ( x , y , z ) obtained by subtracting the gravity from the total acceleration, and the angular velocity vector ( x , y , z ) measured by the smartphone gyroscope.

Appendix D.1.3. PAMAP2 Dataset

The PAMAP2 dataset [53,54], publicly available in [55], contains data of different physical activities, performed by 9 subjects wearing 3 inertial measurement units and a heart rate monitor with a sampling rate of 100 Hz. According to the dataset’s protocol, there are 12 physical activities—lying, sitting, standing, walking, running, cycling, nordic walking, ascending stairs, descending stairs, vacuum cleaning, and rope jumping. All the collected data above include two 3-axis accelerometer data, 3-axis gyroscope data, 3-axis magnetometer data, 3-axis orientation data, and temperature. This dataset has a total of 3,850,505 samples and each sample has a timestamp, a user id, an activity id, and the corresponding features. Here we pick 40 features listed by the dataset. Similar to the WISDM dataset, a sliding window approach with a window size of 128 readings (1.28 s) is used for segmenting the sequences with a 50% overlapping.

Appendix D.2. 1D CNN Model

In general, sensor data, such as accelerometers and gyroscopes, can be classified as time-series data. We can encode these time-series data as images to allow machines to recognize human behavior virtually. Inspired by the recent successes of deep learning techniques in computer vision, we convert sequence data into image data according to Gramain Angular Field transform algorithm [66] to obtain a one-dimensional convolutional neural network (1D CNN) model. The benefit of using CNNs for sequence classification is that we can learn from the raw time series data directly, and this method does not require domain expertise to manually engineer input features. This 1D CNN model can learn an internal representation of the time series data and could achieve good performance to models that fit on the version of the dataset with engineered features.
Similar to the general CNN model, the convolutional layer that uses the convolution kernel for the input data is the most essential part in CNN. The convolution kernel works as a filter and is activated by a non-linear activation function. In this paper, a sequential model with a PyTorch backend is built via Google Colab [56] and 5 consecutive 1D convolutional layers with 128 neurons each are selected. Each layer uses a ReLU activation function. In order to compress this CNN model, we cannot add a 1D max-pooling because of the random down-sampling. For the WISDM and PAMAP2 datasets, 10 convolution kernels are used for each layer. For the UCI-HAR dataset, 5 convolution kernels are used. All these values are selected based on tons of experiments. In addition, for the sake of efficiency, we select the number of epochs to be 150 for all three datasets during the training stage.

Appendix D.3. Data Pre-Processing

In order to provide a certain data dimension and improve the performance of the proposed 1D CNN Model, the above collected raw data need to be pre-processed as the following.

Appendix D.3.1. Re-Scaling and Standardization

If we use the above datasets’ raw data directly to train our model, the final results may cause training bias because of those large values. In order to remove such bias, standardizing a dataset is necessary. Standardizing a dataset involves re-scaling the distribution of the values for each channel such that the mean is 0 and the standard deviation is 1, as shown in Equation (A20):
X i j = X i j m e a n ( X i ) s t d ( X i )
where i = 1 , 2 , , n and n denotes the number of channels, j = 1 , 2 , , m and m denotes the number of elements in each channel.

Appendix D.3.2. Segmentation

As mentioned above, the input to the model consists of a data sequence extracted from the raw sensor data. The data were recorded continuously in the data collection process. In order to preserve the temporal relationship between the collected data points and their corresponding activity, a sliding window approach is used to segment the collected data points. For the HCI-HAR and PAMAP2 datasets, a fixed-length of 128 sliding windows with an overlapping rate of 50% is applied. For the WISDM dataset, a fixed-length of 80 sliding windows with an overlapping rate of 50% is applied. After segmenting the raw data, we obtain 27,455 samples for the WISDM dataset, 10,299 samples for the UCI-HAR dataset, and 30,356 samples for the PAMAP2 dataset. Next, we select 80% data samples randomly from WISDM and PAMAP2 datasets as the training samples while the remaining 20% data samples are the testing samples. For the UCI-HAR dataset, it is already split.

Appendix D.3.3. K-Fold Cross-Validation

In order to improve the proposed model’s final performance, k-fold cross-validation is used after the above segmentation step. The principle of the k-fold cross-validation method is to split the input samples as the number of k groups. It can lead to a less biased or less positive assessment of the ability of the model than other methods [67]. All the training data samples are considered for both training and validation in a k-fold cross-validation approach. First, we divide the training data samples into k equal subsets. Then, we pick one subset as the validation set and the remaining k 1 subsets as the training set. There are k different ways to select the validation set, and therefore we have k different pairs of testing and validation datasets. In this paper, we choose k = 5 and evaluate all the 5 different pairs of testing and validation datasets for our proposed method and all state-of-the-art pruning methods. The final training and validation dataset will be selected according to the final performance with the testing data set.

Appendix D.4. Hyper-Parameters Tuning

Hyper-parameters have a great impact on the deep learning model performance. In the following context, we will present how to select the training subset, validation subset based on the 5-fold cross-validation before the training stage, how to pick the learning rate during the training stage, how to select the model by the epochs during the training stage, and how to pick pruning threshold to compress the final model. The experiments are implemented on all three datasets and the model performance is evaluated by varying several model parameters.

Appendix D.4.1. Cross-Validation Tuning

In order to improve the proposed model’s final performance, k-fold cross-validation is used after segmenting the input samples. This approach can lead to a less biased or less positive assessment of the ability of the model than other methods [67]. Table A9 shows the results on the test set corresponding to different validation set choices. The pruned accuracy results are obtained when we use the fold numbers 4, 5, and 1 as the validation set based on both pruned accuracy and the nonzero parameters’ percent for the WISDM, UCI-HAR, PAMAP2 datasets, respectively.

Appendix D.4.2. Learning Rate Tuning

The learning rate is a hyper-parameter that controls how much to change the model in response to the estimated error each time the model weights are updated [68]. Table A9 demonstrates the experiment results of different learning-rate settings. For the WISDM dataset, we observe that the best performance of pruned accuracy and parameter remaining percentage is achieved when the learning rate equals 1.5 × 10 4 . For the UCI-HAR dataset, when the learning rate is 2 × 10 4 , we obtain the best results of pruned accuracy and parameter remaining percentage. For the PAMAP2 dataset, the best-pruned accuracy is achieved when the learning rate is 1.5 × 1 . 0 4 . To make the experiment settings consistent and comparable, we set the learning rate to be 1.0 × 10 4 for all three datasets.
Table A9. Impact of different cross-validation fold numbers and learning rates on the proposed 0 sparse group lasso approach on each HAR dataset—WISDM, UCI-HAR, and PAMAP2, respectively. (We highlight our selection in both fold number and learning rate for each dataset).
Table A9. Impact of different cross-validation fold numbers and learning rates on the proposed 0 sparse group lasso approach on each HAR dataset—WISDM, UCI-HAR, and PAMAP2, respectively. (We highlight our selection in both fold number and learning rate for each dataset).
DatasetTypeValueBase/Pruned Accuracy (%)Parameter Nonzero (%)Parameter Remaining (%)Node Remaining (%)
WISDMFold Number193.52/92.6811.6432.4957.42
294.88/93.7010.0330.3555.08
394.45/93.489.4527.9752.13
494.97/94.659.5228.0353.52
593.52/92.6811.6432.4957.42
Learning Rate1.0 × 10 5 89.55/86.7227.0993.5096.68
5.0 × 10 5 92.93/84.369.4140.4464.06
1.0 × 10 4 94.97/94.6527.0928.0353.52
1.5 × 10 4 94.96/94.8810.3827.2652.54
1.0 × 10 4 94.65/94.5710.5432.3856.84
UCI-HARFold Number178.42/78.0815.5331.9956.64
289.89/89.2832.4964.2980.27
379.13/79.3716.0232.2556.84
478.22/78.2218.6940.0263.48
590.06/89.9623.0046.8368.75
Learning Rate1.0 × 10 5 85.27/85.5177.9894.6697.27
5.0 × 10 5 89.38/89.2416.6985.7792.58
1.0 × 10 4 90.06/89.9623.0046.8368.75
1.5 × 10 4 90.94/90.9116.6931.0456.45
2.0 × 10 4 90.40/90.4313.2429.1054.10
PAMAP2Fold Number196.89/96.951.263.7210.74
292.29/92.281.273.1510.35
396.49/96.281.814.7414.84
495.08/94.991.203.4210.55
594.81/94.811.463.7111.52
Learning Rate1.0 × 10 5 93.63/85.807.9328.6149.22
5.0 × 10 5 94.25/93.893.9011.8128.32
1.0 × 10 4 96.89/96.951.263.7210.74
1.5 × 10 4 96.57/96.621.122.367.62
2.0 × 10 4 94.89/94.990.682.027.62

Appendix D.4.3. Number of Epochs Tuning

The number of epochs is a hyper-parameter that defines the number of times that the learning algorithm will work through the entire training dataset. Figure A1a–c show the training accuracy, validation accuracy, and testing accuracy versus the number of epochs for WISDM, UCI-HAR, and PAMAP2 datasets, respectively. For all three datasets, the number of epochs is 150; however, we only pick the epoch number based on the highest accuracy of the validation set for each dataset. The validation dataset is different from the test dataset, but is instead used to give an unbiased estimate of our final model. Based on the final results, the highest validation accuracy occurs at 150, 102, 113 epochs for WISDM, UCI-HAR, and PAMAP2 datasets, respectively.
Figure A1. Impact of the different number of epochs (ac) and prune thresholds (df) on the proposed 0 sparse group lasso approach on each HAR dataset—WISDM (a,d), UCI-HAR (b,e), and PAMAP2 (c,f), respectively.
Figure A1. Impact of the different number of epochs (ac) and prune thresholds (df) on the proposed 0 sparse group lasso approach on each HAR dataset—WISDM (a,d), UCI-HAR (b,e), and PAMAP2 (c,f), respectively.
Sensors 22 04081 g0a1

Appendix D.4.4. Prune Threshold Tuning

For all the pruning methods including 0 sparse group lasso, after pruning, the weights cannot be exact zero due to the binary bits computation. Therefore, if weight is less than a costumed prune threshold, we set the weight to be zero in those pruned models. Figure A1d–f shows the experiment results of different prune threshold settings for WISDM, UCI-HAR, and PAMAP2 datasets, respectively. For the WISDM dataset, we observe that the best performance of pruned accuracy and parameter remaining percentage is achieved when the threshold equals 0.015. For the UCI-HAR dataset, when the pruned threshold is 0.005, we obtain the best results of pruned accuracy and parameter remaining percentage. For the PAMAP2 dataset, the best-pruned accuracy is achieved when the pruned threshold is 0.01.

References

  1. Frankle, J.; Carbin, M. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019; Available online: https://arxiv.org/abs/1803.03635 (accessed on 24 May 2022).
  2. Hassibi, B.; Stork, D. Second order derivaties for network prunning: Optimal brain surgeon. In Advances in Neural Information Processing Systems 5 (NIPS 1992); Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1993. [Google Scholar]
  3. Castellano, G.; Fanelli, A.M.; Pelillo, M. An iterative pruning algorithm for feedforward neural networks. IEEE Trans. Neural Netw. 1997, 8, 519–531. [Google Scholar] [CrossRef] [PubMed]
  4. Han, S.; Pool, J.; Tran, J.; Dally, W. Learning both Weights and Connections for Efficient Neural Network. In Advances in Neural Information Processing Systems 28; Curran Associates, Inc.: Red Hook, NY, USA, 2015; pp. 1135–1143. [Google Scholar]
  5. Chetlur, S.; Woolley, C.; Vandermersch, P.; Cohen, J.; Tran, J.; Catanzaro, B.; Shelhamer, E. cuDNN: Efficient Primitives for Deep Learning. arXiv 2014, arXiv:1410.0759. [Google Scholar]
  6. Ding, X.; Ding, G.; Guo, Y.; Han, J.; Yan, C. Approximated oracle filter pruning for destructive cnn width optimization. In Proceedings of the 36th International Conference on Machine Learning, PMLR 97, Long Beach, CA, USA, 9–15 June 2019; pp. 1607–1616. [Google Scholar]
  7. Neklyudov, K.; Molchanov, D.; Ashukha, A.; Vetrov, D.P. Structured Bayesian Pruning via Log-Normal Multiplicative Noise. In Advances in Neural Information Processing Systems 30; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 6775–6784. [Google Scholar]
  8. Louizos, C.; Welling, M.; Kingma, D.P. Learning Sparse Neural Networks through L_0 Regularization. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  9. Louizos, C.; Ullrich, K.; Welling, M. Bayesian compression for deep learning. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 3288–3298. [Google Scholar]
  10. Liu, Z.G.; Whatmough, P.N.; Mattina, M. Sparse Systolic Tensor Array for Efficient CNN Hardware Acceleration. arXiv 2020, arXiv:2009.02381. [Google Scholar]
  11. Liu, Z.; Whatmough, P.N.; Mattina, M. Systolic Tensor Array: An Efficient Structured-Sparse GEMM Accelerator for Mobile CNN Inference. IEEE Comput. Archit. Lett. 2020, 19, 34–37. [Google Scholar] [CrossRef]
  12. Pilanci, M.; Wainwright, M.J.; El Ghaoui, L. Sparse learning via Boolean relaxations. Math. Prog. 2015, 151, 63–87. [Google Scholar] [CrossRef]
  13. Bertsimas, D.; Pauphilet, J.; Parys, B.V. Sparse Regression: Scalable algorithms and empirical performance. arXiv 2019, arXiv:1902.06547. [Google Scholar]
  14. Tibshirani, R. Regression Shrinkage and Selection Via the Lasso. J. R. Stat. Soc. Ser. B 1994, 58, 267–288. [Google Scholar] [CrossRef]
  15. Zou, H.; Hastie, T. Regularization and variable selection via the Elastic Net. J. R. Stat. Soc. Ser. B 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  16. Fan, J.; Li, R. Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  17. Zhang, C.H. Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 2010, 38, 894–942. [Google Scholar] [CrossRef] [Green Version]
  18. Hazimeh, H.; Mazumder, R. Fast Best Subset Selection: Coordinate Descent and Local Combinatorial Optimization Algorithms. arXiv 2018, arXiv:1803.01454. [Google Scholar] [CrossRef]
  19. Guo, Y.; Yao, A.; Chen, Y. Dynamic Network Surgery for Efficient DNNs. In Advances in Neural Information Processing Systems 29; Curran Associates, Inc.: Red Hook, NY, USA, 2016; pp. 1379–1387. [Google Scholar]
  20. Ding, X.; Ding, G.; Zhou, X.; Guo, Y.; Han, J.; Liu, J. Global Sparse Momentum SGD for Pruning Very Deep Neural Networks. In Advances in Neural Information Processing Systems 32; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 6379–6391. [Google Scholar]
  21. Xiao, X.; Wang, Z.; Rajasekaran, S. AutoPrune: Automatic Network Pruning by Regularizing Auxiliary Parameters. In Advances in Neural Information Processing Systems 32; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 13681–13691. [Google Scholar]
  22. Alvarez, J.M.; Salzmann, M. Learning the Number of Neurons in Deep Networks. In Advances in Neural Information Processing Systems 29; Curran Associates, Inc.: Red Hook, NY, USA, 2016; pp. 2270–2278. [Google Scholar]
  23. Liu, B.; Wang, M.; Foroosh, H.; Tappen, M.; Penksy, M. Sparse Convolutional Neural Networks. In Proceedings of the Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; p. 7298681242022. Available online: https://ieeexplore.ieee.org/document/7298681 (accessed on 24 May 2022).
  24. Wen, W.; Wu, C.; Wang, Y.; Chen, Y.; Li, H. Learning Structured Sparsity in Deep Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Available online: https://arxiv.org/abs/1608.03665 (accessed on 24 May 2022).
  25. Yang, C.; Yang, Z.; Khattak, A.M.; Yang, L.; Zhang, W.; Gao, W.; Wang, M. Structured Pruning of Convolutional Neural Networks via L1 Regularization. IEEE Access 2019, 7, 106385–106394. [Google Scholar] [CrossRef]
  26. Yang, H.; Gui, S.; Zhu, Y.; Liu, J. Automatic Neural Network Compression by Sparsity-Quantization Joint Learning: A Constrained Optimization-based Approach. arXiv 2020, arXiv:1910.05897. [Google Scholar]
  27. Zhang, T.; Ye, S.; Zhang, K.; Tang, J.; Wen, W.; Fardad, M.; Wang, Y. A Systematic DNN Weight Pruning Framework using Alternating Direction Method of Multipliers. arXiv 2018, arXiv:1804.03294. [Google Scholar]
  28. He, Y.; Liu, P.; Wang, Z.; Hu, Z.; Yang, Y. Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  29. Ren, A.; Zhang, T.; Ye, S.; Li, J.; Xu, W.; Qian, X.; Lin, X.; Wang, Y. ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using Alternating Direction Methods of Multipliers. In Proceedings of the Twenty-Fourth International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS ’19), Providence, RI, USA, 13–17 April 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 925–938. [Google Scholar] [CrossRef]
  30. Liu, J.; Ye, J. Moreau-Yosida Regularization for Grouped Tree Structure Learning. In Advances in Neural Information Processing Systems 23; Lafferty, J.D., Williams, C.K.I., Shawe-Taylor, J., Zemel, R.S., Culotta, A., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2010; pp. 1459–1467. [Google Scholar]
  31. Collins, M.D.; Kohli, P. Memory Bounded Deep Convolutional Networks. arXiv 2014, arXiv:1412.1442. [Google Scholar]
  32. Li, H.; Kadav, A.; Durdanovic, I.; Samet, H.; Graf, H.P. Pruning Filters for Efficient ConvNets. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  33. Liu, Z.; Li, J.; Shen, Z.; Huang, G.; Yan, S.; Zhang, C. Learning efficient convolutional networks through network slimming. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2736–2744. [Google Scholar]
  34. Yoon, J.; Hwang, S.J. Combined Group and Exclusive Sparsity for Deep Neural Networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, NSW, Australia, 6–11 August 2017; PMLR: Proceedings of Machine Learning Research, Precup, D., Teh, Y.W., Eds.; International Convention Centre: Sydney, Australia, 2017; Volume 70, pp. 3958–3966. [Google Scholar]
  35. Scardapane, S.; Comminiello, D.; Hussain, A.; Uncini, A. Group sparse regularization for deep neural networks. Neurocomputing 2017, 241, 81–89. [Google Scholar] [CrossRef] [Green Version]
  36. Liu, Z.G.; Whatmough, P.N.; Zhu, Y.; Mattina, M. S2TA: Exploiting Structured Sparsity for Energy-Efficient Mobile CNN Acceleration. In Proceedings of the 2022 IEEE International Symposium on High-Performance Computer Architecture (HPCA), Seoul, Korea, 2–6 April 2022. [Google Scholar]
  37. Bolte, J.; Sabach, S.; Teboulle, M. Proximal alternating linearized minimization for nonconvex and nonsmooth problems. Math. Program. 2014, 146, 459–494. [Google Scholar] [CrossRef]
  38. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. Siam J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef] [Green Version]
  39. Dai, B.; Zhu, C.; Guo, B.; Wipf, D. Compressing Neural Networks using the Variational Information Bottleneck. In Proceedings of the 35th International Conference on Machine Learning (ICML 2018), Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  40. LeCun, Y.; Denker, J.S.; Solla, S.A. Optimal Brain Damage. In Advances in Neural Information Processing Systems 2; Touretzky, D.S., Ed.; Morgan-Kaufmann: Burlington, MA, USA, 1990; pp. 598–605. [Google Scholar]
  41. Zeng, W.; Urtasun, R. MLPrune: Multi-Layer Pruning for Automated Neural Network Compression. 2019. Available online: https://openreview.net/forum?id=r1g5b2RcKm (accessed on 24 May 2022).
  42. Wang, C.; Grosse, R.; Fidler, S.; Zhang, G. EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis. In Proceedings of the Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 6566–6575. [Google Scholar]
  43. Deng, L. The mnist database of handwritten digit images for machine learning research. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
  44. Zagoruyko, S. 92.45% on CIFAR-10 in Torch. 2015. Available online: http://torch.ch/blog/2015/07/30/cifar.html (accessed on 24 May 2022).
  45. Zhang, G.; Wang, C.; Xu, B.; Grosse, R. Three Mechanisms of Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  46. Krizhevsky, A.; Nair, V.; Hinton, G. CIFAR-10 (Canadian Institute for Advanced Research). Available online: http://www.cs.toronto.edu/~kriz/cifar.html (accessed on 24 May 2022).
  47. Le, Y.; Yang, X.S. Tiny ImageNet Visual Recognition Challenge. CS 231N 2015, 7, 3. [Google Scholar]
  48. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  49. Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. Activity Recognition Using Cell Phone Accelerometers. SIGKDD Explor. Newsl. 2011, 12, 74–82. [Google Scholar] [CrossRef]
  50. WISDM: Wireless Sensor Data Mining. Available online: https://www.cis.fordham.edu/wisdm/dataset.php (accessed on 24 May 2022).
  51. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J. A Public Domain Dataset for Human Activity Recognition using Smartphones. In Proceedings of the 21th international European symposium on artificial neural networks, computational intelligence and machine learning, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
  52. Human Activity Recognition Using Smartphones Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/human+activity+recognition+using+smartphones (accessed on 24 May 2022).
  53. Reiss, A.; Stricker, D. Introducing a New Benchmarked Dataset for Activity Monitoring. In Proceedings of the 2012 16th International Symposium on Wearable Computers, Newcastle, UK, 18–22 June 2012. [Google Scholar] [CrossRef]
  54. Reiss, A.; Stricker, D. Creating and Benchmarking a New Dataset for Physical Activity Monitoring. In PETRA ’12, Proceedings of the 5th International Conference on PErvasive Technologies Related to Assistive Environments, Crete, Greece, 6–8 June 2012; Association for Computing Machinery: New York, NY, USA, 2012. [Google Scholar] [CrossRef]
  55. PAMAP2 Physical Activity Monitoring Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/PAMAP2+Physical+Activity+Monitoring (accessed on 24 May 2022).
  56. Google Colab. Available online: https://research.google.com/colaboratory/faq.html (accessed on 24 May 2022).
  57. Pytorch Mobile. Available online: https://pytorch.org/mobile/android/ (accessed on 24 May 2022).
  58. Profile Battery Usage with Batterystats and Battery Historian. Available online: https://developer.android.com/topic/performance/power/setup-battery-historian (accessed on 24 May 2022).
  59. Yuan, L.; Liu, J.; Ye, J. Efficient Methods for Overlapping Group Lasso. In Advances in Neural Information Processing Systems 24; Shawe-Taylor, J., Zemel, R.S., Bartlett, P.L., Pereira, F., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2011; pp. 352–360. [Google Scholar]
  60. Zeng, J.; Lau, T.T.K.; Lin, S.B.; Yao, Y. Global convergence of block coordinate descent in deep learning. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; Available online: https://arXiv:1803.00225 (accessed on 24 May 2022).
  61. Bao, C.; Ji, H.; Quan, Y.; Shen, Z. L0 norm based dictionary learning by proximal methods with global convergence. In Proceedings of the Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar] [CrossRef]
  62. Lau, T.T.K.; Zeng, J.; Wu, B.; Yao, Y. A Proximal Block Coordinate Descent Algorithm for Deep Neural Network Training. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018—Workshop Track Proceedings, Vancouver, BC, Canada, 3 May–30 April 2018; Available online: arxiv.org/abs/1803.09082 (accessed on 24 May 2022).
  63. Attouch, H.; Bolte, J. On the convergence of the proximal algorithm for nonsmooth functions involving analytic features. Math. Program. 2009, 116, 5–16. [Google Scholar] [CrossRef]
  64. Bach, F.R.; Mairal, J.; Ponce, J. Convex Sparse Matrix Factorizations. arXiv 2008, arXiv:0812.1869. [Google Scholar]
  65. Shomron, G.; Weiser, U. Non-Blocking Simultaneous Multithreading: Embracing the Resiliency of Deep Neural Networks. In Proceedings of the 2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO), Athens, Greece, 17–21 October 2020; Available online: arxiv.org/abs/2004.09309 (accessed on 24 May 2022).
  66. Wang, Z.; Oates, T. Imaging Time-Series to Improve Classification and Imputation. arXiv 2015, arXiv:1506.00327. [Google Scholar]
  67. Tamilarasi, P.; Rani, R. Diagnosis of Crime Rate against Women using k-fold Cross Validation through Machine Learning. In Proceedings of the 2020 Fourth International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 11–13 March 2020; pp. 1034–1038. [Google Scholar] [CrossRef]
  68. Brownlee, J. Understand the Impact of Learning Rate on Neural Network Performance. 2019. Available online: https://machinelearningmastery.com/understand-the-dynamics-of-learning-rate-on-deep-learning-neural-networks (accessed on 24 May 2022).
Figure 1. Observations of different strategies’ pruned filter matrix for hardware acceleration with software implementation of convolution in cuDNN. (a) General Matrix Multiply (GEMM) is applied in cuDNN. (b) Different strategies such as no pruning, individual sparsity, column-wise group sparsity, and both individual sparsity and column-wise group sparsity on pruning the filter matrix. (c) The pruned filter matrix implemented in cuDNN and determined whether it can be used for hardware acceleration or not.
Figure 1. Observations of different strategies’ pruned filter matrix for hardware acceleration with software implementation of convolution in cuDNN. (a) General Matrix Multiply (GEMM) is applied in cuDNN. (b) Different strategies such as no pruning, individual sparsity, column-wise group sparsity, and both individual sparsity and column-wise group sparsity on pruning the filter matrix. (c) The pruned filter matrix implemented in cuDNN and determined whether it can be used for hardware acceleration or not.
Sensors 22 04081 g001
Figure 2. Overview of the proposed MobilePrune method. (a) Group sparsity for weights of a neuron for fully connected layers. (b) Sparsity on individual weights for fully connected layers. (c) Pruning strategy for fully connected layers and their effect where sparsity is induced on both neuron-wise groups and individual weights. (d) Group and individual sparsity for convolutional layers.
Figure 2. Overview of the proposed MobilePrune method. (a) Group sparsity for weights of a neuron for fully connected layers. (b) Sparsity on individual weights for fully connected layers. (c) Pruning strategy for fully connected layers and their effect where sparsity is induced on both neuron-wise groups and individual weights. (d) Group and individual sparsity for convolutional layers.
Sensors 22 04081 g002
Table 1. Comparison of pruned models with state-of-the-art methods on different datasets – MNIST, CIFAR-10, and Tiny-ImageNet, respectively. (We highlight our MobilePrune results and mark the best performance as blue among different methods for each model in each dataset).
Table 1. Comparison of pruned models with state-of-the-art methods on different datasets – MNIST, CIFAR-10, and Tiny-ImageNet, respectively. (We highlight our MobilePrune results and mark the best performance as blue among different methods for each model in each dataset).
DatasetModelMethodsBase/Pruned Accuracy (%)Original/Remaining Parameters (Mil)FLPOs (Mil)
MNIST BC-GNJ [9]98.40/98.20267.00/28.7328.64
BC-GHS [9]98.40/98.20267.00/28.1728.09
LeNet-300-100L0 [8]-/98.60-69.27
L0-sep [8]-/98.20-26.64
MobilePrune98.24/98.23267.00/5.2525.79
SBP [7]-/99.14-212.80
BC-GNJ [9]99.10/99.00431.00/3.88282.87
BC-GHS [9]99.10/99.00431.00/2.59153.38
LeNet-5L0 [8]-/99.10-1113.40
L0-sep [8]-/99.00-390.68
MobilePrune99.12/99.11431.00/2.31113.50
CIFAR-10 Original [44]-/92.4515.00/-313.5
PF [32]-/93.4015.00/5.4206.3
VGG-likeSBP [7]92.80/92.5015.00/-136.0
SBPa [7]92.80/91.0015.00/-99.20
VIBNet [39]-/93.5015.00/0.8786.82
MobilePrune92.96/92.9415.00/0.6077.83
C-OBD [40]95.30/95.277.42/2.92488.85
C-OBS [2]95.30/95.307.42/3.04378.22
ResNet32Kron-OBD [40,41]95.30/95.307.42/3.26526.17
Kron-OBS [2,41]95.30/95.467.42/3.23524.52
EigenDamage [42]95.30/95.287.42/2.99457.46
MobilePrune95.29/95.477.42/2.93371.30
NN slimming [33]61.56/40.0520.12/5.83158.62
C-OBD [40]61.56/47.3620.12/4.21481.90
C-OBS [2]61.56/39.8020.12/6.55210.05
Tiny-ImageNetVGG-19Kron-OBD [40,41]61.56/44.4120.12/4.72298.28
Kron-OBS [2,41]61.56/44.5420.12/5.26266.43
EigenDamage [42]61.56/56.9220.12/5.21408.17
MobilePrune61.56/56.2720.12/4.05407.37
Table 2. Comparison of pruning method on the desktop with state-of-the-art methods for pruning accuracy, pruning rate and response delay on HAR datasets—WISDM, HCI-HAR, and PAMAP2, respectively. (We highlight our MobilePrune results and mark the best performance as blue among different penalties for each dataset).
Table 2. Comparison of pruning method on the desktop with state-of-the-art methods for pruning accuracy, pruning rate and response delay on HAR datasets—WISDM, HCI-HAR, and PAMAP2, respectively. (We highlight our MobilePrune results and mark the best performance as blue among different penalties for each dataset).
DatasetPenaltyBase/Pruned Accuracy (%)Parameter Nonzero (%)Parameter Remaining (%)Node Remaining (%)Base/Pruned Response Delay (s)Time Saving Percentage (%)
WISDM l 0 norm94.72/94.7963.36100.00100.000.38/0.390.00
l 1 norm94.30/93.8413.5846.2668.160.38/0.2436.84
l 2 norm94.61/94.5456.2890.4695.120.38/0.357.89
Group lasso94.68/94.3248.2389.7394.730.38/0.357.89
l 1 sparse Group lasso94.81/94.7917.9153.4173.830.41/0.2636.59
MobilePrune94.97/94.659.5228.0352.520.50/0.1766.00
UCI-HAR l 0 norm91.52/91.4888.49100.00100.000.84/0.804.76
l 1 norm90.46/90.3381.5898.4799.220.81/0.820.00
l 2 norm91.01/90.9488.35100.00100.000.79/0.800.00
Group lasso90.80/90.8482.91100.00100.000.83/0.786.02
l 1 sparse Group lasso91.11/91.0481.2197.7098.830.84/0.804.76
MobilePrune90.06/89.9623.0046.8368.751.01/0.4357.43
PAMAP2 l 0 norm93.15/93.0769.27100.00100.000.41/0.410.00
l 1 norm95.22/95.291.467.2819.730.40/0.0880.00
l 2 norm92.08/92.0965.3294.9397.270.41/0.394.88
Group lasso93.30/93.2861.78100.00100.000.41/0.410.00
l 1 sparse Group lasso96.87/97.202.679.7226.170.40/0.1075.00
MobilePrune96.89/96.951.263.7210.740.51/0.0590.20
Table 3. Comparison of pruning method on the mobile devices with other state-of-the-art pruning methods for computational cost and battery usage on HAR dataset—WISDM, HCI-HAR, and PAMAP2, respectively. (We highlight our MobilePrune results and mark the best performance as blue among different penalties for each device in each dataset).
Table 3. Comparison of pruning method on the mobile devices with other state-of-the-art pruning methods for computational cost and battery usage on HAR dataset—WISDM, HCI-HAR, and PAMAP2, respectively. (We highlight our MobilePrune results and mark the best performance as blue among different penalties for each device in each dataset).
DatasetDevicePenaltyBase/Pruned Response Delay (s)Time Saving Percentage (%)Based/Pruned Device Estimated Battery Use (%/h)Battery Saving Percentage (%)
WISDM Huawei P20 l 0 norm1.40/1.279.290.71/0.701.41
l 1 norm1.33/0.7146.620.74/0.6512.16
l 2 norm1.28/1.215.470.74/0.770.00
Group lasso1.27/1.270.000.74/0.770.00
l 1 sparse Group lasso1.25/0.8135.200.74/0.688.11
MobilePrune1.34/0.5161.940.72/0.4537.50
OnePlus 8 Pro l 0 norm0.57/0.4914.040.34/0.325.88
l 1 norm0.48/0.3429.170.35/0.3014.29
l 2 norm0.48/0.4016.670.34/0.340.00
Group lasso0.49/0.458.160.34/0.350.00
l 1 sparse Group lasso0.48/0.3331.250.35/0.3014.29
MobilePrune0.48/0.2352.080.34/0.2332.35
HCI-HAR Huawei P20 l 0 norm1.43/1.430.000.84/0.840.00
l 1 norm1.42/1.420.000.85/0.841.18
l 2 norm1.43/1.430.000.84/0.840.00
Group lasso1.43/1.430.000.84/0.822.38
l 1 sparse Group lasso1.42/1.410.700.85/0.823.53
MobilePrune1.42/0.8540.140.84/0.5534.52
OnePlus 8 Pro l 0 norm0.53/0.530.000.35/0.350.00
l 1 norm0.54/0.515.560.37/0.362.70
l 2 norm0.54/0.531.850.37/0.370.00
Group lasso0.53/0.521.890.36/0.360.00
l 1 sparse Group lasso0.53/0.521.890.36/0.360.00
MobilePrune0.54/0.4222.220.36/0.2919.44
PAMAP2 Huawei P20 l 0 norm2.64/2.720.000.76/0.790.00
l 1 norm2.74/0.4583.580.79/0.5332.91
l 2 norm2.67/2.564.120.78/0.780.00
Group lasso2.67/2.680.000.78/0.780.00
l 1 sparse Group lasso2.69/0.5579.550.79/0.5727.85
MobilePrune2.70/0.3288.150.79/0.5036.71
OnePlus 8 Pro l 0 norm0.94/0.931.060.88/0.880.00
l 1 norm0.93/0.2573.120.87/0.5536.78
l 2 norm0.93/0.912.150.88/0.871.14
Group lasso0.94/0.950.000.89/0.890.00
l 1 sparse Group lasso0.95/0.2969.470.88/0.5932.95
MobilePrune0.94/0.2177.660.87/0.5437.93
Table 4. Alation studies on various network models. (We mark the best performance as blue among different penalties for each model).
Table 4. Alation studies on various network models. (We mark the best performance as blue among different penalties for each model).
Network ModelPenaltyBase/Pruned Accuracy (%)Original/Remaining Parameters (Mil)FLOPsSparsity (%)
LetNet-300 0 norm98.24/98.46267 K/57.45 K143.2021.55
Group lasso98.24/98.17267 K/32.06 K39.7012.01
1 sparse group lasso98.24/98.00267 K/15.80 K25.885.93
0 sparse group lasso98.24/98.23267 K/5.25 K25.791.97
LetNet-5 0 norm99.12/99.20431 K/321.0 K2293.074.48
Group lasso99.12/99.11431 K/8.81 K187.002.04
1 sparse group lasso99.12/99.03431 K/9.98 K183.832.32
0 sparse group lasso99.12/99.11431 K/2.31 K113.500.54
VGG-like 0 norm92.96/93.4015 M/3.39 M210.9422.6
Group lasso92.96/92.4715 M/0.84 M78.075.60
1 sparse group lasso92.96/92.9015 M/0.61 M134.354.06
0 sparse group lasso92.96/92.9415 M/0.60 M77.834.00
ResNet-32 0 norm95.29/95.687.42 M/6.74 M993.1190.84
Group lasso95.29/95.307.42 M/3.03 M373.0940.84
1 sparse group lasso95.29/95.047.42 M/5.66 M735.1276.28
0 sparse group lasso95.29/95.477.42 M/2.93 M371.3039.49
VGG-19 0 norm61.56/61.99138 M/19.29 M1519.2313.98
Group lasso61.56/53.25138 M/5.93 M683.994.30
1 sparse group lasso61.56/53.97138 M/0.21 M1282.820.15
0 sparse group lasso61.56/56.27138 M/4.05 M407.372.93
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shao, Y.; Zhao, K.; Cao, Z.; Peng, Z.; Peng, X.; Li, P.; Wang, Y.; Ma, J. MobilePrune: Neural Network Compression via 0 Sparse Group Lasso on the Mobile System. Sensors 2022, 22, 4081. https://doi.org/10.3390/s22114081

AMA Style

Shao Y, Zhao K, Cao Z, Peng Z, Peng X, Li P, Wang Y, Ma J. MobilePrune: Neural Network Compression via 0 Sparse Group Lasso on the Mobile System. Sensors. 2022; 22(11):4081. https://doi.org/10.3390/s22114081

Chicago/Turabian Style

Shao, Yubo, Kaikai Zhao, Zhiwen Cao, Zhehao Peng, Xingang Peng, Pan Li, Yijie Wang, and Jianzhu Ma. 2022. "MobilePrune: Neural Network Compression via 0 Sparse Group Lasso on the Mobile System" Sensors 22, no. 11: 4081. https://doi.org/10.3390/s22114081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop