IEEE Computational Intelligence Magazine - August 2021 - 11

DNNs that are suitable for a broad range of different memory
usage and energy consumption requirements. Under negligible
accuracy loss, EMOMC improves the energy efficiency and
model compression rate of VGG-16 on CIFAR-10 by a factor of
more than
89 . # and .42 #, respectively.
I. Introduction
D
eep neural networks (DNNs) are artificial neural networks
with more than three layers (i.e., more than one
hidden layer), which progressively extract higher-level
features from the raw input in the learning process.
They have delivered the state-of-the-art accuracy on various
real-world problems, such as image classification, face recognition,
and language translation [1]. The superior accuracy of
DNNs, however, comes at the cost of high computational and
space complexity. For example, the VGG-16 model [2] has about
138 million parameters, which requires over 500 MB memory
for storage and 15.5G multiply-and-accumulates (MACs) to
process an input image with 224 × 224 pixels. In myriad application
scenarios, it is desirable to make the inference on edge
devices rather than on cloud, for reducing the latency and
dependency on connectivity and improving privacy and security.
Many of the edge devices that draw the DNNs inference have
stringent limitations on energy consumption, memory capacity,
etc. The large-scale DNNs [3], [4] are usually difficult to be
deployed on edge devices, thus hindering their wide application.
Efficient processing of DNNs for inference has become increasingly
important for the deployment on edge devices. For generating
efficient DNNs, many neural architecture search (NAS)
approaches have been developed in recent years [5]-[7]. One way
of carrying out NAS is to search from scratch [8], [9]. In contrast,
model compression1 [10] searches for the optimal networks starting
from a well-trained network. For instance, to reduce the storage
requirement of DNNs, Han et al. proposed a three-stage pipeline
(i.e., pruning, trained quantization, and Huffman coding) to compress
redundant weights [10]. Wang et al. suggested removing
redundant convolution filters to reduce the model size [11]. Rather
than reducing the model size, a few attempts [12], [13] are conducted
to compress DNNs directly by taking the energy consumption
as the feedback signals. They have achieved promising results in
reducing the size of weight parameters (or energy consumption).
However, these approaches require the model to achieve approximately
no loss of accuracy, rendering the solution less flexible.
In practice, different users often have distinct preferences on
desirable objectives, e.g., accuracy, model size, energy efficiency, and
latency, when they select the optimal DNN model for their applications.
In this paper, a novel approach, called Evolutionary MultiObjective
Model Compression (EMOMC), is proposed to
optimize energy efficiency/model size and accuracy simultaneously.
By considering network pruning and quantization, the model
compression is formulated as a multi-objective problem under different
dataflow designs and parameter coding schemes. Each candidate
architecture can be regarded as an individual in the
1The technique aims to shrink the size of the neural network model without a significant
drop of accuracy.
evolutionary population. Owing to the cooperation and interplay
of the evolution between different architectures in the population,
a set of compact DNNs that offer trade-offs on different objectives
(e.g., accuracy, energy efficiency, and model size) can be obtained in
a single run. Unlike most existing approaches which aim to reduce
the size of weight parameters or the energy consumption with no
significant loss of accuracy, the proposed approach attempts to
achieve a good balance between desired objectives, for meeting the
requirements of different edge devices. Experimental results demonstrate
that the proposed approach can obtain a diverse population
of compact DNNs for customized requirements of accuracy,
memory capacity, and energy consumption.
The novelty and main contributions of this work can be
summarized as follows:
❏ The model compression problem is formulated as a multiobjective
problem. The optimal solutions are searched in the
network pruning and quantization space using a population-based
algorithm.
❏ To speed up the population evolution, a two-stage pruning/
quantization co-optimization strategy is developed based on
the orthogonality between pruning and quantization.
❏ The trade-offs between accuracy, energy efficiency, and
model size in model compression are explored by considering
different dataflow designs and parameter coding schemes.
The experimental results demonstrate that the proposed
method can obtain a set of diverse Pareto optimal solutions in
a single run. Also, it achieves a considerably higher energy
efficiency than current state-of-the-art methods.
II. Preliminaries
Network pruning and quantization are two commonly used
model compression techniques to improve the energy efficiency
in model inference and/or to shrink the size of the model.
Moreover, the dataflow design employed by edge devices and the
coding scheme applied to store the weight matrix both have a
significant impact on the performance of model compression.
A. Network Pruning and Quantization
For making the training easy, the networks are usually overparameterized
[14]. Pruning is a widely-used model compression
technique that can effectively reduce the energy consumption of
edge devices and shrink the model size [10]. Network pruning
removes some of the redundant parameters in the network by
setting their values as zeros. A well-trained neural network usually
contains a large number of weights whose values are relatively
small compared to other parameters. In most cases, these parameters
are not particularly important when performing model inference.
Hence, one can sort all the parameters in the model and
replace those parameters with the least absolute values by zeros,
while the accuracy of the model can still be maintained. For
instance, the pruning amount to be 33%, then one-third of the
parameters in the model will be replaced by zeros. In the inference
process, if the processing elements (PEs)2 whose input
2The PE is a basic unit to conduct computation in processors.
AUGUST 2021 | IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE 11

IEEE Computational Intelligence Magazine - August 2021

Table of Contents for the Digital Edition of IEEE Computational Intelligence Magazine - August 2021

Contents
IEEE Computational Intelligence Magazine - August 2021 - Cover1
IEEE Computational Intelligence Magazine - August 2021 - Cover2
IEEE Computational Intelligence Magazine - August 2021 - Contents
IEEE Computational Intelligence Magazine - August 2021 - 2
IEEE Computational Intelligence Magazine - August 2021 - 3
IEEE Computational Intelligence Magazine - August 2021 - 4
IEEE Computational Intelligence Magazine - August 2021 - 5
IEEE Computational Intelligence Magazine - August 2021 - 6
IEEE Computational Intelligence Magazine - August 2021 - 7
IEEE Computational Intelligence Magazine - August 2021 - 8
IEEE Computational Intelligence Magazine - August 2021 - 9
IEEE Computational Intelligence Magazine - August 2021 - 10
IEEE Computational Intelligence Magazine - August 2021 - 11
IEEE Computational Intelligence Magazine - August 2021 - 12
IEEE Computational Intelligence Magazine - August 2021 - 13
IEEE Computational Intelligence Magazine - August 2021 - 14
IEEE Computational Intelligence Magazine - August 2021 - 15
IEEE Computational Intelligence Magazine - August 2021 - 16
IEEE Computational Intelligence Magazine - August 2021 - 17
IEEE Computational Intelligence Magazine - August 2021 - 18
IEEE Computational Intelligence Magazine - August 2021 - 19
IEEE Computational Intelligence Magazine - August 2021 - 20
IEEE Computational Intelligence Magazine - August 2021 - 21
IEEE Computational Intelligence Magazine - August 2021 - 22
IEEE Computational Intelligence Magazine - August 2021 - 23
IEEE Computational Intelligence Magazine - August 2021 - 24
IEEE Computational Intelligence Magazine - August 2021 - 25
IEEE Computational Intelligence Magazine - August 2021 - 26
IEEE Computational Intelligence Magazine - August 2021 - 27
IEEE Computational Intelligence Magazine - August 2021 - 28
IEEE Computational Intelligence Magazine - August 2021 - 29
IEEE Computational Intelligence Magazine - August 2021 - 30
IEEE Computational Intelligence Magazine - August 2021 - 31
IEEE Computational Intelligence Magazine - August 2021 - 32
IEEE Computational Intelligence Magazine - August 2021 - 33
IEEE Computational Intelligence Magazine - August 2021 - 34
IEEE Computational Intelligence Magazine - August 2021 - 35
IEEE Computational Intelligence Magazine - August 2021 - 36
IEEE Computational Intelligence Magazine - August 2021 - 37
IEEE Computational Intelligence Magazine - August 2021 - 38
IEEE Computational Intelligence Magazine - August 2021 - 39
IEEE Computational Intelligence Magazine - August 2021 - 40
IEEE Computational Intelligence Magazine - August 2021 - 41
IEEE Computational Intelligence Magazine - August 2021 - 42
IEEE Computational Intelligence Magazine - August 2021 - 43
IEEE Computational Intelligence Magazine - August 2021 - 44
IEEE Computational Intelligence Magazine - August 2021 - 45
IEEE Computational Intelligence Magazine - August 2021 - 46
IEEE Computational Intelligence Magazine - August 2021 - 47
IEEE Computational Intelligence Magazine - August 2021 - 48
IEEE Computational Intelligence Magazine - August 2021 - 49
IEEE Computational Intelligence Magazine - August 2021 - 50
IEEE Computational Intelligence Magazine - August 2021 - 51
IEEE Computational Intelligence Magazine - August 2021 - 52
IEEE Computational Intelligence Magazine - August 2021 - 53
IEEE Computational Intelligence Magazine - August 2021 - 54
IEEE Computational Intelligence Magazine - August 2021 - 55
IEEE Computational Intelligence Magazine - August 2021 - 56
IEEE Computational Intelligence Magazine - August 2021 - 57
IEEE Computational Intelligence Magazine - August 2021 - 58
IEEE Computational Intelligence Magazine - August 2021 - 59
IEEE Computational Intelligence Magazine - August 2021 - 60
IEEE Computational Intelligence Magazine - August 2021 - 61
IEEE Computational Intelligence Magazine - August 2021 - 62
IEEE Computational Intelligence Magazine - August 2021 - 63
IEEE Computational Intelligence Magazine - August 2021 - 64
IEEE Computational Intelligence Magazine - August 2021 - 65
IEEE Computational Intelligence Magazine - August 2021 - 66
IEEE Computational Intelligence Magazine - August 2021 - 67
IEEE Computational Intelligence Magazine - August 2021 - 68
IEEE Computational Intelligence Magazine - August 2021 - 69
IEEE Computational Intelligence Magazine - August 2021 - 70
IEEE Computational Intelligence Magazine - August 2021 - 71
IEEE Computational Intelligence Magazine - August 2021 - 72
IEEE Computational Intelligence Magazine - August 2021 - 73
IEEE Computational Intelligence Magazine - August 2021 - 74
IEEE Computational Intelligence Magazine - August 2021 - 75
IEEE Computational Intelligence Magazine - August 2021 - 76
IEEE Computational Intelligence Magazine - August 2021 - 77
IEEE Computational Intelligence Magazine - August 2021 - 78
IEEE Computational Intelligence Magazine - August 2021 - 79
IEEE Computational Intelligence Magazine - August 2021 - 80
IEEE Computational Intelligence Magazine - August 2021 - 81
IEEE Computational Intelligence Magazine - August 2021 - 82
IEEE Computational Intelligence Magazine - August 2021 - 83
IEEE Computational Intelligence Magazine - August 2021 - 84
IEEE Computational Intelligence Magazine - August 2021 - 85
IEEE Computational Intelligence Magazine - August 2021 - 86
IEEE Computational Intelligence Magazine - August 2021 - 87
IEEE Computational Intelligence Magazine - August 2021 - 88
IEEE Computational Intelligence Magazine - August 2021 - 89
IEEE Computational Intelligence Magazine - August 2021 - 90
IEEE Computational Intelligence Magazine - August 2021 - 91
IEEE Computational Intelligence Magazine - August 2021 - 92
IEEE Computational Intelligence Magazine - August 2021 - 93
IEEE Computational Intelligence Magazine - August 2021 - 94
IEEE Computational Intelligence Magazine - August 2021 - 95
IEEE Computational Intelligence Magazine - August 2021 - 96
IEEE Computational Intelligence Magazine - August 2021 - Cover3
IEEE Computational Intelligence Magazine - August 2021 - Cover4
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202311
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202308
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202305
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202302
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202211
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202208
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202205
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202202
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202111
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202108
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202105
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202102
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202011
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202008
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202005
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202002
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201911
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201908
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201905
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201902
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201811
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201808
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201805
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201802
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter12
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall12
https://www.nxtbookmedia.com