Systems, Man & Cybernetics - July 2015 - 22

To summarize, CNNs have the following apparent
drawbacks and limitations:
◆ BP learning speed and generalization performance for
limited training data sets
◆ tweaks and tricks should be adopted to define the
architecture parameters.
Extreme Learning Machines
ANNs and CNNs have been widely explored as architectures to handle complex and challenging tasks. While new
learning techniques have been proposed by the community, almost all of them are derived from the BP algorithm.
Hence, such networks lack faster learning procedures. It
is, therefore, not surprising to see that it may take hours
[40], days [4], and even longer to train a single ANN/CNN
using the traditional BP-based methods.
Biological-Inspired Learning
Extreme learning machines (ELMs) [41] were inspired by
biological learning and proposed to overcome the issues
faced by the BP-based learning algorithms. ELMs were
devised following the idea that some parts of human brain
systems have random neurons with parameters that are
independent of their environment. All of the hidden nodes
in an ELM are independent of the training data as well as
each other (Figure 3). Hidden nodes need not be tuned, and
the input-to-hidden weights can be randomly generated
before seeing the training data. The solution to learn the
optimal hidden-to-output weights (to obtain the correct
input-output mappings) is the Moore-Penrose generalized
inverse [42].
ELMs: Advantages and Shortcomings
ELMs are extremely fast learning algorithms for SLFNs,
and they have the following important properties.

Input Layer

x1

Hidden Layer

Output Layer

Forward Propagation
b1,
1
h1
Random
ts
h
ig
e
W
h2

yt 1
yt 2

x2
b 1,

hL - 2

D

xd

yt D

hL - 1

Figure 3. The architecture of an ELM. Unlike previously

seen neural architectures, there is no BP in the
learning process. With the ELM, the input-to-hidden
weights are randomly picked. Only the set of hiddento-output connection weights b are learned.

22

IEEE Systems, Man, & Cybernetics Magazine July 2015 	

◆ Minimum training error: The special solution for

learning the hidden-to-output layer connection
weights, derived from exploiting the properties of the
network structure and the Moore-Penrose generalized inverse matrix [42], is one of the least-square
solutions of a linear system.
◆ The smallest norm of weights and best generalization performance: Bartlett's neural network generalization theory [43] for feedforward neural networks
asserts that the smaller the norms of weights are,
the better generalization performance the networks
tend to have. ELMs have this property; indeed, the
solution has the smallest norm among all the leastsquare solutions.
◆ Unique minimum norm least-square solution: The
minimum norm solution for learning the hidden-tooutput weights is unique.
While ELMs have valuable properties, their success
or failure, like ANN and NT architectures, hinges on the
adopted data representation.
The Future of Brain-Inspired Learning
Architectures: Extreme Deep Learning Trees
Considering the strengths and weaknesses of each of the
previously presented architectures, we propose a novel
architecture that borrows their strengths and combines
them into a unique system. We named the architecture the
extreme deep learning tree (EDLT).
EDLT Structure
The EDLT exploits an NT architecture whose in-node
models are composed of an ELM stacked at the output of a
three-layer CNN (i.e., a CNN composed of a convolutional,
an LCN, and a pooling layer). See Figure 4 for the in-node
model architecture.
The Tree: Less Prior Knowledge
One of the main current issues of ANNs is correctly determining the number and the arrangement of hidden layers and neurons before the training starts. A common
approach is to train many different networks (each with a
different structure) and then to adopt the configuration of
the one that has the minimum validation error. However,
even if the optimal configuration for a particular task can
be found, it is not certain that the chosen configuration
will be the best one. The possible configurations of a
single network with only one hidden layer are infinite (we
can have one layer with infinite hidden neurons). The tree
structure of an NT, together with the adoption of appropriate split nodes, allows us to solve such a problem.
CNN Layers: Random Feature Extraction
We have shown that CNNs are able to learn meaningful
feature extractors only by looking at the data. However,
this process requires a network to perform several forward/backward propagations over the data. In addition



Table of Contents for the Digital Edition of Systems, Man & Cybernetics - July 2015

Systems, Man & Cybernetics - July 2015 - Cover1
Systems, Man & Cybernetics - July 2015 - Cover2
Systems, Man & Cybernetics - July 2015 - 1
Systems, Man & Cybernetics - July 2015 - 2
Systems, Man & Cybernetics - July 2015 - 3
Systems, Man & Cybernetics - July 2015 - 4
Systems, Man & Cybernetics - July 2015 - 5
Systems, Man & Cybernetics - July 2015 - 6
Systems, Man & Cybernetics - July 2015 - 7
Systems, Man & Cybernetics - July 2015 - 8
Systems, Man & Cybernetics - July 2015 - 9
Systems, Man & Cybernetics - July 2015 - 10
Systems, Man & Cybernetics - July 2015 - 11
Systems, Man & Cybernetics - July 2015 - 12
Systems, Man & Cybernetics - July 2015 - 13
Systems, Man & Cybernetics - July 2015 - 14
Systems, Man & Cybernetics - July 2015 - 15
Systems, Man & Cybernetics - July 2015 - 16
Systems, Man & Cybernetics - July 2015 - 17
Systems, Man & Cybernetics - July 2015 - 18
Systems, Man & Cybernetics - July 2015 - 19
Systems, Man & Cybernetics - July 2015 - 20
Systems, Man & Cybernetics - July 2015 - 21
Systems, Man & Cybernetics - July 2015 - 22
Systems, Man & Cybernetics - July 2015 - 23
Systems, Man & Cybernetics - July 2015 - 24
Systems, Man & Cybernetics - July 2015 - 25
Systems, Man & Cybernetics - July 2015 - 26
Systems, Man & Cybernetics - July 2015 - 27
Systems, Man & Cybernetics - July 2015 - 28
Systems, Man & Cybernetics - July 2015 - 29
Systems, Man & Cybernetics - July 2015 - 30
Systems, Man & Cybernetics - July 2015 - 31
Systems, Man & Cybernetics - July 2015 - 32
Systems, Man & Cybernetics - July 2015 - 33
Systems, Man & Cybernetics - July 2015 - 34
Systems, Man & Cybernetics - July 2015 - 35
Systems, Man & Cybernetics - July 2015 - 36
Systems, Man & Cybernetics - July 2015 - 37
Systems, Man & Cybernetics - July 2015 - 38
Systems, Man & Cybernetics - July 2015 - 39
Systems, Man & Cybernetics - July 2015 - 40
Systems, Man & Cybernetics - July 2015 - 41
Systems, Man & Cybernetics - July 2015 - 42
Systems, Man & Cybernetics - July 2015 - 43
Systems, Man & Cybernetics - July 2015 - 44
Systems, Man & Cybernetics - July 2015 - 45
Systems, Man & Cybernetics - July 2015 - 46
Systems, Man & Cybernetics - July 2015 - 47
Systems, Man & Cybernetics - July 2015 - 48
Systems, Man & Cybernetics - July 2015 - 49
Systems, Man & Cybernetics - July 2015 - 50
Systems, Man & Cybernetics - July 2015 - 51
Systems, Man & Cybernetics - July 2015 - 52
Systems, Man & Cybernetics - July 2015 - 53
Systems, Man & Cybernetics - July 2015 - 54
Systems, Man & Cybernetics - July 2015 - 55
Systems, Man & Cybernetics - July 2015 - 56
Systems, Man & Cybernetics - July 2015 - 57
Systems, Man & Cybernetics - July 2015 - 58
Systems, Man & Cybernetics - July 2015 - 59
Systems, Man & Cybernetics - July 2015 - 60
Systems, Man & Cybernetics - July 2015 - 61
Systems, Man & Cybernetics - July 2015 - 62
Systems, Man & Cybernetics - July 2015 - 63
Systems, Man & Cybernetics - July 2015 - 64
Systems, Man & Cybernetics - July 2015 - 65
Systems, Man & Cybernetics - July 2015 - 66
Systems, Man & Cybernetics - July 2015 - 67
Systems, Man & Cybernetics - July 2015 - 68
Systems, Man & Cybernetics - July 2015 - 69
Systems, Man & Cybernetics - July 2015 - 70
Systems, Man & Cybernetics - July 2015 - 71
Systems, Man & Cybernetics - July 2015 - 72
Systems, Man & Cybernetics - July 2015 - Cover3
Systems, Man & Cybernetics - July 2015 - Cover4
https://www.nxtbook.com/nxtbooks/ieee/smc_202310
https://www.nxtbook.com/nxtbooks/ieee/smc_202307
https://www.nxtbook.com/nxtbooks/ieee/smc_202304
https://www.nxtbook.com/nxtbooks/ieee/smc_202301
https://www.nxtbook.com/nxtbooks/ieee/smc_202210
https://www.nxtbook.com/nxtbooks/ieee/smc_202207
https://www.nxtbook.com/nxtbooks/ieee/smc_202204
https://www.nxtbook.com/nxtbooks/ieee/smc_202201
https://www.nxtbook.com/nxtbooks/ieee/smc_202110
https://www.nxtbook.com/nxtbooks/ieee/smc_202107
https://www.nxtbook.com/nxtbooks/ieee/smc_202104
https://www.nxtbook.com/nxtbooks/ieee/smc_202101
https://www.nxtbook.com/nxtbooks/ieee/smc_202010
https://www.nxtbook.com/nxtbooks/ieee/smc_202007
https://www.nxtbook.com/nxtbooks/ieee/smc_202004
https://www.nxtbook.com/nxtbooks/ieee/smc_202001
https://www.nxtbook.com/nxtbooks/ieee/smc_201910
https://www.nxtbook.com/nxtbooks/ieee/smc_201907
https://www.nxtbook.com/nxtbooks/ieee/smc_201904
https://www.nxtbook.com/nxtbooks/ieee/smc_201901
https://www.nxtbook.com/nxtbooks/ieee/smc_201810
https://www.nxtbook.com/nxtbooks/ieee/smc_201807
https://www.nxtbook.com/nxtbooks/ieee/smc_201804
https://www.nxtbook.com/nxtbooks/ieee/smc_201801
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_1017
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0717
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0417
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0117
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_1016
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0716
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0416
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0116
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_1015
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0715
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0415
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0115
https://www.nxtbookmedia.com