tu-logo ur-logo. Outline. Outline. 1. Neural Networks. Single-Hidden Layer Feedforward Networks (SLFNs). Conventional ...... support vector machines. ELM Web ...
Outline
Extreme Learning Machine for MultiCategories Classification Applications Hai-Jun Rong1,2 , Guang-Bin Huang1 and Yew-Soon Ong2 1 School
of Electrical and Electronic Engineering
2 School
of Computer Engineering Nanyang Technological University Nanyang Avenue, Singapore 639798 E-mail: {hjrong, egbhuang, asysong}@ntu.edu.sg
IEEE World Congress on Computational Intelligence Hong Kong, June 1-6 2008 ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Outline
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Outline
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Outline
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Outline
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Outline
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Feedforward Neural Networks with Additive Nodes
Output of hidden nodes G(ai , bi , x) = g(ai · x + bi )
(1)
ai : the weight vector connecting the ith hidden node and the input nodes. bi : the threshold of the ith hidden node.
Output of SLFNs
fL (x) =
L X
βi G(ai , bi , x)
(2)
i=1
Figure 1:
Feedforward Network Architecture: additive hidden
tu-logo βi : the weight vector connecting the ith hidden node and the output nodes.
nodes ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Feedforward Neural Networks with Additive Nodes
Output of hidden nodes G(ai , bi , x) = g(ai · x + bi )
(1)
ai : the weight vector connecting the ith hidden node and the input nodes. bi : the threshold of the ith hidden node.
Output of SLFNs
fL (x) =
L X
βi G(ai , bi , x)
(2)
i=1
Figure 1:
Feedforward Network Architecture: additive hidden
tu-logo βi : the weight vector connecting the ith hidden node and the output nodes.
nodes ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Feedforward Neural Networks with Additive Nodes
Output of hidden nodes G(ai , bi , x) = g(ai · x + bi )
(1)
ai : the weight vector connecting the ith hidden node and the input nodes. bi : the threshold of the ith hidden node.
Output of SLFNs
fL (x) =
L X
βi G(ai , bi , x)
(2)
i=1
Figure 1:
Feedforward Network Architecture: additive hidden
tu-logo βi : the weight vector connecting the ith hidden node and the output nodes.
nodes ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Feedforward Neural Networks with RBF Nodes
Output of hidden nodes G(ai , bi , x) = g (bi kx − ai k)
(3)
ai : the center of the ith hidden node. bi : the impact factor of the ith hidden node.
Output of SLFNs
fL (x) =
L X
βi G(ai , bi , x)
(4)
i=1
Figure 2:
Feedforward Network Architecture: RBF hidden
nodes
tu-logo βi : the weight vector connecting the ith hidden node and the output nodes. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Feedforward Neural Networks with RBF Nodes
Output of hidden nodes G(ai , bi , x) = g (bi kx − ai k)
(3)
ai : the center of the ith hidden node. bi : the impact factor of the ith hidden node.
Output of SLFNs
fL (x) =
L X
βi G(ai , bi , x)
(4)
i=1
Figure 2:
Feedforward Network Architecture: RBF hidden
nodes
tu-logo βi : the weight vector connecting the ith hidden node and the output nodes. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Feedforward Neural Networks with RBF Nodes
Output of hidden nodes G(ai , bi , x) = g (bi kx − ai k)
(3)
ai : the center of the ith hidden node. bi : the impact factor of the ith hidden node.
Output of SLFNs
fL (x) =
L X
βi G(ai , bi , x)
(4)
i=1
Figure 2:
Feedforward Network Architecture: RBF hidden
nodes
tu-logo βi : the weight vector connecting the ith hidden node and the output nodes. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Function Approximation of Neural Networks
Learning Model For N arbitrary distinct samples (xi , ti ) ∈ Rn × Rm , SLFNs with L hidden nodes and activation function g(x) are mathematically modeled as fL (xj ) = oj , j = 1, · · · , N
Cost function: E =
Figure 3:
PN
j=1
(5)
‚ ‚ ‚ ‚ ‚oj − tj ‚ .
The target is to minimize the cost function E by adjusting the network parameters: βi , ai , bi .
2
tu-logo
Feedforward Network Architecture. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Function Approximation of Neural Networks
Learning Model For N arbitrary distinct samples (xi , ti ) ∈ Rn × Rm , SLFNs with L hidden nodes and activation function g(x) are mathematically modeled as fL (xj ) = oj , j = 1, · · · , N
Cost function: E =
Figure 3:
PN
j=1
(5)
‚ ‚ ‚ ‚ ‚oj − tj ‚ .
The target is to minimize the cost function E by adjusting the network parameters: βi , ai , bi .
2
tu-logo
Feedforward Network Architecture. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Function Approximation of Neural Networks
Learning Model For N arbitrary distinct samples (xi , ti ) ∈ Rn × Rm , SLFNs with L hidden nodes and activation function g(x) are mathematically modeled as fL (xj ) = oj , j = 1, · · · , N
Cost function: E =
Figure 3:
PN
j=1
(5)
‚ ‚ ‚ ‚ ‚oj − tj ‚ .
The target is to minimize the cost function E by adjusting the network parameters: βi , ai , bi .
2
tu-logo
Feedforward Network Architecture. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Function Approximation of Neural Networks
Learning Model For N arbitrary distinct samples (xi , ti ) ∈ Rn × Rm , SLFNs with L hidden nodes and activation function g(x) are mathematically modeled as fL (xj ) = oj , j = 1, · · · , N
Cost function: E =
Figure 3:
PN
j=1
(5)
‚ ‚ ‚ ‚ ‚oj − tj ‚ .
The target is to minimize the cost function E by adjusting the network parameters: βi , ai , bi .
2
tu-logo
Feedforward Network Architecture. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Learning Algorithms of Neural Networks
Learning Methods Many learning methods mainly based on gradient-descent/iterative approaches have been developed over the past two decades. Back-Propagation (BP) and its variants are most popular.
Figure 4:
tu-logo
Feedforward Network Architecture. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Learning Algorithms of Neural Networks
Learning Methods Many learning methods mainly based on gradient-descent/iterative approaches have been developed over the past two decades. Back-Propagation (BP) and its variants are most popular.
Figure 4:
tu-logo
Feedforward Network Architecture. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Learning Algorithms of Neural Networks
Learning Methods Many learning methods mainly based on gradient-descent/iterative approaches have been developed over the past two decades. Back-Propagation (BP) and its variants are most popular.
Figure 4:
tu-logo
Feedforward Network Architecture. ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Advantagnes and Disadvantages Popularity Widely used in various applications: regression, classification, etc. Limitations Usually different learning algorithms used in different SLFNs architectures. Some parameters have to be tuned mannually. tu-logo
Overfitting. Local minima.
ur-logo
Time-consuming. ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
SLFN Models Learning Methods
Advantagnes and Disadvantages Popularity Widely used in various applications: regression, classification, etc. Limitations Usually different learning algorithms used in different SLFNs architectures. Some parameters have to be tuned mannually. tu-logo
Overfitting. Local minima.
ur-logo
Time-consuming. ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) New Learning Theory If continuous target function f (x) can be approximated by SLFNs with adjustable hidden nodes then the hidden node parameters of such SLFNs needn’t be tuned. Instead, all these hidden node parameters can be randomly generated without the knowledge of the training data. Given any nonconstant piecewise continuous function g, for any continuous target function f and any randomly generated sequence {(ai , bi )Li=1 }, lim kf (x) − fL (x)k = 0
L→∞
Figure 5:
Feedforward Network Architecture: any type of
G(ai , bi , x).
holds with probability one if βi is chosen to tu-logo minimize the kf (x) − fL (x)k, i = 1, · · · , L.
G.-B. Huang, et al., “Universal Approximation Using Incremental Constructive Feedforward Networks with Random ur-logo Hidden Nodes,” IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Convex Incremental Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) New Learning Theory If continuous target function f (x) can be approximated by SLFNs with adjustable hidden nodes then the hidden node parameters of such SLFNs needn’t be tuned. Instead, all these hidden node parameters can be randomly generated without the knowledge of the training data. Given any nonconstant piecewise continuous function g, for any continuous target function f and any randomly generated sequence {(ai , bi )Li=1 }, lim kf (x) − fL (x)k = 0
L→∞
Figure 5:
Feedforward Network Architecture: any type of
G(ai , bi , x).
holds with probability one if βi is chosen to tu-logo minimize the kf (x) − fL (x)k, i = 1, · · · , L.
G.-B. Huang, et al., “Universal Approximation Using Incremental Constructive Feedforward Networks with Random ur-logo Hidden Nodes,” IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Convex Incremental Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Unified Learning Platform
Mathematical Model For N arbitrary distinct samples (xi , ti ) ∈ Rn × Rm , standard SLFNs with L hidden nodes and output function g(x) are mathematically modeled as L X
βi G(ai , bi , xj ) = tj , j = 1, · · · , N
i=1
(6) (ai , bi ): Hidden node parameters.
Figure 6:
Feedforward Network Architecture: any type of
tu-logo βi : the weight vector connecting the ith hidden node and the output node.
G(ai , bi , x). ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Unified Learning Platform
Mathematical Model For N arbitrary distinct samples (xi , ti ) ∈ Rn × Rm , standard SLFNs with L hidden nodes and output function g(x) are mathematically modeled as L X
βi G(ai , bi , xj ) = tj , j = 1, · · · , N
i=1
(6) (ai , bi ): Hidden node parameters.
Figure 6:
Feedforward Network Architecture: any type of
tu-logo βi : the weight vector connecting the ith hidden node and the output node.
G(ai , bi , x). ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) Mathematical Model PL
i=1
βi G(ai , bi , xj ) = tj , j = 1, · · · , N is equivalent to Hβ = T, where H(a1 , · · · , aL , b1 , · · · , bL , x1 , · · · , xN ) 3 ··· G(aL , bL , x1 ) 7 . 7 . 5 ··· .
G(a1 , b1 , x1 ) 6 . =6 . 4 . G(a1 , b1 , xN ) 2
2 6 6 β =6 4
β1T . . . βL T
···
G(aL , bL , xN )
3
2
7 7 7 5
6 6 and T = 6 4
L×m
(7)
N×L
3 tT 1 7 . 7 . 7 . 5 tT N N×m
(8)
H is called the hidden layer output matrix of the neural network; the ith column of H is the output of the ith hidden node with respect to inputs x1 , x2 , · · · , xN .
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Outline 1
Neural Networks Single-Hidden Layer Feedforward Networks (SLFNs) Conventional Learning Algorithms of SLFNs
2
Extreme Learning Machine Unified Learning Platform ELM Algorithm
3
ELM for Multi-Categories Classification Problems
4
Performance Evaluations
5
Summary
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
tu-logo
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) Three-Step Learning Model Given a training set ℵ = {(xi , ti )|xi ∈ Rn , ti ∈ Rm , i = 1, · · · , N}, activation function g, and the number of hidden nodes L, 1
Assign randomly hidden node parameters (ai , bi ), i = 1, · · · , L.
2
Calculate the hidden layer output matrix H.
3
Calculate the output weight β: β = H† T.
where H† is the Moore-Penrose generalized inverse of hidden layer output matrix H. tu-logo
Source Codes of ELM http://www.ntu.edu.sg/home/egbhuang/
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) Three-Step Learning Model Given a training set ℵ = {(xi , ti )|xi ∈ Rn , ti ∈ Rm , i = 1, · · · , N}, activation function g, and the number of hidden nodes L, 1
Assign randomly hidden node parameters (ai , bi ), i = 1, · · · , L.
2
Calculate the hidden layer output matrix H.
3
Calculate the output weight β: β = H† T.
where H† is the Moore-Penrose generalized inverse of hidden layer output matrix H. tu-logo
Source Codes of ELM http://www.ntu.edu.sg/home/egbhuang/
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) Three-Step Learning Model Given a training set ℵ = {(xi , ti )|xi ∈ Rn , ti ∈ Rm , i = 1, · · · , N}, activation function g, and the number of hidden nodes L, 1
Assign randomly hidden node parameters (ai , bi ), i = 1, · · · , L.
2
Calculate the hidden layer output matrix H.
3
Calculate the output weight β: β = H† T.
where H† is the Moore-Penrose generalized inverse of hidden layer output matrix H. tu-logo
Source Codes of ELM http://www.ntu.edu.sg/home/egbhuang/
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) Three-Step Learning Model Given a training set ℵ = {(xi , ti )|xi ∈ Rn , ti ∈ Rm , i = 1, · · · , N}, activation function g, and the number of hidden nodes L, 1
Assign randomly hidden node parameters (ai , bi ), i = 1, · · · , L.
2
Calculate the hidden layer output matrix H.
3
Calculate the output weight β: β = H† T.
where H† is the Moore-Penrose generalized inverse of hidden layer output matrix H. tu-logo
Source Codes of ELM http://www.ntu.edu.sg/home/egbhuang/
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
Extreme Learning Machine (ELM) Three-Step Learning Model Given a training set ℵ = {(xi , ti )|xi ∈ Rn , ti ∈ Rm , i = 1, · · · , N}, activation function g, and the number of hidden nodes L, 1
Assign randomly hidden node parameters (ai , bi ), i = 1, · · · , L.
2
Calculate the hidden layer output matrix H.
3
Calculate the output weight β: β = H† T.
where H† is the Moore-Penrose generalized inverse of hidden layer output matrix H. tu-logo
Source Codes of ELM http://www.ntu.edu.sg/home/egbhuang/
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
ur-logo
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Unified Learning Platform ELM Algorithm
ELM Learning Algorithm
Salient Features “Simple Math is Enough.” ELM is a simple tuning-free three-step algorithm. The learning speed of ELM is extremely fast. Unlike the conventional learning methods which MUST see the training data before generating the hidden node parameters, ELM could generate the hidden node parameters before seeing the training data. Unlike the traditional classic gradient-based learning algorithms which only work for differentiable activation functions, ELM works for all bounded nonconstant piecewise continuous activation functions including non-differential activation functions. Unlike the traditional classic gradient-based learning algorithms facing several issues like local minima, improper learning rate and overfitting, etc, ELM tends to reach the solutions straightforward without such trivial issues. The ELM learning algorithm looks much simpler than many learning algorithms: neural networks and support vector machines.
tu-logo
ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
ELM for Multi-Categories Classification Problems Three basic methods 1
Single ELM classifier: m output nodes of ELM for m-class applications. We say x is in class l if output node l has the highest output value.
2
One-Against-All ELM (ELM-OAA): m-class classification problems are implemented by m binary ELM classifiers, each of which is trained independently to classify one of the m pattern classes.
3
One-Against-One ELM (ELM-OAO): the m pattern classes are pairwise decomposed into m(m − 1)/2 two different classes, and each of them is trained by one binary ELM classifier. tu-logo
Exponential loss based decoding approach used in ELM-OAA and ELM-OAO E. L. Allwein, et al., “Reducing multiclass to binary: a unifying approach for margin classifiers,” Journal of Machine ur-logo
Learning Research, vol. 1, pp. 113-141, 2001.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
ELM for Multi-Categories Classification Problems Three basic methods 1
Single ELM classifier: m output nodes of ELM for m-class applications. We say x is in class l if output node l has the highest output value.
2
One-Against-All ELM (ELM-OAA): m-class classification problems are implemented by m binary ELM classifiers, each of which is trained independently to classify one of the m pattern classes.
3
One-Against-One ELM (ELM-OAO): the m pattern classes are pairwise decomposed into m(m − 1)/2 two different classes, and each of them is trained by one binary ELM classifier. tu-logo
Exponential loss based decoding approach used in ELM-OAA and ELM-OAO E. L. Allwein, et al., “Reducing multiclass to binary: a unifying approach for margin classifiers,” Journal of Machine ur-logo
Learning Research, vol. 1, pp. 113-141, 2001.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
ELM for Multi-Categories Classification Problems Three basic methods 1
Single ELM classifier: m output nodes of ELM for m-class applications. We say x is in class l if output node l has the highest output value.
2
One-Against-All ELM (ELM-OAA): m-class classification problems are implemented by m binary ELM classifiers, each of which is trained independently to classify one of the m pattern classes.
3
One-Against-One ELM (ELM-OAO): the m pattern classes are pairwise decomposed into m(m − 1)/2 two different classes, and each of them is trained by one binary ELM classifier. tu-logo
Exponential loss based decoding approach used in ELM-OAA and ELM-OAO E. L. Allwein, et al., “Reducing multiclass to binary: a unifying approach for margin classifiers,” Journal of Machine ur-logo
Learning Research, vol. 1, pp. 113-141, 2001.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
ELM for Multi-Categories Classification Problems Three basic methods 1
Single ELM classifier: m output nodes of ELM for m-class applications. We say x is in class l if output node l has the highest output value.
2
One-Against-All ELM (ELM-OAA): m-class classification problems are implemented by m binary ELM classifiers, each of which is trained independently to classify one of the m pattern classes.
3
One-Against-One ELM (ELM-OAO): the m pattern classes are pairwise decomposed into m(m − 1)/2 two different classes, and each of them is trained by one binary ELM classifier. tu-logo
Exponential loss based decoding approach used in ELM-OAA and ELM-OAO E. L. Allwein, et al., “Reducing multiclass to binary: a unifying approach for margin classifiers,” Journal of Machine ur-logo
Learning Research, vol. 1, pp. 113-141, 2001.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
ELM for Multi-Categories Classification Problems Three basic methods 1
Single ELM classifier: m output nodes of ELM for m-class applications. We say x is in class l if output node l has the highest output value.
2
One-Against-All ELM (ELM-OAA): m-class classification problems are implemented by m binary ELM classifiers, each of which is trained independently to classify one of the m pattern classes.
3
One-Against-One ELM (ELM-OAO): the m pattern classes are pairwise decomposed into m(m − 1)/2 two different classes, and each of them is trained by one binary ELM classifier. tu-logo
Exponential loss based decoding approach used in ELM-OAA and ELM-OAO E. L. Allwein, et al., “Reducing multiclass to binary: a unifying approach for margin classifiers,” Journal of Machine ur-logo
Learning Research, vol. 1, pp. 113-141, 2001.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Table 1: Specification of Real-World Classification Benchmark Problems Type
Datasets
Type I
Glass Vehicle Page Blocks Image Segmentation Satellite Image Shuttle DNA Optical Recognition of Handwritten Digits Pen-Based Recognition of Handwritten Digits Cancer Arrhythmia Letter Recognition
Type II
# Attributes
# Classes
9 18 10 19 36 9 180 64 16 98 279 16
6 4 5 7 6 7 3 10 10 14 16 26
# Observations Training Testing 110 104 420 426 2,700 2,773 1,100 1,210 4,400 2,035 43,500 14,500 1,000 1,186 3,823 1,797 7,494 3,498 144 46 300 152 10,000 10,000
tu-logo
ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Table 2: Comparison of network complexity of ELM-OAA, ELM and ELM-OAO Datasets Glass Vehicle Page Blocks Image Segmentation Satellite Image Shuttle DNA Handwritten Optical Recognition Handwritten Pen-Based Recognition Cancer Arrhythmia Letter Recognition
ELM-OAA 30 90 130 200 450 260 350 420
ELM 30 110 160 210 470 340 450 470
ELM-OAO 20 50 50 30 150 80 290 110
650
690
190
40 50 1100
50 70 1500
10 20 510
tu-logo
ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Table 3: Comparison of testing accuracy of ELM-OAA, ELM and ELM-OAO Datasets Glass Vehicle Page Blocks Image Segmentation Satellite Image Shuttle DNA Handwritten Optical Recognition Handwritten Pen-Based Recognition Cancer Arrhythmia Letter Recognition
ELM-OAA 66.346 79.513 95.873 95.216 89.670 99.667 94.866 96.928
ELM 64.423 79.489 95.761 94.667 89.663 99.715 94.828 96.948
ELM-OAO 65.385 81.378 95.397 94.493 89.955 99.581 94.496 96.343
98.294
98.302
97.819
78.652 65.441 93.417
77.522 65.395 93.029
73.000 61.770 93.055
tu-logo
ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
Table 4: Comparison of training and testing time (seconds) of ELM-OAA, ELM and ELM-OAO Datasets Glass Vehicle Page Blocks Image Segmentation Satellite Image Shuttle DNA Handwritten Optical Recognition Handwritten Pen-Based Recognition Cancer Arrhythmia Letter Recognition
ELM-OAA (s) Training Testing 0.0250 0.0617 0.1508 0.0617 1.7125 0.5023 3.3547 0.2953 43.838 1.6500 151.88 7.4594 7.0797 0.6227 39.002 2.0734 140.17 5.3305 0.0491 0.0191 0.1250 0.0586 2295.6 73.006
ELM (s) Training Testing 0.0117 0.0773 0.0680 0.0531 0.6641 0.2070 0.6000 0.0906 8.4320 0.2758 36.614 1.6516 5.0328 0.2359 6.7414 0.3461 18.127 0.7742 0.0072 0.0194 0.0266 0.0273 76.066 1.9700
ELM-OAO (s) Training Testing 0.0133 0.0219 0.0547 0.0313 0.3422 0.2836 0.1453 0.1102 5.1039 1.0383 30.395 4.8422 4.7070 0.4578 4.2766 2.5727 17.923 7.3648 0.0469 0.0422 0.1414 0.2281 1429.5 279.57
tu-logo
ur-logo
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
Summary ELM, ELM-OAO and ELM-OAA obtain similar testing accuracies. ELM-OAO usually requires smaller number of hidden nodes than the single ELM classifier and ELM-OAA. The training time required by ELM-OAO is similar or less than ELM and ELM-OAA when the number of pattern classes is small (say, not larger than 10). However when the number of pattern classes is large (say, tu-logo larger than 10), the training time cost by ELM-OAO is most likely higher than the single ELM classifier but still smaller ur-logo than ELM-OAA. ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo
Neural Networks ELM ELM for Multi-Categories Classification Problems Performance Evaluations Summary
References
References G.-B. Huang, et al., “Universal Approximation Using Incremental Networks with Random Hidden Computational Nodes”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 879-892, 2006. G.-B. Huang, et al., “Extreme Learning Machine: Theory and Applications,” Neurocomputing, vol. 70, pp. 489-501, 2006. G.-B. Huang, et al., “Convex Incremental Extreme Learning Machine,” Neurocomputing, vol. 70, pp. 3056-3062, 2007. M.-B. Li, et al., “Fully complex extreme learning machine,” Neurocomputing, vol. 68, pp. 306-314, 2005. N.-Y. Liang, et al., “A Fast and Accurate On-line Sequential Learning Algorithm for Feedforward Networks,” IEEE Transactions on Neural Networks, vol. 17, no. 6, pp. 1411-1423, 2006. Q.-Y. Zhu, et al., “Evolutionary Extreme Learning Machine”, Pattern Recognition, vol. 38, no. 10, pp. 1759-1763, 2005. G.-B. Huang, et al., “Can Threshold Networks Be Trained Directly?” IEEE Transactions on Circuits and Systems II, vol. 53, no. 3, pp. 187-191, 2006. G.-B. Huang, et al., “Real-Time Learning Capability of Neural Networks”, IEEE Transactions on Neural Networks, vol. 17, no. 4, pp. 863-878, 2006.
ELM Web Portal: www.ntu.edu.sg/home/egbhuang
Extreme Learning Machines
tu-logo
ur-logo