Extreme learning machines are
feedforward neural network
A feedforward neural network (FNN) is an artificial neural network wherein connections between the nodes do ''not'' form a cycle. As such, it is different from its descendant: recurrent neural networks.
The feedforward neural network was the ...
s for
classification Classification is a process related to categorization, the process in which ideas and objects are recognized, differentiated and understood.
Classification is the grouping of related facts into classes.
It may also refer to:
Business, organizat ...
,
regression
Regression or regressions may refer to:
Science
* Marine regression, coastal advance due to falling sea level, the opposite of marine transgression
* Regression (medicine), a characteristic of diseases to express lighter symptoms or less extent ( ...
,
clustering,
sparse approximation, compression and
feature learning with a single layer or multiple layers of hidden nodes, where the parameters of hidden nodes (not just the weights connecting inputs to hidden nodes) need to be tuned. These hidden nodes can be randomly assigned and never updated (i.e. they are
random projection but with nonlinear transforms), or can be inherited from their ancestors without being changed. In most cases, the output weights of hidden nodes are usually learned in a single step, which essentially amounts to learning a linear model. The name "extreme learning machine" (ELM) was given to such models by its main inventor Guang-Bin Huang.
According to their creators, these models are able to produce good generalization performance and learn thousands of times faster than networks trained using
backpropagation. In literature, it also shows that these models can outperform
support vector machine
In machine learning, support vector machines (SVMs, also support vector networks) are supervised learning models with associated learning algorithms that analyze data for classification and regression analysis. Developed at AT&T Bell Laboratorie ...
s in both classification and regression applications.
History
From 2001-2010, ELM research mainly focused on the unified learning framework for "generalized" single-hidden layer feedforward neural networks (SLFNs), including but not limited to sigmoid networks, RBF networks, threshold networks, trigonometric networks, fuzzy inference systems, Fourier series,
Laplacian transform, wavelet networks, etc. One significant achievement made in those years is to successfully prove the universal approximation and classification capabilities of ELM in theory.
From 2010 to 2015, ELM research extended to the unified learning framework for kernel learning, SVM and a few typical feature learning methods such as
Principal Component Analysis
Principal component analysis (PCA) is a popular technique for analyzing large datasets containing a high number of dimensions/features per observation, increasing the interpretability of data while preserving the maximum amount of information, and ...
(PCA) and
Non-negative Matrix Factorization (NMF). It is shown that SVM actually provides suboptimal solutions compared to ELM, and ELM can provide the whitebox kernel mapping, which is implemented by ELM random feature mapping, instead of the blackbox kernel used in SVM. PCA and NMF can be considered as special cases where linear hidden nodes are used in ELM.
From 2015 to 2017, an increased focus has been placed on hierarchical implementations
of ELM. Additionally since 2011, significant biological studies have been made that support certain ELM theories.
From 2017 onwards, to overcome low-convergence problem during training
LU decomposition,
Hessenberg decomposition and
QR decomposition based approaches with
regularization have begun to attract attention
In a 2017 announcement from
Google Scholar:
Classic Papers: Articles That Have Stood The Test of Time, two ELM papers have been listed in the
Top 10 in Artificial Intelligence for 2006" taking positions 2 and 7.
Algorithms
Given a single hidden layer of ELM, suppose that the output function of the
-th hidden node is
, where
and
are the parameters of the
-th hidden node. The output function of the ELM for single hidden layer feedforward networks (SLFN) with
hidden nodes is:
, where
is the output weight of the
-th hidden node.