A Sigma-Pi-Sigma Neural Network (SPSNN) |
| |
Authors: | Li Chien-Kuo |
| |
Affiliation: | (1) Department of Information Management, Shih Chien University, Taipei, Taiwan, ROC |
| |
Abstract: | This letter presents a sigma-pi-sigma neural network (SPSNN) structure. The SPSNN can learn to implement static mapping that
multilayer neural networks and radial basis function networks usually do. The output of the SPSNN has the sum of product-of-sum
form
, where x
j's are inputs, N
v
is the number of inputs, f
nij() is a function to be generated through the network training, and K is the number of pi-sigma network (PSN) which is the
basic building block for SPSNN. A linear memory array can be used to implement f
nij
(). The function f
nij
(x
j
) can be expressed as
, where B
ijk() is a single-variable basis function, w
nijk's are weight values stored in memory, N
q
is the quantized element number for x
j
, and N
e
is the number of basis functions in the neighborhood used for storing information for x
j. If all B
ijk()'s are Gaussian functions, the new neural network degenerates to a Gaussian function network. This paper focuses on the
use of overlapped rectangular pulses as the basis functions. With such basis functions,
will equal either zero or w
nijk, and the computation of f
nij (x
j) becomes a simple addition of retrieved w
nijk's. The new neural network structure demonstrates excellent learning convergence characteristics and requires small memory
space. It has merits over multilayer neural networks, radial basis function networks and CMAC.
This revised version was published online in June 2006 with corrections to the Cover Date. |
| |
Keywords: | function approximation memory-based neural network ridge polynomial network self-generated basis function sigma-pi-sigma neural network |
本文献已被 SpringerLink 等数据库收录! |
|