- Volume 21 Issue 1
In this paper, a higher-order feedforward neural network called ridge polynomial network (RPN) which shows good approximation capability for nonlnear continuous functions defined on compact subsets in multi-dimensional Euclidean spaces, is presented. This network provides more efficient and regular structure as compared to ordinary higher-order feedforward networks based on Gabor-Kolmogrov polynomial expansions, while maintating their fast learning property. the ridge polynomial network is a generalization of the pi-sigma network (PSN) and uses a specialform of ridge polynomials. It is shown that any multivariate polynomial can be exactly represented in this form, and thus realized by a RPN. The approximation capability of the RPNs for arbitrary continuous functions is shown by this representation theorem and the classical weierstrass polynomial approximation theorem. The RPN provides a natural mechanism for incremental function approximation based on learning algorithm of the PSN. Simulation results on several applications such as multivariate function approximation and pattern classification assert nonlinear approximation capability of the RPN.