Abstract

It is easy for a multi-layered perception (MLP) to form open plane classification borders, and for a radial basis function network (RBFN) to form closed circular or elliptic classification borders. In contrast, it is difficult for a MLP to form closed circular or elliptic classification borders, and for RBFN to form open plane classification borders. Hence, MLP and RBFN have their own advantages and disadvantages in dealing with various classification problems. To combine their advantages, in this paper, we proposed a novel neural network, Hybrid Transfer Function Network (HTFN), whose hidden layer contains sigmoid and Gaussian units at the same time. Although there are two kinds of processing units in HTFN, in this study, we used the principle of minimizing error sum of squares to derive the supervised learning rules for all the network parameters. When HTFN contains only either sigmoid units or Gaussian units in its hidden layer, HTFN can be transferred into MLP and RBFN, respectively. Hence, MLP and RBFN can be considered as a special case of HTFN. To verify that HTFN is superior to MLP and RBFN, this study employed three man-made examples to test the three networks. The results showed that HTFN is more accurate than MLP and RBFN, confirming that combining sigmoid and Gaussian units into hidden layer can combine advantages of MLP and RBFN.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.