英文摘要 |
This study proposes adaptive probabilistic neural networks (APNN), which include three kinds of parameters: the variable weights representing the importance of input variables, the core-width-reciprocal representing the effective range of patterns, and the data weights representing the reliability of patterns. This study proposes a algorithm to adapt these parameters, and maximize the accuracy of the model by optimizing these parameters in the learning process. In order to prove the performance of APNN, three artificial function mapping problems as well as an actual classification problem are employed to test it and compare it with a back-propagation network (BPN) and a probabilistic neural network (PNN). The results proved that the accuracy of APNN is only slightly lower than BPN, and is strongly superior to PNN. Furthermore, the variable weights of APNN really express the importance of input variables for output variables, which provides the model with explanation abilities. |