Content deleted Content added
categorization/tagging using AWB |
|||
Line 1:
A '''Probabilistic Neural Network (PNN)''' is a [[Feedforward neural network]] , which was derived from [[Bayesian network]]<ref>http://herselfsai.com/2007/03/probabilistic-neural-networks.html</ref> and a statistical algorithm called [[Kernel Fisher discriminant analysis]].<ref>http://www.psi.toronto.edu/~vincent/research/presentations/PNN.pdf</ref>
* Input layer
* Hidden layer
Line 15:
For PNN networks there is one pattern neuron for each category of the target variable. The actual target category of each training case is stored with each hidden neuron; the weighted value coming out of a hidden neuron is fed only to the pattern neuron that corresponds to the hidden neuron’s category. The pattern neurons add the values for the class they represent
PNN often use in classification problems.<ref>http://www.mathworks.in/help/toolbox/nnet/ug/bss38ji-1.html</ref>
===Output layer===
The output layer compares the weighted votes for each target category accumulated in the pattern layer and uses the largest vote to predict the target category.
== Advantages==
There are several advantages and disadvantages using PNN instead of [[multilayer perceptron]].<ref>http://www.dtreg.com/pnn.htm</ref>
* PNN is much faster compare to multilayer perceptron networks.
Line 46:
== References ==
{{Reflist}}
{{Uncategorized|date=April 2012}}
|