Prior knowledge for pattern recognition: Difference between revisions

Content deleted Content added
m Reverted edits by 222.154.232.180 to last version by 203.39.32.102 (HG)
Protez (talk | contribs)
mNo edit summary
Line 4:
== Definition ==
 
Prior knowledge, as<ref>B. definedScholkopf inand [Scholkopf02]A. Smola, "Learning with Kernels", MIT Press 2002.</ref> refers to all information about the problem available in addition to the training data. However, in this most general form, determining a [[Model (abstract)|model]] from a finite set of samples without prior knowledge is an [[ill-posed]] problem, in the sense that a unique model may not exist. Many classifiers incorporate the general smoothness assumption that a test pattern similar to one of the training samples tends to be assigned to the same class.
 
The importance of prior knowledge in machine learning is suggested by its role in search and optimization. Loosely, the [[No free lunch in search and optimization|no free lunch theorem]] states that all search algorithms have the same average performance over all problems, and thus implies that to gain in performance on a certain application one must use a specialized algorithm that includes some prior knowledge about the problem. <!-- This sentence is still not right. Read the "no free lunch" article to see why.
Line 54:
Prior knowledge on these can enhance the quality of the recognition if included in the learning. Moreover, not taking into account the poor quality of some data or a large imbalance between the classes can mislead the decision of a classifier.
 
== ReferencesNotes ==
 
<references/>
* [Scholkopf02], B. Scholkopf and A. Smola, "Learning with Kernels", MIT Press 2002.
== References ==
 
* [Krupka07], E. Krupka and N. Tishby, "Incorporating Prior Knowledge on Features into Learning", Eleventh International Conference on Artificial Intelligence and Statistics (AISTATS 07)
 
[[Category:Machine learning]]