Content deleted Content added
m Fix links to disambiguation page Transformation |
Flod logic (talk | contribs) m Spelling/grammar/punctuation/typographical correction |
||
(23 intermediate revisions by 16 users not shown) | |||
Line 1:
[[Pattern recognition]] is a very active field of research intimately bound to [[machine learning]]. Also known as classification or [[statistical classification]], pattern recognition aims at building a [[classifier (mathematics)|classifier]] that can determine the class of an input pattern. This procedure, known as training, corresponds to learning an unknown decision function based only on a set of input-output pairs <math>(\boldsymbol{x}_i,y_i)</math> that form the training data (or training set). Nonetheless, in real world applications such as [[character recognition]], a certain amount of information on the problem is usually known beforehand. The incorporation of this prior knowledge into the training is the key element that will allow an increase of performance in many applications.▼
▲[[Pattern recognition]] is a very active field of research intimately bound to [[machine learning]]. Also known as classification or [[statistical classification]], pattern recognition aims at building a [[classifier]] that can determine the class of an input pattern. This procedure, known as training, corresponds to learning an unknown decision function based only on a set of input-output pairs <math>(\boldsymbol{x}_i,y_i)</math> that form the training data (or training set). Nonetheless, in real world applications such as [[character recognition]], a certain amount of information on the problem is usually known beforehand. The incorporation of this prior knowledge into the training is the key element that will allow an increase of performance in many applications.
==
Prior knowledge
The importance of prior knowledge in machine learning is suggested by its role in search and optimization. Loosely, the [[No free lunch in search and optimization|no free lunch theorem]] states that all search algorithms have the same average performance over all problems, and thus implies that to gain in performance on a certain application one must use a specialized algorithm that includes some prior knowledge about the problem. <!-- This sentence is still not right. Read the "no free lunch" article to see why.
Line 16 ⟶ 15:
A very common type of prior knowledge in pattern recognition is the invariance of the class (or the output of the classifier) to a [[Transformation (geometry)|transformation]] of the input pattern. This type of knowledge is referred to as '''transformation-invariance'''. The mostly used transformations used in image recognition are:
* [[Translation (geometry)|translation]];
* [[Rotation (mathematics)|rotation]];
* [[Shear mapping|skewing]];
* [[Scaling (geometry)|scaling]].
Incorporating the invariance to a transformation <math>T_{\theta}: \boldsymbol{x} \mapsto T_{\theta}\boldsymbol{x}</math> parametrized in <math>\theta</math> into a classifier of output <math>f(\boldsymbol{x})</math> for an input pattern <math>\boldsymbol{x}</math> corresponds to
:<math>
f(\boldsymbol{x}) = f(T_{\theta}\boldsymbol{x}), \quad \forall \boldsymbol{x}, \theta .</math>
Local invariance can also be considered for a transformation centered at <math>\theta=0</math>, so that <math>T_0\boldsymbol{x} = \boldsymbol{x}</math>, by using the constraint
:<math>
\left.\frac{\partial}{\partial \theta}\right|_{\theta=0} f(T_{\theta} \boldsymbol{x}) = 0 .
</math>
Another approach is to consider
:<math>
f(\boldsymbol{x}) = y_{\mathcal{P}},\ \forall \boldsymbol{x}\in \mathcal{P} ,
</math>
where <math>y_{\mathcal{P}}</math> is the membership class of the region <math>\mathcal{P}</math> of the input space.
A different type of class-invariance found in pattern recognition is
== Knowledge
Other forms of prior knowledge than class-invariance concern the data more specifically and are thus of particular interest for real-world applications. The three particular cases that most often occur when gathering data are:
* '''Unlabeled samples''' are available with supposed class-memberships;
* '''Imbalance''' of the training set due to a high proportion of samples of a class;
* '''Quality of the data''' may vary from a sample to another.
Prior knowledge
== Notes ==
<references/>
== References ==
* E. Krupka and N. Tishby, "[https://proceedings.mlr.press/v2/krupka07a.html Incorporating Prior Knowledge on Features into Learning]", Eleventh International Conference on Artificial Intelligence and Statistics (AISTATS 07)
[[Category:Machine learning]]
|