Content deleted Content added
GreenC bot (talk | contribs) Move 1 url. Wayback Medic 2.5 |
|||
(42 intermediate revisions by 33 users not shown) | |||
Line 1:
The '''winnow algorithm'''
▲The '''winnow algorithm''' <ref>Littlestone, N. (1988) "Learning Quickly When Irrelevant Attributes About: A New Linear-threshold Algorithm" Machine Learning 285-318(2)</ref> is a technique from machine learning. It is closely related to the [[Perceptron]], but it uses a multiplicative weight-update scheme that allows it perform much better than the perceptron when many dimensions are irrelevant (hence its name). It is not a sophisticated algorithm but it scales well to high-dimensional spaces. During training, winnow is shown a sequence of positive and negative examples. From these it learns a decision hyperplane.
==
The basic algorithm, Winnow1, is as follows. The instance space is <math>X=\{0,1\}^n</math>, that is, each instance is described as a set of [[Boolean-valued]] [[features (pattern recognition)|features]]. The algorithm maintains non-negative weights <math>w_i</math> for <math>i\in \{1
▲The instance space is <math>X=\{0,1\}^n</math>. The algorithm maintains non-negative weights <math>w_i</math> for <math>i\in \{1...n\}</math> which are initially set to 1. When the learner is given an example <math>(x_1,...x_n)</math>, the learner follows the following prediction rule:
* '''If''' <math>\sum_{i=1}^n w_i x_i > \Theta </math>, '''then'''
* '''Otherwise'''
For each example with which it is presented, the learner applies the following update rule:
* If an example is correctly classified, do nothing.
* If an example is predicted
*: <math>\forall x_{i} = 1, w_{i} = 0</math>
* If an example is predicted
*: <math>\forall x_{i} = 1, w_{i} = \alpha w_{i}</math>
A
==Mistake bounds==
In certain circumstances, it can be shown that the number of mistakes Winnow makes as it learns has an [[Upper and lower bounds|upper bound]] that is independent of the number of instances with which it is presented. If the Winnow1 algorithm uses <math>\alpha > 1</math> and <math>\Theta \geq 1/\alpha</math> on a target function that is a <math>k</math>-literal monotone disjunction given by <math>f(x_1,\ldots,x_n)=x_{i_1}\cup \cdots \cup x_{i_k}</math>, then for any sequence of instances the total number of mistakes is bounded by:
<math>\alpha k ( \log_\alpha \Theta+1)+\frac{n}{\Theta}</math>.<ref>
Nick Littlestone (1989). "Mistake bounds and logarithmic linear-threshold learning algorithms".
Technical report UCSC-CRL-89-11, University of California, Santa Cruz.</ref>
==References==
<references/>
[[Category:Classification algorithms]]
|