Content deleted Content added
expanded definition some more, said that algo looks for smallest inco solution, but this may cost time |
{{1r}} |
||
(29 intermediate revisions by 15 users not shown) | |||
Line 1:
{{mi|{{Nf|date=November 2024}}{{1r|date=November 2024}}}}
In [[constraint satisfaction problem|constraint satisfaction]] [[backtracking]] [[algorithm]]s, '''constraint learning''' is a
==Definition==
Backtracking algorithms work by choosing an
If the partial solution <math>x_1=a_1,\ldots,x_k=a_k</math> is inconsistent, the problem instance implies the constraint stating that <math>x_i=a_i</math> cannot
On the other hand, if a subset of this evaluation is inconsistent, the corresponding constraint may be useful in the subsequent search, as the same subset of the partial evaluation may occur again in the search. For example, the algorithm may
{| cellpadding=20
The efficiency of constraint learning algorithm is balanced between two factors. On one hand, the more often a recorded constraint is violated, the more often block backtracking from doing useless search. As a result, algorithms search for small inconsistent subset of the current partial solution, as they correspond to constraints that are easier to violate. On the other hand, finding a small inconsistent subset of the current partial evaluation may require time, and the benefit may not be balanced by the subsequent reduction of the search time.▼
|-
| [[Image:Constraint-learning-1.svg]]
| [[Image:Constraint-learning-2.svg]]
| [[Image:Constraint-learning-3.svg]]
|-
| Search has reached a dead end.
| Inconsistency may be caused by the values of <math>x_1</math> and <math>x_4</math> only. This fact can be stored in a new constraint.
| If the algorithm reaches the same values of <math>x_1</math> and <math>x_4</math> again, the new constraint blocks the search.
|}
==Efficiency of constraint learning==
▲The efficiency gain of constraint learning
Size is however not the only feature of learned constraints to take into account. Indeed, a small constraint may be useless in a particular state of the search space because the values that violate it will not be encountered again. A larger constraint whose violating values are more similar to the current partial assignment may be preferred in such cases.
Various constraint learning techniques exist, differing in strictness of recorded constraints and cost of finding them.
==Graph-based learning==
If the algorithm proves all values of <math>x_{k+1}</math> to be inconsistent with <math>x_1=a_1,\ldots,x_k=a_k</math>, then this evaluation was consistent, as otherwise the algorithm would not have evaluated <math>x_{k+1}</math> at all; as a result, the constraints violated by a value of <math>x_{k+1}</math> together with <math>x_1=a_1,\ldots,x_k=a_k</math> all contain <math>x_{k+1}</math>.
As a result, an inconsistent evaluation is the restriction of the truth evaluation of <math>x_1,\ldots,x_k</math> to variables that are in a constraint with <math>x_{k+1}</math>, provided that this constraint contains no unassigned variable.
Learning constraints representing these partial evaluation is called graph-based learning. It uses the same rationale of [[graph-based backjumping]]. These methods are called "graph-based" because they are based on pairs of variables in the same constraint, which can be found from the graph associated to the constraint satisfaction problem.
==Jumpback learning==
Jumpback learning is based on storing as constraints the inconsistent assignments that would be found by [[conflict-based backjumping]]. Whenever a partial assignment is found inconsistent, this algorithm selects the violated constraint that is minimal according to an ordering based on the order of instantiation of variables. The evaluation restricted to the variables that are in this constraint is inconsistent and is usually shorter than the complete evaluation. Jumpback learning stores this fact as a new constraint.
The ordering on constraints is based on the order of assignment of variables. In particular, the least of two constraints is the one whose latest non-common variable has been instantiated first. When an inconsistent assignment is reached, jumpback learning selects the violated constraint that is minimal according to this ordering, and restricts the current assignment to its variables. The constraint expressing the inconsistency of this assignment is stored.
==Constraint maintenance==
Constraint learning algorithms differ not only on the choice of constraint corresponding to a given inconsistent partial evaluation, but also on the choice of which constraints they retain and which ones they discard.
In general, learning all inconsistencies in the form of constraints and keeping them indefinitely may exhaust the available memory and increase the cost of checking consistency of partial evaluations. These problems can be solved either by storing only some learned constraints or by occasionally discarding constraints.
''Bounded learning'' only stores constraints if the inconsistent partial evaluation they represent is smaller than a given constraint number. ''Relevance-bounded learning'' discards constraints (or does not store them at all) that are considered not relevant given the current point of the search space; in particular, it discards or does not store all constraints that represent inconsistent partial evaluations that differ from the current partial evaluation on no more than a given fixed number of variables.
==See also==
Line 16 ⟶ 56:
*[[Backjumping]]
==
*{{
|
|
|authorlink = Rina Dechter
|
|
|
|
}} {{ISBN
[[Category:
|