Content deleted Content added
m Dating maintenance tags: {{Citation needed}} |
Marcocapelle (talk | contribs) →External links: Removed grandparent category of Category:Reliability engineering Tags: Mobile edit Mobile web edit Advanced mobile edit |
||
(8 intermediate revisions by 7 users not shown) | |||
Line 1:
{{more cn|date=September 2022}}
'''Robustification''' is a form of [[optimisation]] whereby a [[system]] is made less sensitive to the effects of [[Random variable|random variability]], or [[Noisy data|noise]], that is present in that system's input variables and [[parameter]]s. The process is typically associated with [[engineering systems]], but the process can also be applied to a [[political policy]], a [[Business Strategy|business strategy]] or any other system that is subject to the effects of random
==Clarification on definition==
Line 11:
===Non-linearities===
Consider the [[Graph (discrete mathematics)|graph]] below of a relationship between an input variable ''x'' and the output ''Y'', for which it is desired that a value of 7 is taken, of a system of interest. It can be seen that there are two possible values that ''x'' can take, 5 and 30. If the tolerance for ''x'' is independent of the nominal value, then it can also be seen that when ''x'' is set equal to 30, the expected variation of ''Y'' is less than if ''x'' were set equal to 5. The reason is that the gradient at ''x'' = 30 is less than at ''x'' = 5, and the random variability in ''x'' is suppressed as it flows to ''Y''.
[[File:Robustification.JPG]]
This basic principle underlies all robustification, but in practice there are typically a number of inputs and it is the suitable point with the lowest gradient on a multi-dimensional surface that must be found.
===Non-constant variability===
Line 22:
''Z'' = ''x y''
For any target value of ''Z'' there is an infinite number of combinations for the nominal values of ''x'' and ''y'' that will be suitable. However, if the standard deviation of ''x'' was proportional to the nominal value and the standard deviation of ''y'' was constant, then ''x'' would be reduced (to limit the random variability that will flow from the right hand side of the equation to the left hand side) and ''y'' would be increased (with no expected increase random variability because the standard deviation is constant) to bring the value of ''Z'' to the target value. By doing this, ''Z'' would have the desired nominal value and it would be expected that its standard deviation would be at a minimum: robustified.
By taking advantage of the two principles covered above, one is able to optimise a system so that the nominal value of a systems output is kept at its desired level while also minimising the likelihood of any deviation from that nominal value. This is despite the presence of random variability within the input variables.
==Methods==
There are three distinct methods of robustification, but a
===Experimental===
The experimental approach is probably the most widely known. It involves the identification of those variables that can be adjusted and those variables that are treated as [[Statistical noise|noises]]. An experiment is then designed to investigate how changes to the nominal value of the adjustable variables can limit the transfer of noise from the noise variables to the output. This approach is attributed to [[Genichi Taguchi|Taguchi]] and is often associated with [[Taguchi methods]]. While many have found the approach to provide impressive results, the techniques have also been criticised for being statistically erroneous and inefficient. Also, the time and effort required can be significant.
Another experimental method that was used for robustification is the Operating Window. It was developed in the [[United States]] before the wave of quality methods from [[Japan]] came to the [[Western world|West]], but still remains unknown to many.<ref name="Clausing">See Clausing (2004) reference for more details</ref> In this approach, the noise of the inputs is continually increased as the system is modified to reduce sensitivity to that noise. This increases robustness, but also provides a clearer measure of the variability that is flowing through the system. After optimisation, the random variability of the inputs is controlled and reduced, and the system exhibits improved quality.
===Analytical===
Line 40:
===Numerical===
In the numerical approach a model is run a number of times as part of a [[Monte Carlo method|Monte Carlo simulation]] or a numerical propagation of errors to predict the variability of the outputs. Numerical optimisation methods such as hill climbing or evolutionary algorithms are then used to find the optimum nominal values for the inputs. This approach typically requires less human time and effort than the other two, but it can be very demanding on computational resources during simulation and optimization.
==See also==
Line 46:
==Footnotes==
{{reflist}}
==References==
Line 55:
==External links==
* {{usurped|1=[https://web.archive.org/web/20070314202927/http://grassmannalgebra.info/probabilisticdesign/ Probabilistic design]}}
[[Category:Reliability engineering]]
|