Inverse function theorem: Difference between revisions

Content deleted Content added
Ac44ck (talk | contribs)
Clarified that the example is for a vector-valued function in text and image. The first sentence of the page should make that clear, but it may not be obvious to those who jump to the example.
Thric3 (talk | contribs)
mNo edit summary
Line 1:
In [[mathematics]], the '''inverse function theorem''' gives sufficient conditions for a [[vector-valued function]] to be [[invertible]] on an [[open region]] containing a point in its ___domain. The theorem can be generalized to maps defined on [[manifold|manifolds]], and on infinite dimensional [[Banach space]]s. Loosely, a ''[[smooth function|C<sup>1</sup>]]'' function ''F'' is invertible at a point ''p'' if its [[Jacobian]] ''J<sub>F</sub>(p)'' is invertible.
 
More precisely, the theorem states that if the [[total derivative]] of a [[continuously differentiable]] function ''F'' defined from an open set U of '''R'''<sup>''n''</sup> into '''R'''<sup>''n''</sup> is invertible at a point ''p'' (i.e., the [[Jacobian determinant]] determinant of ''F'' at ''p'' is nonzero), then F is an invertible function near ''p''. That is, an [[inverse function]] to ''F'' exists in some [[neighbourhood (mathematics)|neighborhood]] of ''F''(''p''). Moreover, the inverse function ''F<sup>-1</sup>'' is also continuously differentiable. In the infinite dimensional case it is required that the [[Frechet derivative]] have a [[bounded linear map|bounded]] inverse near ''p''.
 
The Jacobian matrix of ''F''<sup>-1</sup> at ''F''(''p'') is then the inverse of the Jacobian of ''F'', evaluated at ''p''. This can be understood as a special case of the [[chain rule]], which states that for [[linear transformations]] ''f'' and ''g'',