<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://en.formulasearchengine.com/index.php?action=history&amp;feed=atom&amp;title=Characteristic_function_%28convex_analysis%29</id>
	<title>Characteristic function (convex analysis) - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://en.formulasearchengine.com/index.php?action=history&amp;feed=atom&amp;title=Characteristic_function_%28convex_analysis%29"/>
	<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/index.php?title=Characteristic_function_(convex_analysis)&amp;action=history"/>
	<updated>2026-04-18T07:43:06Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.43.0-wmf.28</generator>
	<entry>
		<id>https://en.formulasearchengine.com/index.php?title=Characteristic_function_(convex_analysis)&amp;diff=15207&amp;oldid=prev</id>
		<title>en&gt;Helpful Pixie Bot: ISBNs (Build KC)</title>
		<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/index.php?title=Characteristic_function_(convex_analysis)&amp;diff=15207&amp;oldid=prev"/>
		<updated>2012-05-05T23:40:03Z</updated>

		<summary type="html">&lt;p&gt;ISBNs (Build KC)&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{refimprove|date=May 2010}}&lt;br /&gt;
&lt;br /&gt;
In [[numerical analysis]], &amp;#039;&amp;#039;&amp;#039;fixed-point iteration&amp;#039;&amp;#039;&amp;#039; is a method of computing [[fixed point (mathematics)|fixed points]] of [[iterated function]]s.&lt;br /&gt;
&lt;br /&gt;
More specifically, given a  function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; defined on the [[real number]]s with real values and given a point &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; in the [[domain of a function|domain]] of &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;, the fixed point [[iteration]] is&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;x_{n+1}=f(x_n), \, n=0, 1, 2, \dots&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which gives rise to the [[sequence]] &amp;lt;math&amp;gt;x_0, x_1, x_2, \dots&amp;lt;/math&amp;gt; which is hoped to [[limit (mathematics)|converge]] to a point &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is continuous, then one can prove that the obtained &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; is a fixed point of &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;, i.e.,&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;f(x)=x&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
More generally, the function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; can be defined on any [[metric space]] with values in that same space.&lt;br /&gt;
&lt;br /&gt;
==Examples==&lt;br /&gt;
&lt;br /&gt;
* A first simple and useful example is the [[Babylonian method]] for computing the [[square root]] of &amp;#039;&amp;#039;a&amp;#039;&amp;#039;&amp;gt;0, which consists in taking &amp;lt;math&amp;gt;f(x)=\frac 12\left(\frac ax + x\right)&amp;lt;/math&amp;gt;, i.e. the mean value of &amp;#039;&amp;#039;x&amp;#039;&amp;#039; and &amp;#039;&amp;#039;a/x&amp;#039;&amp;#039;, to approach the limit &amp;lt;math&amp;gt;x = \sqrt a&amp;lt;/math&amp;gt; (from whatever starting point &amp;lt;math&amp;gt;x_0 \gg 0 &amp;lt;/math&amp;gt;). This is a special case of [[Newton&amp;#039;s method]] quoted below.&lt;br /&gt;
&lt;br /&gt;
[[File:Sine fixed point.svg|250px|thumb|The fixed-point iteration &amp;#039;&amp;#039;x&amp;#039;&amp;#039;&amp;lt;sub&amp;gt;&amp;#039;&amp;#039;n&amp;#039;&amp;#039;+1&amp;lt;/sub&amp;gt;&amp;amp;nbsp;=&amp;amp;nbsp;sin &amp;#039;&amp;#039;x&amp;#039;&amp;#039;&amp;lt;sub&amp;gt;&amp;#039;&amp;#039;n&amp;#039;&amp;#039;&amp;lt;/sub&amp;gt; with initial value &amp;#039;&amp;#039;x&amp;#039;&amp;#039;&amp;lt;sub&amp;gt;0&amp;lt;/sup&amp;gt; = 2 converges to 0. This example does not satisfy the assumptions of the [[Banach fixed point theorem]] and so its speed of convergence is very slow.]]&lt;br /&gt;
&lt;br /&gt;
* The fixed-point iteration &amp;lt;math&amp;gt;x_{n+1}=\cos x_n\,&amp;lt;/math&amp;gt; converges to the unique fixed point of the function &amp;lt;math&amp;gt;f(x)=\cos x\,&amp;lt;/math&amp;gt; for any starting point &amp;lt;math&amp;gt;x_0.&amp;lt;/math&amp;gt; This example does satisfy the assumptions of the [[Banach fixed point theorem]]. Hence, the error after n steps satisfies &amp;lt;math&amp;gt;|x_n-x_0| \leq { q^n \over 1-q } | x_1 - x_0 | = C q^n&amp;lt;/math&amp;gt; (where we can take &amp;lt;math&amp;gt;q = 0.85&amp;lt;/math&amp;gt;, if we start from &amp;lt;math&amp;gt;x_0=1&amp;lt;/math&amp;gt;.) When the error is less than a multiple of &amp;lt;math&amp;gt;q^n&amp;lt;/math&amp;gt; for some constant &amp;#039;&amp;#039;q&amp;#039;&amp;#039;, we say that we have [[linear convergence]]. The Banach fixed-point theorem allows one to obtain fixed-point iterations with linear convergence.&lt;br /&gt;
&lt;br /&gt;
* The fixed-point iteration &amp;lt;math&amp;gt;x_{n+1}=2x_n\,&amp;lt;/math&amp;gt; will diverge unless &amp;lt;math&amp;gt;x_0=0&amp;lt;/math&amp;gt;. We say that the fixed point of &amp;lt;math&amp;gt;f(x)=2x\,&amp;lt;/math&amp;gt; is repelling.&lt;br /&gt;
&lt;br /&gt;
* The requirement that &amp;#039;&amp;#039;f&amp;#039;&amp;#039; is continuous is important, as the following example shows. The iteration&lt;br /&gt;
:&amp;lt;math&amp;gt; x_{n+1} = &lt;br /&gt;
\begin{cases}&lt;br /&gt;
\frac{x_n}{2}, &amp;amp; x_n \ne 0\\&lt;br /&gt;
1, &amp;amp; x_n=0&lt;br /&gt;
\end{cases}&amp;lt;/math&amp;gt;&lt;br /&gt;
converges to 0 for all values of  &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt;.&lt;br /&gt;
However, 0 is &amp;#039;&amp;#039;not&amp;#039;&amp;#039; a fixed point of the function&lt;br /&gt;
:&amp;lt;math&amp;gt;f(x) = &lt;br /&gt;
\begin{cases}&lt;br /&gt;
\frac{x}{2}, &amp;amp; x \ne 0\\&lt;br /&gt;
1, &amp;amp; x = 0&lt;br /&gt;
\end{cases}&amp;lt;/math&amp;gt;&lt;br /&gt;
as this function is &amp;#039;&amp;#039;not&amp;#039;&amp;#039; continuous at &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;, and in fact has no fixed points.&lt;br /&gt;
&lt;br /&gt;
==Applications==&lt;br /&gt;
* [[Newton&amp;#039;s method]] for a given differentiable function &amp;lt;math&amp;gt;f(x)&amp;lt;/math&amp;gt; is &amp;lt;math&amp;gt;x_{n+1}=x_n-\frac{f(x_n)}{f&amp;#039;(x_n)}&amp;lt;/math&amp;gt;. If we write &amp;lt;math&amp;gt;g(x)=x-\frac{f(x)}{f&amp;#039;(x)}&amp;lt;/math&amp;gt; we may rewrite the Newton iteration as the fixed point iteration &amp;lt;math&amp;gt;x_{n+1}=g(x_n)&amp;lt;/math&amp;gt;. If this iteration converges to a fixed point &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; of &amp;lt;math&amp;gt;g&amp;lt;/math&amp;gt; then &amp;lt;math&amp;gt;x=g(x)=x-\frac{f(x)}{f&amp;#039;(x)}&amp;lt;/math&amp;gt; so &amp;lt;math&amp;gt;\frac{f(x)}{f&amp;#039;(x)}=0&amp;lt;/math&amp;gt;. The inverse of anything is nonzero, therefore &amp;lt;math&amp;gt;f(x)=0&amp;lt;/math&amp;gt;: x is a &amp;#039;&amp;#039;root&amp;#039;&amp;#039; of f. Under the assumptions of the [[Banach fixed point theorem]], the Newton iteration, framed as the fixed point method, demonstrates [[linear convergence]]. However, a more detailed analysis shows [[quadratic convergence]], i.e., &amp;lt;math&amp;gt;|x_n-x|&amp;lt;Cq^{2^n}&amp;lt;/math&amp;gt;, under certain circumstances.&lt;br /&gt;
&lt;br /&gt;
* [[Halley&amp;#039;s method]] is similar to [[Newton&amp;#039;s method]] but, when it works correctly, its error is &amp;lt;math&amp;gt;|x_n-x|&amp;lt;Cq^{3^n}&amp;lt;/math&amp;gt; ([[cubic convergence]]). In general, it is possible to design methods that converge with speed &amp;lt;math&amp;gt;Cq^{k^n}&amp;lt;/math&amp;gt; for any &amp;lt;math&amp;gt;k\in \Bbb N&amp;lt;/math&amp;gt;. As a general rule, the higher the &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt;, the less stable it is, and the more computationally expensive it gets. For these reasons, higher order methods are typically not used.&lt;br /&gt;
&lt;br /&gt;
* [[Runge-Kutta method]]s and numerical [[Ordinary Differential Equation]] solvers in general can be viewed as fixed point iterations. Indeed, the core idea when analyzing the [[A-stability]] of ODE solvers is to start with the special case &amp;lt;math&amp;gt;y&amp;#039;=ay&amp;lt;/math&amp;gt;, where a is a [[complex number]], and to check whether the ODE solver converges to the fixed point &amp;lt;math&amp;gt;y=0&amp;lt;/math&amp;gt; whenever the real part of a is negative.&amp;lt;ref&amp;gt;One may also consider certain iterations A-stable if the iterates stay bounded for a long time, which is beyond the scope of this article.&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* The [[Picard–Lindelöf theorem]], which shows that ordinary differential equations have solutions, is essentially an application of the [[Banach fixed point theorem]] to a special sequence of functions which forms a fixed point iteration.&lt;br /&gt;
&lt;br /&gt;
* The [[goal seeking]] function in Excel can be used to find solutions to the [[Colebrook equation]] to an accuracy of 15 significant figures.&amp;lt;ref&amp;gt;M A Kumar (2010), Solve Implicit Equations (Colebrook) Within Worksheet, Createspace, ISBN 1-4528-1619-0&lt;br /&gt;
&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Some of the &amp;quot;successive approximation&amp;quot; schemes used in [[dynamic programming]] to solve  [[Bellman equation|Bellman&amp;#039;s functional equation]] are based on fixed point iterations in the space of the return function.&amp;lt;ref&amp;gt;Bellman, R. (1957). Dynamic programming, Princeton University Press.&amp;lt;/ref&amp;gt;&amp;lt;ref&amp;gt;Sniedovich, M. (2010). Dynamic Programming: Foundations and Principles, [[Taylor &amp;amp; Francis]].&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Properties==&lt;br /&gt;
&lt;br /&gt;
* If a function &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; defined on the real line with real values is [[Lipschitz continuity|Lipschitz continuous]] with Lipschitz constant &amp;lt;math&amp;gt;L&amp;lt;1&amp;lt;/math&amp;gt;, then this function has precisely one fixed point, and the fixed-point iteration converges towards that fixed point for any initial guess &amp;lt;math&amp;gt;x_0.&amp;lt;/math&amp;gt; This theorem can be generalized to any metric space.&lt;br /&gt;
&lt;br /&gt;
:Proof of this theorem:&lt;br /&gt;
&lt;br /&gt;
:Since &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt; is Lipschitz continuous with Lipschitz constant &amp;lt;math&amp;gt;L&amp;lt;1&amp;lt;/math&amp;gt;, then for the sequence &amp;lt;math&amp;gt;\{x_n,n=0,1,2...\}&amp;lt;/math&amp;gt;, we have:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;|x_2-x_1|=|f(x_1)-f(x_0)|\leq L|x_1-x_0|&amp;lt;/math&amp;gt;,&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;|x_3-x_2|=|f(x_2)-f(x_1)|\leq L|x_2-x_1|&amp;lt;/math&amp;gt;,&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;\cdots&amp;lt;/math&amp;gt;,&lt;br /&gt;
&lt;br /&gt;
:and&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;|x_n-x_{n-1}|=|f(x_{n-1})-f(x_{n-2})|\leq L|x_{n-1}-x_{n-2}|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
:Combining the above inequalities yields:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;|x_n-x_{n-1}|\leq L^{n-1}|x_1-x_0|&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
:Since &amp;lt;math&amp;gt;L&amp;lt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;L^{n-1}\rightarrow 0&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;n \rightarrow \infty&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
:Therefore, we can show  &amp;lt;math&amp;gt;\{x_n\}&amp;lt;/math&amp;gt; is a [[Cauchy sequence]] and thus it converges to a point &amp;lt;math&amp;gt;x^*&amp;lt;/math&amp;gt;. &lt;br /&gt;
&lt;br /&gt;
:For the iteration &amp;lt;math&amp;gt;x_n=f(x_{n-1})&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; go to infinity on both sides of the equation, we obtain &amp;lt;math&amp;gt;x^*=f(x^*)&amp;lt;/math&amp;gt;. This shows that &amp;lt;math&amp;gt;x^*&amp;lt;/math&amp;gt; is the fixed point for &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. So we proved the iteration will eventually converge to a fixed-point.&lt;br /&gt;
&lt;br /&gt;
:This property is very useful because not all iterations can arrive at a convergent fixed-point. When constructing a fixed-point iteration, it is very important to make sure it converges. There are several [[fixed-point theorem]]s to guarantee the existence of the fixed point, but since the iteration function is continuous, we can usually use the above theorem to test if an iteration converges or not. The proof of the generalized theorem to metric space is similar.&lt;br /&gt;
&lt;br /&gt;
* The speed of convergence of the iteration sequence can be increased by using a [[convergence acceleration]] method such as [[Aitken&amp;#039;s delta-squared process]]. The application of Aitken&amp;#039;s method to fixed-point iteration is known as [[Steffensen&amp;#039;s method]], and it can be shown that Steffensen&amp;#039;s method yields a rate of convergence that is at least quadratic.&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;br /&gt;
* {{cite book| last1=Burden | first1=Richard L. | last2=Faires | first2=J. Douglas &lt;br /&gt;
| title=Numerical Analysis | publisher=PWS Publishers | edition=3rd | isbn=0-87150-857-5 &lt;br /&gt;
| year=1985 | chapter=2.2 Fixed-Point Iteration}}.&lt;br /&gt;
&lt;br /&gt;
==See also==&lt;br /&gt;
&lt;br /&gt;
* [[Root-finding algorithm]]&lt;br /&gt;
* [[Fixed-point theorem]]&lt;br /&gt;
* [[Fixed-point combinator]]&lt;br /&gt;
* [[Banach fixed-point theorem]]&lt;br /&gt;
* [[Cobweb plot]]&lt;br /&gt;
* [[Markov chain]]&lt;br /&gt;
&lt;br /&gt;
==External links==&lt;br /&gt;
*[http://algonum.appspot.com/#fixpoint.picard Fixed-point algorithms online]&lt;br /&gt;
*[http://www.maccery.com/maths/fixed_point.php Fixed-point iteration online calculator]&lt;br /&gt;
*[http://user.mendelu.cz/marik/maw/index.php?lang=en&amp;amp;form=banach Fixed-point iteration online calculator (Mathematical Assistant on Web)]&lt;br /&gt;
&lt;br /&gt;
{{DEFAULTSORT:Fixed-Point Iteration}}&lt;br /&gt;
[[Category:Root-finding algorithms]]&lt;br /&gt;
[[Category:Iterative methods]]&lt;/div&gt;</summary>
		<author><name>en&gt;Helpful Pixie Bot</name></author>
	</entry>
</feed>