Alternating series: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
mNo edit summary
No edit summary
 
(199 intermediate revisions by more than 100 users not shown)
Line 1: Line 1:
{{Short description|Infinite series whose terms alternate in sign}}
In [[mathematics]], an '''alternating series''' is an [[infinite series]] of the form
{{More citations needed|date=January 2010}}
{{Calculus |Series}}


In [[mathematics]], an '''alternating series''' is an [[infinite series]] of the form
:<math>\sum_{n=0}^\infty (-1)^n\,a_n,</math>
<math display="block">\sum_{n=0}^\infty (-1)^n a_n</math> or <math display="block">\sum_{n=0}^\infty (-1)^{n+1} a_n</math>
with {{math|''a<sub>n</sub>'' > 0}} for all&nbsp;{{mvar|n}}. The signs of the general terms alternate between positive and negative. Like any series, an alternating [[Convergent series|series converges]] if and only if the associated sequence of partial sums [[Limit of a sequence|converges]].


==Examples==
with ''a<sub>n</sub>'' &ge; 0. A ''sufficient'' condition for the [[series (mathematics)|series]] to converge is that it [[absolute convergence|converges absolutely]]. But this is often too strong a condition to ask:it is not ''necessary''. For example, the [[harmonic series (mathematics)|harmonic series]]
The geometric series [[1/2 − 1/4 + 1/8 − 1/16 + ⋯|{{sfrac|1|2}} − {{sfrac|1|4}} + {{sfrac|1|8}} − {{sfrac|1|16}} + ⋯]] sums to {{sfrac|1|3}}.


The [[harmonic series (mathematics)#Alternating harmonic series|alternating harmonic series]] has a finite sum but the [[harmonic series (mathematics)|harmonic series]] does not.
:<math>\sum_{n=0}^\infty \frac1n,</math>


The [[Mercator series]] provides an analytic expression of the [[natural logarithm]]:
<math display="block"> \sum_{n=1}^\infty \frac{(-1)^{n+1}}{n} x^n \;=\; \ln (1+x).</math>

The functions sine and cosine used in [[trigonometry]] can be defined as alternating series in [[calculus]] even though they are introduced in elementary algebra as the ratio of sides of a right triangle. In fact,
<math display="block">\sin x = \sum_{n=0}^\infty (-1)^n \frac{x^{2n+1}}{(2n+1)!},</math> and
<math display="block">\cos x = \sum_{n=0}^\infty (-1)^n \frac{x^{2n}}{(2n)!} .</math>
When the alternating factor {{math|(–1)<sup>''n''</sup>}} is removed from these series one obtains the [[hyperbolic function]]s sinh and cosh used in calculus.

For integer or positive index α the [[Bessel function]] of the first kind may be defined with the alternating series
<math display="block"> J_\alpha(x) = \sum_{m=0}^\infty \frac{(-1)^m}{m! \, \Gamma(m+\alpha+1)} {\left(\frac{x}{2}\right)}^{2m+\alpha} </math> where {{math|Γ(''z'')}} is the [[gamma function]].

If {{mvar|s}} is a [[complex number]], the [[Dirichlet eta function]] is formed as an alternating series
<math display="block">\eta(s) = \sum_{n=1}^{\infty}{(-1)^{n-1} \over n^s} = \frac{1}{1^s} - \frac{1}{2^s} + \frac{1}{3^s} - \frac{1}{4^s} + \cdots</math>
that is used in [[analytic number theory]].

== Alternating series test ==
{{main|Alternating series test}}

The theorem known as "Leibniz Test" or the [[alternating series test]] tells us that an alternating series will converge if the terms {{math|''a<sub>n</sub>''}} converge to 0 [[monotonic function|monotonically]].

Proof: Suppose the sequence <math>a_n</math> converges to zero and is monotone decreasing. If <math>m</math> is odd and <math>m<n</math>, we obtain the estimate <math>S_n - S_m \le a_{m}</math> via the following calculation:
<math display="block">\begin{align}
S_n - S_m & =
\sum_{k=0}^n(-1)^k\,a_k\,-\,\sum_{k=0}^m\,(-1)^k\,a_k\ = \sum_{k=m+1}^n\,(-1)^k\,a_k \\
& =a_{m+1} - a_{m+2} + a_{m+3} - a_{m+4} + \cdots + a_n\\
& = a_{m+1}-(a_{m+2}-a_{m+3}) - (a_{m+4}-a_{m+5}) - \cdots - a_n \le a_{m+1} \le a_{m}.
\end{align}</math>

Since <math>a_n</math> is monotonically decreasing, the terms <math>-(a_m - a_{m+1})</math> are negative. Thus, we have the final inequality: <math>S_n - S_m \le a_m</math>. Similarly, it can be shown that <math>-a_m \le S_n - S_m </math>. Since <math>a_m</math> converges to <math>0</math>, our partial sums <math>S_m</math> form a [[Cauchy sequence]] (i.e., the series satisfies the [[Cauchy criterion]]) and therefore converge. The argument for <math>m</math> even is similar.

== Approximating sums ==
The estimate above does not depend on <math>n</math>. So, if <math>a_n</math> is approaching 0 monotonically, the estimate provides an [[error bound]] for approximating infinite sums by partial sums:
<math display="block">\left|\sum_{k=0}^\infty(-1)^k\,a_k\,-\,\sum_{k=0}^m\,(-1)^k\,a_k\right|\le |a_{m+1}|.</math>That does not mean that this estimate always finds the very first element after which error is less than the modulus of the next term in the series. Indeed if you take <math>1-1/2+1/3-1/4+... = \ln 2</math> and try to find the term after which error is at most 0.00005, the inequality above shows that the partial sum up through <math>a_{20000}</math> is enough, but in fact this is twice as many terms as needed. Indeed, the error after summing first 9999 elements is 0.0000500025, and so taking the partial sum up through <math>a_{10000}</math> is sufficient. This series happens to have the property that constructing a new series with <math>a_n -a_{n+1}</math> also gives an alternating series where the Leibniz test applies and thus makes this simple error bound not optimal. This was improved by the Calabrese bound,<ref>{{Cite journal |last=Calabrese |first=Philip |date=March 1962 |title=A Note on Alternating Series |url=https://www.jstor.org/stable/2311056 |journal=The American Mathematical Monthly |volume=69 |issue=3 |pages=215–217 |doi=10.2307/2311056|jstor=2311056 }}</ref> discovered in 1962, that says that this property allows for a result 2 times less than with the Leibniz error bound. In fact this is also not optimal for series where this property applies 2 or more times, which is described by [[Richard Johnsonbaugh|Johnsonbaugh]] error bound.<ref>{{Cite journal |last=Johnsonbaugh |first=Richard |date=October 1979 |title=Summing an Alternating Series |url=https://www.jstor.org/stable/2321292 |journal=The American Mathematical Monthly |volume=86 |issue=8 |pages=637–648 |doi=10.2307/2321292|jstor=2321292 }}</ref> If one can apply the property an infinite number of times, [[Series acceleration#Euler's transform|Euler's transform]] applies.<ref>{{cite arXiv |last=Villarino |first=Mark B. |date=2015-11-27 |title=The error in an alternating series |class=math.CA |eprint=1511.08568 }}</ref>

== Absolute convergence ==
A series <math display=inline>\sum a_n</math> [[absolute convergence|converges absolutely]] if the series <math display=inline>\sum |a_n|</math> converges.

Theorem: Absolutely convergent series are convergent.

Proof: Suppose <math display=inline>\sum a_n</math> is absolutely convergent. Then, <math display=inline>\sum |a_n|</math> is convergent and it follows that <math display=inline>\sum 2|a_n|</math> converges as well. Since <math display=inline> 0 \leq a_n + |a_n| \leq 2|a_n|</math>, the series <math display=inline>\sum (a_n + |a_n|)</math> converges by the [[Direct comparison test|comparison test]]. Therefore, the series <math display=inline>\sum a_n</math> converges as the difference of two convergent series <math display=inline>\sum a_n = \sum (a_n + |a_n|) - \sum |a_n|</math>.

== Conditional convergence ==
A series is [[Conditional convergence|conditionally convergent]] if it converges but does not converge absolutely.

For example, the [[harmonic series (mathematics)|harmonic series]]
<math display="block">\sum_{n=1}^\infty \frac{1}{n}, </math>
diverges, while the alternating version
diverges, while the alternating version
<math display="block">\sum_{n=1}^\infty \frac{(-1)^{n+1}}{n}, </math>
converges by the [[Alternating series#Alternating series test|alternating series test]].

== Rearrangements ==
For any series, we can create a new series by rearranging the order of summation. A series is [[Series (mathematics)#Unconditionally convergent series|unconditionally convergent]] if any rearrangement creates a series with the same convergence as the original series. [[Absolute convergence#Rearrangements and unconditional convergence|Absolutely convergent series are unconditionally convergent]]. But the [[Riemann series theorem]] states that conditionally convergent series can be rearranged to create arbitrary convergence.<ref>{{cite journal |last1=Mallik |first1=AK |year=2007 |title=Curious Consequences of Simple Sequences |journal=Resonance |volume=12 |issue=1 |pages=23–37 |doi=10.1007/s12045-007-0004-7|s2cid=122327461 }}</ref> The general principle is that addition of infinite sums is only commutative for absolutely convergent series.


For example, one false proof that 1=0 exploits the failure of associativity for infinite sums.
:<math>\sum_{n=0}^\infty \frac{(-1)^n}{n}</math>


As another example, by [[Mercator series]]
converges to <math>\ln2</math>.
<math display="block">\ln(2) = \sum_{n=1}^\infty \frac{(-1)^{n+1}}{n} = 1 - \frac{1}{2} + \frac{1}{3} - \frac{1}{4} + \cdots.</math>


But, since the series does not converge absolutely, we can rearrange the terms to obtain a series for <math display="inline">\tfrac 1 2 \ln(2)</math>:
A broader test for convergence of an alternating series is the ''Cauchy criterion'': if the sequence <math>a_n</math> is monotone decreasing and tends to zero, then the series
<math display="block">\begin{align}
& {} \quad \left(1-\frac{1}{2}\right)-\frac{1}{4} +\left(\frac{1}{3}-\frac{1}{6}\right) -\frac{1}{8}+\left(\frac{1}{5} -\frac{1}{10}\right)-\frac{1}{12}+\cdots \\[8pt]
& = \frac{1}{2}-\frac{1}{4}+\frac{1}{6} -\frac{1}{8}+\frac{1}{10}-\frac{1}{12} +\cdots \\[8pt]
& = \frac{1}{2}\left(1-\frac{1}{2} + \frac{1}{3} -\frac{1}{4}+\frac{1}{5}- \frac{1}{6}+ \cdots\right)= \frac{1}{2} \ln(2).
\end{align}</math>


== Series acceleration ==
:<math>\sum_{n=0}^\infty (-1)^n\,a_n</math>
In practice, the numerical summation of an alternating series may be sped up using any one of a variety of [[series acceleration]] techniques. One of the oldest techniques is that of [[Euler summation]], and there are many modern techniques that can offer even more rapid convergence.


==See also==
converges.
* [[Grandi's series]]
* [[Nörlund–Rice integral]]


==Notes==
A '''conditionally convergent series''' is an infinite series that converges, but does not converge absolutely. The following non-intuitive result is true: if the ''real'' series
{{reflist}}


==References==
:<math>\sum_{n=0}^\infty (-1)^n\,a_n</math>
* [[Earl D. Rainville]] (1967) ''Infinite Series'', pp 73–6, [[Macmillan Publishers]].
*{{MathWorld|title=Alternating Series|urlname=AlternatingSeries}}


{{series (mathematics)}}
converges conditionally, then for every real number <math>\beta</math> there is a ''reordering'' <math>\sigma</math> of the series such that


{{DEFAULTSORT:Alternating Series}}
:<math>\sum_{n=0}^\infty (-1)^n\,a_{\sigma(n)}=\beta.</math>
[[Category:Mathematical series]]
[[Category:Real analysis]]

Latest revision as of 20:33, 23 March 2024

In mathematics, an alternating series is an infinite series of the form

or
with an > 0 for all n. The signs of the general terms alternate between positive and negative. Like any series, an alternating series converges if and only if the associated sequence of partial sums converges.

Examples[edit]

The geometric series 1/21/4 + 1/81/16 + ⋯ sums to 1/3.

The alternating harmonic series has a finite sum but the harmonic series does not.

The Mercator series provides an analytic expression of the natural logarithm:

The functions sine and cosine used in trigonometry can be defined as alternating series in calculus even though they are introduced in elementary algebra as the ratio of sides of a right triangle. In fact,

and
When the alternating factor (–1)n is removed from these series one obtains the hyperbolic functions sinh and cosh used in calculus.

For integer or positive index α the Bessel function of the first kind may be defined with the alternating series

where Γ(z) is the gamma function.

If s is a complex number, the Dirichlet eta function is formed as an alternating series

that is used in analytic number theory.

Alternating series test[edit]

The theorem known as "Leibniz Test" or the alternating series test tells us that an alternating series will converge if the terms an converge to 0 monotonically.

Proof: Suppose the sequence converges to zero and is monotone decreasing. If is odd and , we obtain the estimate via the following calculation:

Since is monotonically decreasing, the terms are negative. Thus, we have the final inequality: . Similarly, it can be shown that . Since converges to , our partial sums form a Cauchy sequence (i.e., the series satisfies the Cauchy criterion) and therefore converge. The argument for even is similar.

Approximating sums[edit]

The estimate above does not depend on . So, if is approaching 0 monotonically, the estimate provides an error bound for approximating infinite sums by partial sums:

That does not mean that this estimate always finds the very first element after which error is less than the modulus of the next term in the series. Indeed if you take and try to find the term after which error is at most 0.00005, the inequality above shows that the partial sum up through is enough, but in fact this is twice as many terms as needed. Indeed, the error after summing first 9999 elements is 0.0000500025, and so taking the partial sum up through is sufficient. This series happens to have the property that constructing a new series with also gives an alternating series where the Leibniz test applies and thus makes this simple error bound not optimal. This was improved by the Calabrese bound,[1] discovered in 1962, that says that this property allows for a result 2 times less than with the Leibniz error bound. In fact this is also not optimal for series where this property applies 2 or more times, which is described by Johnsonbaugh error bound.[2] If one can apply the property an infinite number of times, Euler's transform applies.[3]

Absolute convergence[edit]

A series converges absolutely if the series converges.

Theorem: Absolutely convergent series are convergent.

Proof: Suppose is absolutely convergent. Then, is convergent and it follows that converges as well. Since , the series converges by the comparison test. Therefore, the series converges as the difference of two convergent series .

Conditional convergence[edit]

A series is conditionally convergent if it converges but does not converge absolutely.

For example, the harmonic series

diverges, while the alternating version
converges by the alternating series test.

Rearrangements[edit]

For any series, we can create a new series by rearranging the order of summation. A series is unconditionally convergent if any rearrangement creates a series with the same convergence as the original series. Absolutely convergent series are unconditionally convergent. But the Riemann series theorem states that conditionally convergent series can be rearranged to create arbitrary convergence.[4] The general principle is that addition of infinite sums is only commutative for absolutely convergent series.

For example, one false proof that 1=0 exploits the failure of associativity for infinite sums.

As another example, by Mercator series

But, since the series does not converge absolutely, we can rearrange the terms to obtain a series for :

Series acceleration[edit]

In practice, the numerical summation of an alternating series may be sped up using any one of a variety of series acceleration techniques. One of the oldest techniques is that of Euler summation, and there are many modern techniques that can offer even more rapid convergence.

See also[edit]

Notes[edit]

  1. ^ Calabrese, Philip (March 1962). "A Note on Alternating Series". The American Mathematical Monthly. 69 (3): 215–217. doi:10.2307/2311056. JSTOR 2311056.
  2. ^ Johnsonbaugh, Richard (October 1979). "Summing an Alternating Series". The American Mathematical Monthly. 86 (8): 637–648. doi:10.2307/2321292. JSTOR 2321292.
  3. ^ Villarino, Mark B. (2015-11-27). "The error in an alternating series". arXiv:1511.08568 [math.CA].
  4. ^ Mallik, AK (2007). "Curious Consequences of Simple Sequences". Resonance. 12 (1): 23–37. doi:10.1007/s12045-007-0004-7. S2CID 122327461.

References[edit]