\(\renewcommand {\hspace }{\ifstar \LWRorighspace \LWRorighspace }\) \(\def \l {\unicode {x0142}}\) \(\renewcommand {\ref }{\ifstar \LWRref \LWRref }\) \(\require {textcomp}\) \(\newcommand {\RR }{\mathbb {R}} \) \(\newcommand {\ZZ }{\mathbb Z} \) \(\newcommand {\NN }{\mathbb N} \) \(\renewcommand {\qedhere }{} \) \(\def\alpha{\unicode{x1D6FC}}\) \(\def\beta{\unicode{x1D6FD}}\) \(\def\delta{\unicode{x1D6FF}}\) \(\def\varepsilon{\unicode{x1D700}}\) \(\def\pi{\unicode{x1D70B}}\)

Week 1 Continuity of the inverse function. Infinite series

Version 2025/02/02 Week 1 in PDF All notes in PDF To other weeks

These notes are being developed to reflect the content of the Real Analysis course as taught in the 2024/25 academic year. The first half of the course is lectured by Dr Yuri Bazlov. Questions and comments on these lecture notes should be directed to Yuri.Bazlov@manchester.ac.uk. The second half will be lectured by Dr Mark Coleman.

Pre-requisite: Mathematical Foundations and Analysis

We build upon what was achieved in the Mathematical Foundations and Analysis (MFA) course, taught in Semester 1. Limit of a sequence, continuous function and limit of a function remain key notions in Real Analysis, which develops the “analysis” part of MFA further. Important functions of real variable, used in MFA, will be formally defined, and their properties proved. This includes the power function \(x^\alpha \) with arbitrary real \(\alpha ;\) exponential function \(a^x\) and the logarithm; trigonometric functions.

Hence, the Real Analysis course will constantly require the students to call on their knowledge of definitions and results introduced in MFA.

An informal introduction (not covered in lectures)

The first part of the course will introduce infinite series. Recall Euler’s identity

\(\displaystyle e^{i\pi } + 1 = 0,\)

often brought up as an example of a beautiful mathematical result. But what exactly is there to prove? Let us try to understand this equation and its connection to Real Analysis.

First of all, \(e\) is the irrational number discovered by Jacob Bernoulli in 1683 (and initially written as “\(b\)”) in his study of compound interest. It is defined as

\(\displaystyle e = \lim _{n\to \infty } \Bigl (1+\frac 1n\Bigr )^n,\) equivalently, \(\displaystyle e =\sum _{n=0}^\infty \frac 1{n!}= 1+\frac 11+\frac 1{1\cdot 2}+\frac 1{1\cdot 2\cdot 3}+\dots .\)

Both formulae involve an infinite process, which in practice can only approximate \(e\) to a given precision. Note how the limit of a sequence, defined in MFA, appears next to something new: a “sum of infinitely many numbers”.

What is \(e^{i\pi },\) though? Raising \(e\) to an imaginary power is done via the rule

\(\displaystyle e^z = 1+\frac z{1!}+\frac {z^2}{2!}+\frac {z^3}{3!}+\dots \)

This is where the meaning of Euler’s identity starts to come across. The “infinite sum” for \(e^{i\pi }\) breaks down into the real and imaginary parts, leading to two equations,

\(\displaystyle 1-\frac {\pi ^2}{2!}+\frac {\pi ^4}{4!}-\frac {\pi ^6}{6!}+\dots = -1\) and \(\displaystyle \pi -\frac {\pi ^3}{3!}+\frac {\pi ^5}{5!}-\frac {\pi ^7}{7!}+\dots = 0.\)

Adding up ten terms of each “infinite sum”, we obtain approximately \(-1.000000004\) and \(-5.3\times 10^{-10}\) (check the calculation here!) which suggests that if we keep adding new terms generated by the same rule, in the limit we will indeed get \(-1\) and \(0,\) respectively. This is mysterious: \(\pi \) is transcendental, so no finite sum like this can be a rational number. A proper way to prove that \(\pi \) satisfies these equations is to express \(\cos (x)\) and \(\sin (x),\) functions given by ratios of sides in a right-angled triangle where \(x\) is in radians, as

\(\displaystyle \cos (x) = 1-\frac {x^2}{2!}+\frac {x^4}{4!}-\frac {x^6}{6!}+\dots ,\)   \(\displaystyle \sin (x) = x-\frac {x^3}{3!}+\frac {x^5}{5!}-\frac {x^7}{7!}+\dots \)

These infinite expansions, apparently known already to Madhava (ca. 1400), are enormously important in mathematics and have many compelling applications in the sciences. To prove these formulas is the same as to prove the formula \(e^{ix}=\cos (x) +i\sin (x)\) (used in MFA without proof). Complete and rigorous theory, leading to these expansions, is part of what is covered in the Real Analysis course. In the first part of MATH11112, we will:

  • formally define infinite series and make sense of “sums of infinitely many numbers”;

  • learn about ways to tell whether a given series converges, i.e., has a sum;

  • understand power series which consist of power of \(x\) with some coefficients, and see why they define “smooth” continuous functions of \(x\) if they converge.

Expansion of functions as power series is intimately connected with differentiation, a formal treatment of which begins the second part of the course. Higher derivatives of a function are key to approximating the function by polynomials, called Taylor polynomials. It is this theory that allows us to write a “good” function as sum of a Taylor series.

The course concludes with the third part devoted to rigorous treatment of integration. A key result is the Fundamental Theorem of Calculus, which demonstrates that integration is truly a reverse operation to differentiation.

Further study of series: a power series is a sum of infinitely many functions of the form \(ax^n.\) In 1807, Joseph Fourier publicised a class of scientific problems which require calculating infinite sums of more sophisticated functions, such as \(a \sin (nx).\) The theory of series that we develop in Real Analysis serves as a foundation for the study of Fourier series and other advanced series in mathematics, science and engineering.

End of the informal introduction.

The Inverse Function Theorem

We begin with a result which could have been proved in Mathematical Foundations and Analysis (MFA), given more time. The Inverse Function Theorem will be used to define a function \(x\mapsto x^r\) where \(r\) is rational. First, a definition.

Definition.

Let \(A\subseteq \RR .\) A real-valued function \(f\) on \(A\) is

.
increasing, if \(\forall x,x'\in A,\) \(x<x'\) \(\implies \) \(f(x)\le f(x');\)
strictly increasing, if \(\forall x,x'\in A,\) \(x<x'\) \(\implies \) \(f(x)< f(x');\)
decreasing, if \(\forall x,x'\in A,\) \(x<x'\) \(\implies \) \(f(x)\ge f(x');\)
strictly decreasing, if \(\forall x,x'\in A,\) \(x<x'\) \(\implies \) \(f(x)> f(x').\)

A function satisfying one of the above conditions is called (strictly) monotone.

The above applies to sequences which are functions on \(A=\NN .\)

We can use monotone sequences to calculate limits of functions. The following is an MFA-style result:

Lemma 1.1: limit of \(f\) from the left via strictly increasing sequences.

For a function \(f,\) defined on \((a,b),\) the following are equivalent:

  • (i) \(\lim _{x\to b-} f(x) = \ell .\)

  • (ii) For all strictly increasing sequences \((x_n)\) such that \(x_n\to b\) as \(n\to \infty ,\) one has \(\lim _{n\to \infty } f(x_n) = \ell .\)

Remark: the Lemma can be expressed in words as follows:

The limit of \(f\) at \(b\) from the left is the common limit of all sequences \((f(x_n))_{n\ge 1},\) where a sequence \((x_n)_{n\ge 1}\) is strictly increasing and converges to \(b.\)

  • Proof of the Lemma (not given in class). (i) \(\Rightarrow \) (ii): let \(\varepsilon >0\) be arbitrary. First, we use the definition of \(\lim _{x\to b-} f(x)=\ell \) to generate \(\delta >0\) such that \(|f(x)-\ell |<\varepsilon \) for all \(x\in (b-\delta , b).\)

    Now we let \((x_n)_{n\ge 1}\) be a strictly increasing sequence, and use the above \(\delta >0\) in the definition of “\(x_n\to b\) as \(n\to \infty \)” to generate \(N\in \NN \) such that \(n\ge N\) implies \(|x_n-b|<\delta .\) That is, \(x_N,x_{N+1},\ldots \in (b-\delta , b+\delta ).\) Since the sequence is strictly increasing with limit \(b,\) no term can exceed \(b,\) so in fact \(x_N, x_{N+1},\ldots \in (b-\delta , b).\)

    But then, by the choice of \(\delta ,\) \(|f(x_n) - \ell |<\varepsilon \) for all \(n\ge N.\) We have shown that \(\ell \) satisfies the definition of limit for the sequence \((f(x_n))_{n\ge 1},\) and so (ii) is proved.

    (ii) \(\Rightarrow \) (i): to prove the contrapositive of this implication, we assume that the statement “\(\lim _{x\to b-} f(x) = \ell \)” is false. This means that there exists some \(\varepsilon _0>0\) such that for all \(\delta >0,\) the interval \((b-\delta , b)\) contains a point, say \(x(\delta ),\) with \(|f(x(\delta ))-\ell |\ge \varepsilon _0.\)

    Choose \(\delta _1=1\) and construct \(x(\delta _1)\in (b-\delta _1,b).\) We have \(|f(x(\delta _1))-\ell | \ge \varepsilon _0.\)

    Then, for each \(n\ge 2,\) choose \(\delta _n = \min (\frac 1n, b-x(\delta _{n-1}))\) and construct \(x(\delta _n)\in (b-\delta _n, b).\)

    Since \(b-\frac 1n < x(\delta _n) < b,\) by the Sandwich Rule \(x(\delta _n)\to b\) as \(n\to \infty .\) Also, since \(x(\delta _n) > b-(b-x(\delta _{n-1})) = x(\delta _{n-1}),\) the sequence \(x(\delta _1), x(\delta _2), \dots \) is strictly increasing.

    We still have, by construction, that \(|f(x(\delta _n))-\ell |\ge \varepsilon _0\) for all \(n.\) Therefore, \(\ell \) fails to satisfy the definition of the limit of the sequence \((f(x(\delta _n)))_{n\ge 1},\) i.e. (ii) is false.

(dots on a graph of a function converging to a limit)

Figure 1.1: Lemma 1.1 says that the limit of \(f\) at \(b\) from the left is the same as the common limit of all sequences \(f(x_1),f(x_2),\ldots \) where \(x_1,x_2,\ldots \) strictly increase and converge to \(b\)

The Lemma is illustrated by Figure 1.1. The next result mirrors the Lemma to deal with a limit from the right:

Corollary: limit of \(f\) from the right via strictly decreasing sequences.

For a function \(f,\) defined on \((a,b),\) the following are equivalent:

  • (i) \(\lim _{x\to a+} f(x) = \ell .\)

  • (ii) For all strictly decreasing sequences \((x_n)\) such that \(x_n\to a\) as \(n\to \infty ,\) one has \(\lim _{n\to \infty } f(x_n) = \ell .\)

In other words,

the limit of \(f\) at \(a\) from the right is the common limit of all sequences \((f(x_n))_{n\ge 1},\) where a sequence \((x_n)_{n\ge 1}\) is strictly decreasing and converges to \(a.\)

We are now ready to prove the first theorem of the course.

Theorem 1.2: the Inverse Function Theorem for strictly increasing functions.

A strictly increasing continuous function \(f\colon [a,b] \to [f(a), f(b)]\) has an inverse \(g\colon [f(a), f(b)]\to [a,b]\) which is strictly increasing and continuous.

  • Proof. \(f\colon [a,b]\to [f(a),f(b)]\) is surjective, because for every \(d\in [f(a),f(b)]\) the Intermediate Value Theorem (and its corollary in MFA) gives \(c\) in \([a,b]\) such that \(f(c)=d.\)

    A strictly increasing \(f\) is injective: indeed, if \(x_1\ne x_2,\) then either \(x_1<x_2\) and so \(f(x_1)<f(x_2),\) or \(x_1>x_2\) and so \(f(x_1)>f(x_2).\) In either case \(f(x_1)\ne f(x_2).\)

    We have shown that \(f\) is bijective, hence it has an inverse \(g=f^{-1}\colon [f(a),f(b)] \to [a,b].\)

    We prove that \(g\) is strictly increasing by contradiction. Assume not, then there exist \(y_1,\) \(y_2\) such that \(y_1<y_2\) and \(g(y_1)\ge g(y_2).\) Since \(f\) is increasing, \(f(g(y_1))\ge f(g(y_2)).\) Since \(f=g^{-1},\) this reads \(y_1\ge y_2,\) but at the same time \(y_1<y_2,\) a contradiction.

    We prove that \(g\) is continuous at an arbitrary point \(d\) of its domain by verifying the criterion of continuity, seen in MFA:

    \[ \lim _{y\to d} g(y) = g(d). \]

    By results from MFA, this is equivalent to

    \[ \lim _{y\to d-} g(y) = \lim _{y\to d+} g(y) = g(d). \]

    We first show that \(\lim _{y\to d-} g(y) = g(d).\) We would like to use Lemma 1.1 for this, so we let \((y_n)\) be a strictly increasing sequence in \([f(a), f(b)]\) which converges to \(d.\) Since \(g\) is a strictly increasing function, \((g(y_n))\) is a strictly increasing sequence; it is also bounded (lies in \([a, b]\)), hence by a result from MFA, \((g(y_n))\) has a limit, say \(c,\) in \([a,b].\)

    Then by Lemma 1.1, \(\lim _{n\to \infty } f(g(y_n)) = \lim _{x\to c-} f(x)\) which is \(f(c)\) as \(f\) is continuous. Since \(f=g^{-1},\) this says that \(\lim _{n\to \infty } y_n = f(c).\) Thus, \(d=f(c),\) hence \(g(d)=c.\)

    We have proved that the common limit of all sequences \((g(y_n)),\) where \(y_n\) strictly increases and converges to \(d,\) is \(g(d).\) By Lemma 1.1, this means that \(\lim _{y\to d-} g(y) = g(d).\)

    The proof that \(\lim _{y\to d+} g(y) = g(d)\) is completely similar, based on the Corollary to Lemma 1.1, and we omit it. Continuity of \(g\) at \(d\) is proved.

Example: the \(p\)th root function.

Let \(p\in \NN .\) Show that the \(p\)th power function \([0,+\infty )\to [0,+\infty ),\) \(x\mapsto x^p,\) has a continuous inverse (denoted \(y\mapsto \sqrt [p]y\) and called the \(p\)th root function).

Solution: define \(f\colon [0,+\infty ) \to [0,+\infty )\) by \(f(x)=x^p.\) Then \(f\) is strictly increasing on \([0,+\infty ).\) Apply Inverse Function Theorem 1.2 to the restriction \([0,b]\xrightarrow {f}[0,b^p]\) to get a continuous inverse \(\sqrt [p]{\ }\colon [0,b^p]\to [0,b].\) Since \(b>0\) can be made arbitrarily large, this defines the continuous function \(\sqrt [p]{\ }\) on all of \([0,+\infty ).\)

We compose continuous, strictly increasing functions to define a rational power function:

Example: raising to rational power \(\frac pq\) where \(p\in \ZZ ,\) \(q\in \NN \).

Define \(x^{\frac pq} = \left ( \sqrt [q]x \right )^p.\) This is a continuous function of \(x\) where \(x\in (0,+\infty ).\)

Remark: one can deduce from the definition of a rational power that

\[ x>1, \ r,s\in \mathbb {Q}, \ r<s \quad \implies \quad x^r < x^s. \]

This allows us to formally define arbitrary real powers of \(x:\)

Definition: \(x^\alpha \) where \(x>0\) and \(\alpha \in \RR \).

If \(x\ge 1,\) define \(x^\alpha = \sup \{x^r: r\in \mathbb Q, r\le \alpha \}.\)

If \(0<x<1,\) define \(x^\alpha = (1/x)^{-\alpha }.\)

A disadvantage of this definition is that proving the expected properties of powers such as \(x^\alpha x^\beta = x^{\alpha +\beta }\) requires work. We will soon obtain a more useful expression for powers via the exponential function.

Infinite series: definition

Definition: infinite series, convergent series, sum.

For real numbers \(a_n,\) an infinite series is an expression of the form \(\sum \limits _{n=1}^\infty a_n\) (also written as \(a_1 + a_2 + \dots + a_n + \dots ,\) \(\sum _{n\ge 1}a_n\) or just \(\sum a_n\)).

The \(n\)th partial sum of this series is the finite sum of terms up to and including \(a_n:\) \(s_n=a_1+\dots +a_n=\sum \limits _{i=1}^n a_i.\)

If the sequence of partial sums converges: \(\lim _{n\to \infty }s_n= s,\) we say that the series \(\smash {\sum \limits _{n=1}^\infty a_n}\) is convergent with sum \(s\), and write \(\sum \limits _{n=1}^\infty a_n = s.\)

Remarks on the definition: (i) \(\sum _{n=1}^\infty a_n = s\) is a actually a shorthand which means “the series \(\sum _{n=1}^\infty a_n\) is convergent with sum \(s\)”.

(ii) Any series that is not convergent is said to be a divergent series.

(iii) A series can start from \(n=N\) (any integer) instead of \(n=1:\) \(a_N+a_{N+1}+\dots = \sum \limits _{n=N}^\infty a_n.\) For example, it is common to start from \(n=0.\) The \(n\)th partial sum will still be the sum which ends with \(a_n:\) e.g., for the series \(a_0+a_1+a_2+a_3+\dots ,\)

  • \(a_0\) is the \(0\)th partial sum,

  • \(a_0+a_1\) is the \(1\)st partial sum, \(a_0+a_1+a_2\) is the \(2\)nd partial sum, and so on.

Basic examples of convergent/divergent series are discussed in week 1 supervision classes.

Alert: a strict definition of convergence.

The definition of convergence of the series \(\sum \limits _{n=1}^\infty a_n,\) used in Real Analysis, is very strict: if the number sequence \(a_1,\) \(a_1+a_2,\) \(a_1+a_2+a_3,\dots \) does not have a limit, then the series has no sum.

Weaker definitions can assign a “sum” to some particular types of series which we consider divergent: Cesàro sum, Abel sum etc. They are used in specialist applications which are beyond this course.

The geometric series

The next example is simple yet important: we will see that more complicated series can be studied by comparing them to a geometric series. We revisit a result seen in MFA.

Proposition 1.3: convergence and sum of geometric series.

Let \(a,r\in \RR .\) The geometric series with initial term \(a\) and ratio \(r,\)

\[ a+ar+ar^2+ar^3+\dots = \sum _{n=0}^\infty ar^n, \]

is convergent if \(|r|<1,\) with sum \(\smash [t]{\dfrac a{1-r}}.\)

  • Proof. The \(n\)th partial sum of the series is \(s_n = a(1+r+r^2+\dots +r^n).\) The calculation

    \begin{align*} &(1+r+r^2 + \dots + r^n)(1-r) \\ &= 1-r +r-r^2 + r^2-r^3 + \dots + r^n-r^{n+1} \\ &= 1 - r^{n+1} \end{align*} where the intermediate terms cancel, gives us the formula

    \[ s_n = a\frac {1-r^{n+1}}{1-r}. \]

    If \(|r|<1,\) we recall from MFA that \(r^{n+1}\) tends to \(0\) as \(n\to \infty ,\) so by Algebra of Limits of convergent sequences,

    \[ s = \lim _{n\to \infty } s_n = a\dfrac {1-0}{1-r}=\dfrac {a}{1-r}. \]

    The sum of the series, is, by definition, the limit of partial sums if it exists. Hence the sum of the geometric series is \(a/(1-r)\) as claimed.

Convergence of series with non-negative terms

Unlike the geometric series, usually there is no nice formula for the \(n\)th partial sum \(s_n.\) We still want to decide if a series is convergent, so we prove theorems known as “convergence tests”. Our first few tests work for series where all terms are non-negative.

Theorem 1.4: boundedness test for non-negative series.

Let a series \(a_1+a_2+\dots \) have \(a_n\ge 0\) for all \(n.\) The following are equivalent:

  • (i) the partial sums \(s_1,s_2,\dots \) are bounded above;

  • (ii) the series is convergent.

If (i) and (ii) hold, the sum of the non-negative series is the least upper bound, \(\sup \{s_n:n\ge 1\},\) of its partial sums.

  • Proof. The partial sums of a non-negative series form an increasing sequence, because \(s_{n+1} = s_n + a_{n+1}\ge s_n\) for all \(n.\) We know from MFA that an increasing sequence \((s_n)_{n\ge 1}\) of real numbers has a limit iff it is bounded above, and then the limit is the supremum of the terms of the sequence.

Corollary: only two convergence types for non-negative series.

A non-negative series \(a_1+a_2+\dots \) is either

  • convergent with a non-negative finite sum: \(\sum \limits _{n=1}^\infty a_n = s,\) \(0\le s<+\infty ,\) or

  • divergent if \(\sup \{s_n: n\ge 1\}=+\infty .\)

In the latter case we use the symbolic notation “\(\sum \limits _{n=1}^\infty a_n = +\infty \)” and say that the non-negative series diverges to \(+\infty \).

Alert.

Notation \(\sum \limits _{n=1}^\infty a_n = +\infty \) (to mean that the series is divergent) and \(\sum \limits _{n=1}^\infty a_n < +\infty \) (to mean that the series is convergent) is used only for series with non-negative terms!

The next test is used very often.

Theorem 1.5: the comparison test for non-negative series.

Assume that \(0\le a_n\le b_n\) for all \(n.\) If the series \(b_1+b_2+\dots \) is convergent (with sum \(T\)), then the series \(a_1+a_2+\dots \) is convergent (with sum at most \(T\)).

If \(a_1+a_2+\dots \) is divergent, then \(b_1+b_2+\dots \) is also divergent.

  • Proof. Write \(s_n=a_1+\dots +a_n\) and \(t_n=b_1+\dots +b_n.\) As \(a_1\le b_1,\) \(a_2\le b_2\) etc, we have \(s_n\le t_n,\) where \(t_n\le T\) by Theorem 1.4. Hence \(s_1,s_2,\dots \) have an upper bound \(T,\) so by boundedness test, Theorem 1.4, the series \(a_1+a_2+\dots \) is convergent.

    The sum \(\sum _{n=1}^\infty a_n\) is the least upper bound of \((s_n)_{n\ge 1},\) and \(T\) is an upper bound. Hence \(\sum _{n=1}^\infty a_n \le T.\)

    Now “\(a_1+a_2+\dots \) is divergent \(\Rightarrow \) \(b_1+b_2+\dots \) is divergent” follows by contrapositive.

To use the Comparison Test, we need to compare with some easy series \(\sum b_n,\) yet the Test does not tell us how to find it. Hence we develop further convergence tests.

Theorem 1.6: the Ratio Test for positive series.

For a positive series \(\sum _{n\ge 1}a_n,\) suppose that the limit \(\ell = \lim \limits _{n\to \infty }\frac {a_{n+1}}{a_n} \) exists. Then if \(0\le \ell <1,\) the series is convergent, and if \(\ell >1,\) the series is divergent.

  • Proof. The case \(0\le \ell <1.\) Choose a positive \(\varepsilon \) such that \(\ell +\varepsilon <1.\) For example, \(\varepsilon = (1-\ell )/2\) works.

    Since \(\ell =\lim _{n\to \infty }\frac {a_{n+1}}{a_n},\) there exists \(N\) such that \(\frac {a_{n+1}}{a_n}< \ell +\varepsilon \) for all \(n\ge N.\) Write this as \(a_{n+1}<ra_n,\) where \(r=\ell +\varepsilon .\) Then \(a_{N+2}<r a_{N+1}<r ^2a_N,\) and, repeating this, we obtain \(a_{N+k} < r^k a_N.\) We have

    \begin{align*} a_1+a_2+\dots +a_{N+k} & \le a_1+\dots +a_{N-1} + a_N + ra_N + \dots + r^k a_N \\ & \le (a_1+\dots +a_{N-1}) + \frac {a_N}{1-r}. \end{align*} The upper bound that we have obtained is a finite constant which does not depend on \(k.\) Thus, partial sums of the series \(a_1+a_2+\dots \) are bounded, so by Theorem 1.4, the series is convergent.

    The case \(\ell >1.\) Put \(\varepsilon =\ell - 1.\) There is \(N\) such that \(\ell -\varepsilon <\frac {a_{n+1}}{a_n}\) for \(n\ge N.\) But \(\ell -\varepsilon =1,\) so \(1 < \frac {a_{n+1}}{a_n},\) equivalently \(a_n<a_{n+1},\) for \(n\ge N.\) In particular, all \(a_n\) for \(n>N\) are greater than the positive constant \(a_N.\) Hence \(s_n\ge (n-N)a_N\) which is unbounded.

Alert: the Ratio test may be inconclusive.

If \(\ell =1\) or the limit does not exist, this test does not tell us anything: the harmonic series \(\sum _{n\ge 1}\frac 1n\) is divergent (see the next Chapter), and the series of inverse squares \(\sum _{n\ge 1} \frac 1{n^2}\) is convergent (see the first exercise sheet). Both have \(\ell =1.\)

In the following test (not taught in lectures, not examinable), we use the \(n\)th root function \(\sqrt [n]{\phantom {x^2}},\) defined earlier.

Theorem 1.7: The \(n\)th Root Test for non-negative series.

For a non-negative series \(\sum _{n\ge 1}a_n,\) suppose that the limit \(\ell = \lim _{n\to \infty } \sqrt [n]{a_n} \) exists. Then if \(0\le \ell <1,\) the series is convergent, and if \(\ell >1,\) the series is divergent.

Remark: Again, if \(\ell =1\) or the limit does not exist, this test does not tell us anything.

  • Proof. (not given in class: very similar to the proof of the Ratio Test; not examinable.) The case \(0\le \ell <1.\) Choose a positive \(\varepsilon \) so that \(r = \ell +\varepsilon \) is still less than \(1;\) for example, \(\varepsilon = (1-\ell )/2\) works.

    By definition of limit, there is \(N\) such that \(\sqrt [n]{a_n}<\ell + \varepsilon =r\) for all \(n\ge N.\) Then \(a_n< r^n\) for \(n\ge N,\) so partial sums of the series are bounded by \(a_1+\dots +a_N+\frac 1{1-r},\) implying convergence.

    The case \(\ell >1.\) Put \(\varepsilon =\ell - 1.\) Since \(\ell \) is the limit of \(\sqrt [n]{a_n},\) there is \(N\) such that \(\ell -\varepsilon <\sqrt [n]{a_n}\) for \(n\ge N.\) But \(\ell -\varepsilon =1,\) so \(1 < \sqrt [n]{a_n},\) equivalently \(1<a_n,\) for \(n\ge N.\) We therefore have \(s_n\ge n-N\) which is unbounded.

Version 2025/02/02 Week 1 in PDF All notes in PDF To other weeks