Peel Off
  • Blog
  • About
  • Contact

the extended real line from a topological pov (introduction)

7/11/2014

37 Comments

 
The extended real line, which we will soon define, is a useful setting in the study of measure and integration, for example. But, in that context, it is seen primarily as a tool, and the topological properties that one can acquire from it remains relatively shaded, since the focus is on its useful algebraic properties, and simple limit properties. We shall see that one can arrive in interesting results (that hold even when you consider problems on $\mathbb{R}$ itself) quite easily with it.

IDEA: We want to attach two points to the real line and call them $\infty$ and $-\infty$, and we want them to behave as we expect from something we would call $\infty$ and $-\infty$. For that, if we want to talk about the topology of this resulting space, we essentially have to say what are the neighbourhoods of this topology. We still want $(a-\delta,a+\delta)$ to be a neighbourhood of a real number $a$, for example. But we also want neighbourhoods of $\infty$ now. It seems a reasonable attempt to define a neighbourhood of $\infty$ as $(M,\infty]$ for example (note the closed bracket, indicating that $\infty$ is in the set).

Before proceding, I introduce the concept of a basis of a topology. Essentially, the basis of a topology is a smaller, "controlled" set that generates the topology - it says who the open sets are.

Definition: If $X$ is a set, a basis of a topology in $X$ is a collection $\mathcal{B}$ of sets that satisfy the following properties:

(i) For all $x \in X$ there is a set $B$ in $\mathcal{B}$ that contains $x$
(ii) If $x$ is in the intersection of two basis elements $B_1$ and $B_2$, then there is a basis element $B_3$ that contains $x$ and that is contained in the intersection of $B_1$ and $B_2$. 

We define the topology $\tau$ generated by $\mathcal{B}$ to be the collection of sets $A$ that satisfy the property that for all $x \in A$, there is a basis element $B$ containing $x$ and contained in $A$.

For example, the balls of a metric space are a basis for its topology (draw it in order to understand!)
Another example of a basis (which is, in fact, a corollary of the balls of metric spaces) are the open intervals in the real line.

Of course, there are technical issues (minor ones, easily solved) that I'll overpass. We have to prove that the topology generated by $\mathcal{B}$ is in fact a topology, as defined in a previous post. If you are interested, you can do it as an exercise.

Now, let's jump into what we wanted!

Definition: Take two points that are not in $\mathbb{R}$ and call them $\infty$ and $-\infty$. Now, define
$$\displaystyle \overline{\mathbb{R}}:=\mathbb{R} \cup \{\infty,-\infty\}$$
Furthermore, define the following basis on $\displaystyle \overline{\mathbb{R}}$:
The basis $\mathcal{B}$ will consist of the open intervals and of the sets $(b, \infty]$ and $[-\infty, a)$ for all $b$ and $a$ real numbers.

That this is in fact a basis (which means that this satisfies the properties listed before) is easy to verify.

Now, in order not to introduce a lot of notations and definitions, I'll not define the subspace topology. It is not a difficult definition, but may be abstract and not enlightening at first. Hence, I'll just assume an intuition in it, in order to justify the following: it seems clear that, if you have 
$\displaystyle \overline{\mathbb{R}}$ and pass from it to $\mathbb{R}$, the topology you inherit is exactly the standard topology of $\mathbb{R}$. We will use this fact.

We arrive now at a change of point of view:

In analysis, one often learns the following definition:

We say a sequence $x_n$ converges if there is a real number $L$ such that $\forall \epsilon >0$ there exists a $N \in \mathbb{N}$ such that $n > N \implies |x_n - L| < \epsilon$. In this case, we call $L= \lim x_n$. Otherwise, we say the sequence diverges.

But we also have the following definition: 

($1$) Given a sequence $x_n$, we say $\lim x_n= \infty$ if $\forall A \in \mathbb{R}$ there is a $N \in \mathbb{N}$ such that $n > N \implies x_n> A$. 



Note that this is a slight abuse of notation. The sequence $x_n$ above, BY DEFINITION, does not converge. But we say  $\lim x_n= \infty$, because it makes sense. To be completely honest, we should write something different, like $L ~x_n= \infty$

But note that, according to our topology, we have that the definition of $L ~x_n= \infty$ is in fact the definition of $\lim x_n= \infty$. In fact, ($1$) is precisely telling: For all neighbourhoods $V$ of infinity, there exists an $N$ such that $n > N \implies x_n \in V$. So, $x_n$ CONVERGES, and REALLY CONVERGES to $\infty$.

We come to our first proposition:



Proposition: $\displaystyle \overline{\mathbb{R}}$ is compact.


Proof: Take an open cover $V_i$ of $\displaystyle \overline{\mathbb{R}}$. Choose a $V_{i_1}$ such that it contains $+\infty$, and a $V_{i_2}$ such that it contains $-\infty$. They contain a set of the form $(b,\infty]$ and $[-\infty, a)$ respectively, so the rest of the $V_i$ should cover $[a,b]$, which is contained in the complement of those sets. But, by the Heine-Borel Theorem, $[a,b]$ is compact. Hence, there is a finite subcover of $V_i$ that covers $[a,b]$. So, this finite subcover, together with  $V_{i_1}$ and  $V_{i_2}$ covers $\displaystyle \overline{\mathbb{R}}$. So, we arrived at a finite subcover for $\displaystyle \overline{\mathbb{R}}$. $\blacksquare$.

Corollary: Every sequence in $\displaystyle \overline{\mathbb{R}}$ has a convergent subsequence.


Note the analogy between bolzano-weierstrass theorem and the corollary above. Bolzano-weierstrass says every bounded sequence has a convergent subsequence. 


We arrive now at a result that does not involve $\displaystyle \overline{\mathbb{R}}$ at first sight:

Proposition: Let $f:[0,\infty) \rightarrow \mathbb{R}$ be a continuous function such that $\displaystyle \lim _{x \rightarrow \infty}f(x) =L$, and $L<f(0)$. Then, $f$ has a maximum.

Proof: Define $\overline{f} :[0,\infty] \rightarrow \mathbb{R}$ as $f(x)$ if $x \in [0,\infty ) $ and $L$ if $x=\infty$. Since $\displaystyle \lim _{x \rightarrow \infty} f(x)=L$, $\overline{f}$ is continuous in $[0,\infty]$. Since $[0,\infty]$ is closed (not proved, but easily seen to be true) and $\displaystyle \overline{\mathbb{R}}$ is compact, $[0,\infty]$ is compact. Hence, $\overline{f}$ reaches a maximum on $[0,\infty]$. This maximum cannot be in $\infty$, since $f(0)>f(\infty)$. Hence, this maximum must be achieved in $[0,\infty)$. $\blacksquare$


Note somethings in the previous demonstration:
First, $0$ has nothing special. If there was any other place where $f$ was greater than $L$, it would be enough.
Secondly, this requirement (that $f(0)>L$) is just to guarantee that the maximum is in $[0,\infty)$ and not in $[0,\infty]$. In fact, there is always a maximum in $[0,\infty]$. The problem is, sometimes the maximum can be achieved at infinity. Draw an example of this (any monotonic increasing bounded function will do!).

We conclude by sketching the proof of the following theorem:


Theorem: A continuous bijective function $f$ on an interval has a continuous inverse.

Sketch of Proof: If the interval is of the form $[a,b]$, it is compact, and we are done.
If the interval is of the form $[a,b)$, since $f$ is continuous and bijective, it is monotonic (by the intermediate value theorem), so $\displaystyle \lim_{x \rightarrow b}f(x)$ exists (it can be $\infty$, no problem!). Pass to the extension $\overline{f}$ of $f$ on $[a,b]$. It is continuous. Hence, since $[a,b]$ is compact, the inverse is continuous. Restrict the inverse by taking away $\overline{f}(b)$. This is precisely the inverse of $f$.  The rest is analogous. $\blacksquare$.
37 Comments

Sin and cos

1/11/2014

0 Comments

 
We present a way to define $\sin$ and $\cos$ which is quite traditional, but show a non-canonical way to "prove" that these definitions are equivalent to the geometrical ones.

First, let's define the derivative of a function $f:\mathbb{R} \rightarrow \mathbb{C}$:

Definition: Given a function $f:\mathbb{R} \rightarrow \mathbb{C}$ given by: $f(x)=\Re(f(x))+i \Im(f(x))$, define:

$f'(x)=\Re'(f(x))+i \Im'(f(x))$


OBS: Note that theorems like "derivative of sum is sum of derivatives" still hold, as well as the definition of derivative by the limit.
OBS: Note also that this ISN'T the derivative of a function $f:\mathbb{C} \rightarrow \mathbb{C}$. We are concerned with functions with real domain.


Now, extend the definition of exponentiation (read the first post on this blog) to complex numbers:


Definition: $\displaystyle e^z:=\sum_{n=0}^{\infty}\frac{z^n}{n!}$


The series converges for every complex $z$ by the ratio test, and the formula $e^{(z+w)}=e^ze^w$ still holds by the cauchy product formula. Now, let's calculate the derivative of $e^x$ and $e^{ix}$. Note that $x$ is real.
It's common to do this by theorems of power series. We shall not use them. Instead, we use more elementary methods.
For the derivative of $e^x$:

$\displaystyle \lim_{h\rightarrow 0} \frac{e^{x+h}-e^x}{h}=e^{x}\lim_{h\rightarrow 0} \frac{e^h-1}{h}$

Now, to evaluate the last limit (without using theorems of power series), do the following:

Fix an arbitrary $H >0$.
Now, given an $\epsilon >0$, there exists $n \in \mathbb{N}$ such that:

$$\frac{H^{n}}{(n+1)!}+\frac{H^{n+1}}{(n+2)!}+...  \leq \epsilon$$

since the series $\displaystyle \sum_{k=0}^{\infty}\frac{H^k}{(k+1)!}$ converges by the ratio test. But note that if you multiply $0<h<H$ this implies :

$$\frac{hH^{n}}{(n+1)!}+\frac{hH^{n+1}}{(n+2)!}+... \leq \epsilon.h$$

Since $h<H$:

$$\frac{h^{n+1}}{(n+1)!}+\frac{h^{n+2}}{(n+2)!}+...  \leq \frac{hH^{n}}{(n+1)!}+\frac{hH^{n+1}}{(n+2)!}+... \leq \epsilon.h$$

But then, we have:

$$e^h \leq 1+h+\frac{h^2}{2!}+\frac{h^3}{3!}+...+\frac{h^n}{n!} + \epsilon.h$$

Which gives us:

$$\frac{e^h -1}{h} \leq 1+\frac{h}{2!}+\frac{h^2}{3!}+...+\frac{h^{n-1}}{n!} + \epsilon$$


But $1\leq \frac{e^h -1}{h}$ is obvious from the definition of $e^h$. So, taking limits:


$$1 \leq \displaystyle \lim_{h\rightarrow 0^{+}} \frac{e^h -1}{h} \leq 1+\epsilon$$


But $\epsilon>0$ was arbitrary, which gives:


$$\lim_{h\rightarrow 0^{+}} \frac{e^h -1}{h} =1$$

Now, note that:

 $$\displaystyle \lim_{h\rightarrow 0^{-}} \frac{e^h -1}{h} =
\lim_{h\rightarrow 0^{+}}  \frac{e^{-h}-1}{-h}= \lim_{h\rightarrow 0^{+}} \frac{\frac{1}{e^h}-1}{-h}=
\lim_{h\rightarrow 0^{+}} \frac{e^h-1}{h}.\frac{1}{e^h}=1$$  

Hence, the limit equals $1$, and it is proved that the derivative of $e^x$ is $e^x$. $\blacksquare$

Now, we will calculate the derivative of $e^{ix}$:

$\displaystyle \lim_{h\rightarrow 0} \frac{e^{i(x+h)}-e^{ix}}{h}=e^{ix}\lim_{h\rightarrow 0} \frac{e^{ih}-1}{h}=e^{ix}\lim_{h\rightarrow 0} \frac{e^{ih}-1}{h}$.

But $e^{ih}=1+ih-\frac{h^2}{2!}-i\frac{h^3}{3!}+\frac{h^4}{4!}+...$. Since the series is absolutely convergent, separate the series in two pieces: the part with $i$ and the part without $i$. Similar estimations that were used before now will be able to be used, and will result (since the term in $h$ is $i$):

$$\lim_{h\rightarrow 0} \frac{e^{ih}-1}{h}=i$$

So, the derivative of $e^{ix}$ is $ie^{ix}$.

You may ask at this point: where is $\cos$ and $\sin$?

Definition: 
$\displaystyle \cos(x):=\frac{e^{ix}+e^{-ix}}{2}$
$\displaystyle \sin(x):=\frac{e^{ix}-e^{-ix}}{2i}$

By the definition of $e^z$, $e^{\overline{z}}=\overline{e^z}$. Then, $\cos$ and $\sin$ are real functions. Moreover, it is evident that:

$$e^{ix}=\cos(x)+ i \sin(x)$$

We also have:

$$|e^{ix}|^2=e^{ix}.\overline{e^{ix}}=e^{ix}e^{-ix}=1$$

which implies:

$$|e^{ix}|=1 \Rightarrow \sin^2(x)+\cos^2(x)=1$$

Also, directly from definition:

$$\cos'(x)=-\sin(x), ~~~~~~\sin'(x)=\cos(x)$$

And also directly from definition: $\cos(0)=1$, $\sin(0)=0$

Now, why on earth are those definitions the sine and cosine we know?

We will prove they must be. How?

Proposition: Let $c:\mathbb{R} \rightarrow \mathbb{R}$ and $s: \mathbb{R} \rightarrow \mathbb{R}$ such that:

(1) $c(0)=1$, $s(0)=0$
(2)$c'(x)=-s(x)$, $s'(x)=c(x)$.

So, $s(x)=\sin(x)$ and $c(x)=\cos(x)$.

This way, since the functions sine and cosine we know geometrically satisfy those properties, they must be the $\sin$ and $\cos$ we just defined.

Proof: Suppose we have functions $c, s$ satisfying those properties.
Define the function $f(x):=(\cos(x)-c(x))^2+(\sin(x)-s(x))^2$. We have:

$$f'(x)=2(\cos(x)-c(x))(-\sin(x)+s(x))+2(\sin(x)-s(x))(\cos(x)-c(x))=0$$

Therefore, $f$ is constant.

But $f(0)=(1-1)^2+(0-0)^2=0$. So $f(x)=0$ for all $x \in \mathbb{R}$. 

But this can only be true if $\sin(x)=s(x)$ and $\cos(x)=c(x)$ for all $x \in \mathbb{R}$. $\blacksquare$.









0 Comments

    Author

    My name is Aloizio Macedo, and I am a 21 years old Mathematics student at UFRJ (Universidade Federal do Rio de Janeiro).

    Archives

    October 2015
    August 2015
    November 2014
    September 2014
    August 2014

    Categories

    All

    RSS Feed

Powered by Create your own unique website with customizable templates.