Is integration theory supposed to be this hard?

In summary, there are several different approaches to integration theory, but the most common one is to start by defining integrals of simple functions in a straightforward way and then generalize to more interesting functions. The approach you mention uses the idea of approximating a function by a sequence of simple functions and taking the limit of their integrals. However, this approach requires proving several theorems, such as the Cauchy criterion and the existence of a common limit for different approximating sequences. Other approaches, such as defining the integral as a supremum in the case of positive functions, may be simpler but still require technical proofs. Ultimately, integration theory involves both macro and micro level concepts, with the main idea being to approximate a function with easier integrable
  • #1
Fredrik
Staff Emeritus
Science Advisor
Gold Member
10,877
423
##\newcommand{\dmu}{\ \mathrm{d}\mu}##
I'm using Friedman to try to learn some integration theory. I got the impression that all books start by defining integrals of simple functions in the same way, and then there are several different ways to generalize the definition to more interesting functions. This book's approach is based on the idea that if there's a sequence ##\langle f_n\rangle## of simple functions such that ##f_n\to f## in some sense, then it might make sense to define
$$\int f\dmu=\lim_n\int f_n\dmu.$$ This idea is so simple that I got quite fond of it at first, but things soon get complicated. I'm wondering if other approaches are easier, or if I'm just wrong about how hard this one is.

The idea described above only makes sense if the precise meaning of "##f_n\to f##" is such that the limit on the right always exists, and is independent of the sequence used. The book chooses that meaning by defining f to be integrable if there's a sequence ##\langle f_n\rangle_{n=1}^\infty## that's Cauchy in the mean, and has the property that ##f_n\to f## almost everywhere. It's easy to show that "Cauchy in the mean" implies that the limit exists. The hard part is to prove that it's independent of the sequence. These are some of the statements we'll have to prove to accomplish that, if we follow the book's path: (I'm not going to be very careful with the details in these statements).

  1. If ##\langle f_n\rangle## is Cauchy in the mean, ##\lim_n\int_E f_n\dmu## exists for all E.
  2. The function ##\lambda## defined by ##\lambda(E)=\lim_n\int_E f_n\dmu## for all E is countably additive.
  3. Now suppose that ##\langle f_n\rangle## and ##\langle g_n\rangle## are two Cauchy sequences in the mean that both converge almost everywhere to f. If E is σ-finite, then ##\lim_n\int_E f_n\dmu=\lim_n\int_E g_n\dmu## for all E. (This one uses the result that these sequences also converge to f in measure. I'm putting that on a separate list).
  4. The previous result implies that ##\lim_n\int f_n\dmu=\lim_n\int g_n\dmu##.
This is not too bad, in my opinion. What I find frustrating is that the comment in parentheses spawns a whole new list.
  1. "Cauchy in the mean" implies "Cauchy in measure".
  2. A sequence of measurable functions that's Cauchy in measure has a subsequence that converges almost uniformly.
  3. If ##f_n\to f## almost uniformly, then ##f_n\to f## in measure.
  4. Corollary of the previous two: A sequence of measurable functions that's Cauchy in measure converges in measure. (When we get to this point, we know that if f is integrable and ##\langle f_n\rangle## is a Cauchy sequence in the mean that converges to f almost everywhere, there's a function g such that ##f_n\to g## in measure, but it doesn't seem obvious that f=g a.e. The next theorem should take care of that).
  5. If ##f_n\to f## almost uniformly, then ##f_n\to f## almost everywhere.
So we apparently need lots of theorems just to make sense of the definition of the integral. The above is far from all of it. We also need to show that integrals of simple functions satisfy the basic stuff we expect them to, such as the triangle inequality for integrals and the linearity of the integration operation.

So is this as easy as it gets, or can the theory be made simpler?
 
Physics news on Phys.org
  • #2
There are many different approaches to integration. The approach you mention can certainly be made simpler. That is: we can easily find a very short definition of the integral, without extra work.

The trick is to define the integral first in the case for positive functions. In that case, you can define the integral simply as a supremum, and not as a limit.

Of course, this does not really absolve us of technical difficulties. We still need to prove theorems like the monotone convergence theorem. That theorem is immediate with the definition in the OP.

I find the book "Real Analysis" by Folland to have a lucid and easy exposition of the integral. I learned the theory myself from "Probability and measure" by Billingsley. But it might not suit your needs.

Of course, if you're really lazy, you can just work with the Daniell integral...
 
Last edited:
  • #3
On the macro scale integration theory is straightforward. You've basically captured the main idea: approximate your function by stuff that's easy to integrate, and then the integral of your function should be approximated by the integrals of the approximating functions. This idea is even present in Riemann integration (Riemann sums, step functions, etc.)

On the micro level, however, the theory is mired with technicality (just like any other mathematical theory, really). That said, the approach you've outlined seems overly complicated. Instead of worrying about stuff like "Cauchy in the mean" and whatnot, you could start your theory by explaining what it means for a simple function (=finite linear combination of characteristic functions) to be integrable. This is very easy, and all the theorems have trivial proofs. Then you prove the helpful result that every nonnegative measurable function is the pointwise limit of a monotone sequence of nonnegative simple functions. This result is fundamental and is also very easy to prove. Once in hand, you can then essentially define the integral of a nonnegative measurable function f to be be the limit of the integrals of such an approximating sequence of simple functions. Essentially. To avoid problems with having to explain why this is well-defined, etc., you can instead define the integral of f to be the sup of the integrals of g, where g runs over the set of nonnegative simple functions that are <= f pointwise. This will make a lot of the basic theorems easy to prove, because they would follow from their simple-function analogues.

This is a really standard approach to the theory. E.g., it's the approach in big Rudin, Royden, and in Bartle's integration theory book. The latter is fairly slim and very easy to read.
 
  • #4
@micromass:

OK, thanks. I think the other integration book I own (Capinski & Kopp) does the positive functions first approach. I think I'll have to study the integration chapter in that book to see if I like that approach better.

I'm interested in the things you say can be made easier in the approach I described. In particular, I'm wondering if there's a short proof for the stuff I mentioned in parentheses in item 3 in the first list, a proof that doesn't require me to go through the entire second list. I don't mind using the concepts of "in measure", "in the mean" and so on, but I don't want to prove a whole bunch of theorems about them if it can be avoided.

Daniell integral...never heard of it. I see there's a Wikipedia article with that title. I will check it out when I've finished the pizza I just bought.

@morphism: I haven't read your post yet. I will after the pizza.
 
  • #5
micromass said:
The trick is to define the integral first in the case for positive functions. In that case, you can define the integral simply as a supremum, and not as a limit.

Of course, this does not really absolve us of technical difficulties. We still need to prove theorems like the monotone convergence theorem. That theorem is immediate with the definition in the OP.

+1

Hi. I was going to answer exactly this (swede? Lunds Universitet? very good experience at erasmus there with Alexandru Aleman, who used this book). You are using Rudin's basic Real-analysis book, but there is another very good book, which I suppose you know, Real and Complex Analysis (there is a third: Functional Analysis, but we are not interested at this moment). Rudin uses this approach. The same as other good books like Capinsky-Kopp's

http://cache0.bookdepository.co.uk/assets/images/book/medium/9781/8523/9781852337810.jpg

I do not know both approaches depthly, but, as micromass says: maybe you avoid some difficulties later with the key Monotone and Dominated convergence theorems (and Fatou's lemma). Maybe. I should know both methods better to know which gets longer.
 
  • #6
Fredrik said:
@micromass:

OK, thanks. I think the other integration book I own (Capinski & Kopp) does the positive functions first approach. I think I'll have to study the integration chapter in that book to see if I like that approach better.

I'm interested in the things you say can be made easier in the approach I described. In particular, I'm wondering if there's a short proof for the stuff I mentioned in parentheses in item 3 in the first list, a proof that doesn't require me to go through the entire second list. I don't mind using the concepts of "in measure", "in the mean" and so on, but I don't want to prove a whole bunch of theorems about them if it can be avoided.

Daniell integral...never heard of it. I see there's a Wikipedia article with that title. I will check it out when I've finished the pizza I just bought.

@morphism: I haven't read your post yet. I will after the pizza.

I've edited my post above to say that I like "Real Analysis" by Folland. I feel you would like this approach. I don't know Capinski & Kopp, I should check that out.

Convergence in measure is really handy though, so nothing is wrong with knowing about that. Cauchy in the mean is something I personally never heard about...

Enjoy your pizza!
 
  • #7
Fredrik said:
Daniell integral...never heard of it. I see there's a Wikipedia article with that title. I will check it out when I've finished the pizza I just bought.
.

Hadn't read that (interesting that you also have Cap-Kopp).

But if you have begun recently with integration theory, I guess it is better to leave that for the moment when you want to integrate in general topological vector spaces
 
  • #8
Hmmm, Capinski-Kopp seems like a nice book. A bit sad that they only develop integration theory on [itex]\mathbb{R}[/itex] and not on arbitrary measure spaces...
 
  • #9
morphism said:
That said, the approach you've outlined seems overly complicated. Instead of worrying about stuff like "Cauchy in the mean" and whatnot, you could start your theory by explaining what it means for a simple function (=finite linear combination of characteristic functions) to be integrable. This is very easy, and all the theorems have trivial proofs. Then you prove the helpful result that every nonnegative measurable function is the pointwise limit of a monotone sequence of nonnegative simple functions. This result is fundamental and is also very easy to prove.
Those things are part of Friedman's approach too. I didn't mention what it means for a simple function to be integrable because it seems all books handle simple functions the same way, and differ only in how they generalize integration to measurable functions. The other theorem you mentioned (2.2.5, p. 34) is indeed a nice theorem with a nice and simple proof.

micromass said:
I've edited my post above to say that I like "Real Analysis" by Folland. I feel you would like this approach.
I've had a quick look at it, and my first impression is that it looks very good. It has a few other sections that I'm interested in as well.

Tarantinism said:
swede? Lunds Universitet?
Yes and no. Stockholm.
 
  • #10
I thought that the integral of a positive measurable function,f, was defined as the supremum of the integrals of simple functions (functions that are a finite linear combination of characterisitc functions of sets of finite measure) that are less than or equal to f. This does not require a notion of limit.
 
  • #11
lavinia said:
I thought that the integral of a positive measurable function,f, was defined as the supremum of the integrals of simple functions (functions that are a finite linear combination of characterisitc functions of sets of finite measure) that are less than or equal to f. This does not require a notion of limit.
##\newcommand{\dmu}{\ \mathrm{d}\mu}##
That seems to be what everyone is saying in the posts above. Based on the responses, I have to conclude that the approach described in #1 is a bit unusual. The idea still has certain appeal, since the basic idea is so simple. We want to define
$$\int f\dmu=\lim_n\int f_n\dmu,$$ where the ##f_n## are integrable simple functions such that ##f_n\to f## a.e., but this only makes sense if the limit exists and is independent of the sequence. The limit on the right is just a limit of a sequence of real numbers, so it exists if and only if the sequence is Cauchy. So we also require that the sequence ##\langle f_n\rangle## has the property that for all ε>0, there's an N such that n,m≥N implies
$$\int|f_n-f_m|\dmu<\varepsilon,$$ because this implies that for all n,m≥N
$$\bigg|\int f_n\dmu-\int f_m\dmu\bigg |\leq\int|f_n-f_m|\dmu<\varepsilon.$$ The only problem is that it's hard to prove that the limit is independent of the sequence used. (See the two lists in post #1).
 
Last edited:
  • #12
##\newcommand{\dmu}{\ \mathrm{d}\mu}##
I think I figured out what's different about Friedman's approach, and why it's harder. The simple approach, as explained in e.g. Folland, defines the integral of a non-negative measurable function f as the supremum of all ##\int g\dmu## where g is a non-negative integrable simple function such that g≤f. This implies that there's an increasing sequence ##\langle f_n\rangle## of non-negative integrable simple functions such that
$$\int f\dmu=\lim_n\int f_n\dmu.$$ Here we can see the difference between the simple approach and Friedman's. It's that the simple approach removes the question of whether the limit depends on the sequence, by only considering sequences <fn> that obviously won't give us different values of ##\lim_n\int f_n\dmu##. Friedman's approach is essentially the same as the simple approach. It just does one more thing in addition to the simple stuff: It proves that if we had used some other sequence (e.g. a sequence of simple functions that are ≥f instead of ≤f), the result is the same.

I think it makes sense to do this, because if we don't, it makes perfect sense to ask: Would the result have been different if we had chosen some other sequence, and if yes, why didn't we?
 
Last edited:
  • #13
Fredrik said:
##\newcommand{\dmu}{\ \mathrm{d}\mu}##
I think I figured out what's different about Friedman's approach, and why it's harder. The simple approach, as explained in e.g. Folland, defines the integral of a non-negative measurable function f as the supremum of all ##\int g\dmu## where g is a non-negative integrable simple function such that g≤f. This implies that there's an increasing sequence ##\langle f_n\rangle## of non-negative integrable simple functions such that
$$\int f\dmu=\lim_n\int f_n\dmu.$$ Here we can see the difference between the simple approach and Friedman's. It's that the simple approach removes the question of whether the limit depends on the sequence, by only considering sequences <fn> that obviously won't give us different values of ##\lim_n\int f_n\dmu##. Friedman's approach is essentially the same as the simple approach. It just does one more thing in addition to the simple stuff: It proves that if we had used some other sequence (e.g. a sequence of simple functions that are ≥f instead of ≤f), the result is the same.

I think it makes sense to do this, because if we don't, it makes perfect sense to ask: Would the result have been different if we had chosen some other sequence, and if yes, why didn't we?

Yes. However, note that the goal is eventually to prove the monotone convergence theorem. Once this is proven, we see that Friedman's definition is essentially the same as the simple approach.
 
  • #14
Fredrik said:
##\newcommand{\dmu}{\ \mathrm{d}\mu}##
That seems to be what everyone is saying in the posts above. Based on the responses, I have to conclude that the approach described in #1 is a bit unusual. The idea still has certain appeal, since the basic idea is so simple. We want to define
$$\int f\dmu=\lim_n\int f_n\dmu,$$ where the ##f_n## are integrable simple functions such that ##f_n\to f## a.e., but this only makes sense if the limit exists and is independent of the sequence.

Lebesque Dominated Convergence seems to give you your theory.
 
  • #16
I have never understood integration theory. I will say some ignorant but well meaning things here in case having others correct them sheds light on the topic.

Thinking about it now in reference to your question, I believe one reason is that the objects being integrated, the "interesting" ones, are not functions at all, they are more like probability distributions.

I.e. they have no fixed values at individual points, there is only the probability their value is a certain thing. E.g. one equates two functions in integration theory if they differ on a set of measure zero. That means you can change any value at all and not change the "function".

For instance the function which is identically zero is the same as the function which is zero on the irrationals and equals p on the reduced fraction p/q. This means there is only the probability one that the function has value zero at any given point, but this is not certainty. This is confusing to me.

Of course in this case there is one continuous function in the equivalence class and that is the most natural representative, but the interesting functions are those that are not equivalent to any continuous function, and then there seems to be no natural representative to choose.The most intuitive way of understanding the integral to me is analogous to Riemann integration, but with axes reversed. I.e. take a positive function on the interval [a,b], and subdivide the y-axis into integer intervals. Then define the lower sum to be equal to the sum over all n, of n times the measure of the set where f has values in [n,n+1].

The upper sum is the sum over all n of n times the measure of the set where f has value in the interval [n-1,n].

If these are both finite, then refine the subdivision of the y-axis into tenths and continue.

As suggested above, then one can define the integral as the LUB of the lower sums, provided it equals the GLB of the upper sums.This puts the heat of course on how to define the measure of a set, but it does define the integral of an actual function.

So in some sense, in integration theory, one does define the integral of an actual function.

But when you try to prove completeness results you still have to go back to equivalence classes of functions. I.e. if you want to define a function as a limit of other functions you have a problem defining the values of the limit function, or else you have to give up uniqueness of limits.I.e. if you have a function f and a sequence fn, then fn-->f means the integrals of f-fn go to zero. But if true, then this is still true for any function equivalent to f.

And if you only have the sequence fn and want to define the limit function you cannot do so uniquely.

I.e. the natural way to define Cauchy - ness of a sequence of functions fn, is to say the integrals of the differences fn-fm -->0, but this does not imply the values of the functions converge everywhere. The hard part referred to above is that they do at least converge pointwise a.e. So a Cauchy sequence does define an equivalence class of functions for a limit.Actually this let's you break the problem into two parts:
1) define the integral norm on the space of all continuous functions, making it a metric space, and define the completion of this metric space as the set of all equivalence classes of Cauchy sequences.
The integral of a class of such Cauchy sequences is well defined, as the limit of the integrals.
This is the easy part, i.e. the formal part.

2) Try to find a function that represents each class of Cauchy sequences. This is the hard part. You have to prove that a Cauchy sequence of functions does converge pointwise a.e.

So you get a limit function which is only defined on the complement of a set of measure zero. of course I guess you could say the value is zero on the set of measure zero where the functions do not converge pointwise.Well to me it is a confusing subject, but several people here seem to have studied it. A famous analyst once told me however that the main theorems in the subject are Fubini and dominated convergence, from the point of view of using the theory that is.

In my own experience, I don't know about dominated convergence but i can certainly vouch for Fubini being useful. Lang's Analysis II (maybe now Real analysis), has a good strong statement of Fubini. (And the functions in Lang have values in any Banach space.)
 
Last edited:
  • #17
mathwonk said:
The most intuitive way of understanding the integral to me is analogous to Riemann integration, but with axes reversed. I.e. take a positive function on the interval [a,b], and subdivide the y-axis into integer intervals. Then define the lower sum to be equal to the sum over all n, of n times the measure of the set where f has values in [n,n+1].

The upper sum is the sum over all n of n times the measure of the set where f has value in the interval [n-1,n].

If these are both finite, then refine the subdivision of the y-axis into tenths and continue.

As suggested above, then one can define the integral as the LUB of the lower sums, provided it equals the GLB of the upper sums.
One of the books mentioned above, Capinski & Kopp, starts out saying that this is how Lebesgue did it in 1902, and then they say:
A century of experience with the Lebesgue integral has led to many equivalent definitions, some of them technically (if not always conceptually) simpler. We shall follow a version which, while very similar to Lebesgue's original construction, allows us to make full use of the measure theory developed already.​
Then they state the supremum definition, that everyone in this thread seems to be familiar with. So they're suggesting that the upper/lower sum thing is kind of hard.

The original approach does however have the advantage that it provides motivation for the definition of the term "measurable function": If you chop up the y-axis into intervals ##\{I_k\}##, and want to assign a value that's approximately equal to
$$\sum_k y_k\,\mu(f^{-1}(I_k)),$$ where ##y_k\in I_k##, it's natural (and maybe unavoidable) to require that all the ##f^{-1}(I_k)## are measurable sets.
 
  • #18
Perhaps someone can answer a very simple question: Are all measurable functions integrable? (I'm sure I'll get to the answer soon enough anyway, but if it only takes a minute to reply...)
 
  • #19
Fredrik said:
Perhaps someone can answer a very simple question: Are all measurable functions integrable? (I'm sure I'll get to the answer soon enough anyway, but if it only takes a minute to reply...)

A measurable function may have an infinite integral. In this case it is not called integrable.
 
  • #20
Fredrik said:
One of the books mentioned above, Capinski & Kopp, starts out saying that this is how Lebesgue did it in 1902, and then they say:
A century of experience with the Lebesgue integral has led to many equivalent definitions, some of them technically (if not always conceptually) simpler. We shall follow a version which, while very similar to Lebesgue's original construction, allows us to make full use of the measure theory developed already.​
Then they state the supremum definition, that everyone in this thread seems to be familiar with. So they're suggesting that the upper/lower sum thing is kind of hard.

The original approach does however have the advantage that it provides motivation for the definition of the term "measurable function": If you chop up the y-axis into intervals ##\{I_k\}##, and want to assign a value that's approximately equal to
$$\sum_k y_k\,\mu(f^{-1}(I_k)),$$ where ##y_k\in I_k##, it's natural (and maybe unavoidable) to require that all the ##f^{-1}(I_k)## are measurable sets.

- In classical physics one wants to measure physical quantities in regions of space. The approach is to divide the region into small pieces -- e.g. rectangles - and multiply the volume of the region by the density of the physical quantity in that region. One then adds these up over lots of small regions to estimate the total.

For instance, the flux of a field across a surface is measured as the sum of the normal component of the field density times the areas of small rectangular pieces of the surface.

This procedure is Riemann integration. To me it naturally arises from classical physics. It is a mathematical model of empirical estimation of physical quantities.On the other hand, if one wants to measure the probability of something happening, one starts with the outcome - the thing whose probability one wants to measure - then asks what is the measure of the set of situations where this outcome will occur. one does not divide space up into equal regions. rather, one finds those regions of space where the outcome will occur. This is Lebesque integration.

- Generally, classical physical quantities are assumed to be continuous, in fact differentiable, away from isolated singularities. In fact, they are often harmonic. For these types of mathematical functions the requirement that upper and lower sums converge to the same number is not a restriction.

In probability theory though, continuity - no less differentiability - is of no interest. the only thing that matters is measuring the volume of a region where an outcome will occur.

- I find it interesting that when lower sums and upper sums converge to the same number that the physical quantity must be nearly continuous. For certainly, the estimation of something like flux across a surface should not depend upon how one approximates. Upper sums, lower sums, in between sums should all give good estimates if the region is small enough. If this were not true, it is hard to imagine doing classical physics.
 
Last edited:
  • #21
lavinia said:
A measurable function may have an infinite integral. In this case it is not called integrable.
Ah, of course. Thanks for clearing that up.

I liked your explanation in the post after the one I just quoted. It's always nice to have a good way to explain what the subject is about to people who don't know.

One more question (for anyone):

##\newcommand{\dmu}{\ \mathrm{d}\mu}##
I'm trying to prove that all three definitions mentioned in this thread are equivalent, and I'm struggling a bit with the proof of why the
$$\int f\dmu=\sup\bigg\{\int\phi\dmu\bigg|\,\phi\text{ is simple, and }0\leq\phi\leq f\bigg\}$$ definition implies that there exists a sequence ##\langle f_n\rangle_{n=1}^\infty## that converges to f a.e., and is "Cauchy in the mean" (=Cauchy with respect to the L1 norm). I know that f is measurable, and that this implies that there's an increasing sequence ##\langle f_n\rangle_{n=1}^\infty## of simple functions such that ##f_n\to f## pointwise (and therefore a.e). (See theorem 2.2.5 in Friedman). I'm guessing that it's possible to show that that sequence is Cauchy in the mean, but I don't see how.

Just to make it perfectly clear: what I want to show is that for all ##\varepsilon>0##, there's an ##N\in\mathbb Z^+## such that for all ##n,m\in\mathbb Z^+##,
$$n,m\geq N\ \Rightarrow\ \int|f_n-f_m|\dmu<\varepsilon.$$
 
  • #22
Fredrik said:
Ah, of course. Thanks for clearing that up.

Also when the function is not positive and the measure space has infinite measure you can get things like the sine which doesn't even diverge. It has no limit at all. For instance if you take multiples of pi then the integral oscillates between zero and 2 forever.
 
  • #23
lavinia said:
Also when the function is not positive and the measure space has infinite measure you can get things like the sine which doesn't even diverge. It has no limit at all. For instance if you take multiples of pi then the integral oscillates between zero and 2 forever.

Or take fun things like

[tex]\int_0^{+\infty}{\frac{\sin(x)}{x}dx}[/tex]

This is not Lebesgue integral, but you can integrate it using improper Riemann integration.
 
  • #25
Fredrik said:
Just to make it perfectly clear: what I want to show is that for all ##\varepsilon>0##, there's an ##N\in\mathbb Z^+## such that for all ##n,m\in\mathbb Z^+##,
$$n,m\geq N\ \Rightarrow\ \int|f_n-f_m|\dmu<\varepsilon.$$

Hey Fredrik.

Can you make use of relationship between the norm of the whole expression and the norm inside in the integral?

For example for your expression, consider the L^1 norm of the integral expression against the integral expression itself that involves a 1-norm of fm and fn (sorry I don't know hardly any latex).

I think there is an inequality that is associated with this but I can't remember it off the top of my head. Maybe you will know what inequality I mean: (there are identities that relate norms like the above in analysis).

I might investigate this for myself and get back to you, but this is a very informative thread.
 
  • #26
chiro said:
Can you make use of relationship between the norm of the whole expression and the norm inside in the integral?
The triangle inequality? I can use it for other things (see post #11), but it doesn't seem to help here.

chiro said:
(sorry I don't know hardly any latex).
If you want to, the LaTeX guide for the forum is a good place to start. Link. It hasn't been updated to include simplified tags instead of itex and tex. You can see how they work by quoting one of my posts with LaTeX. I should probably also mention that the code \dmu that I'm using in my posts in this thread isn't a standard code. I've been defining it in my posts with a \newcommand.
 
  • #27
Fredrik said:
The triangle inequality? I can use it for other things (see post #11), but it doesn't seem to help here.

No not the triangle inequality. I'll explain in a non-latexy kind of way (thanks for the latex though!)

Lets say your integral expression is Integral(|fm-fn|dx). What I'm proposing is to use an inequality that relates |Integral([fm-fn]dx)| with Integral(|fm-fn|dx). I remember seeing something like this before in analysis but I can't for the life of me remember the exact identity or appropriate theorem.

If you want to, the LaTeX guide for the forum is a good place to start. Link. It hasn't been updated to include simplified tags instead of itex and tex. You can see how they work by quoting one of my posts with LaTeX. I should probably also mention that the code \dmu that I'm using in my posts in this thread isn't a standard code. I've been defining it in my posts with a \newcommand.

I've been pretty lazy with LaTex so I'll check that out. Thanks!
 
  • #28
chiro said:
No not the triangle inequality. I'll explain in a non-latexy kind of way (thanks for the latex though!)

Lets say your integral expression is Integral(|fm-fn|dx). What I'm proposing is to use an inequality that relates |Integral([fm-fn]dx)| with Integral(|fm-fn|dx). I remember seeing something like this before in analysis but I can't for the life of me remember the exact identity or appropriate theorem.
That's the triangle inequality. :smile: When ##f=\sum_{k=1}^n a_k\chi_{E_k}##, we have ##|f|=\sum_{k=1}^n |a_k|\chi_{E_k}##, and therefore
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
\bigg|\int f\dmu\bigg| =\bigg|\sum_{k=1}^n a_k\mu(E_k)\bigg| \leq\sum_{k=1}^n |a_k|\mu(E_k) =\int |f|\dmu.$$
 
  • #29
Fredrik said:
That's the triangle inequality. :smile: When ##f=\sum_{k=1}^n a_k\chi_{E_k}##, we have ##|f|=\sum_{k=1}^n |a_k|\chi_{E_k}##, and therefore
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
\bigg|\int f\dmu\bigg| =\bigg|\sum_{k=1}^n a_k\mu(E_k)\bigg| \leq\sum_{k=1}^n |a_k|\mu(E_k) =\int |f|\dmu.$$

Haha yeah it is too! XD, Sorry!

Theres no way to prove that you get a bounded result which would satisfy your continuity condition? What about if you considered supremums in the expression? Are there identities that give inequalities for bounds when we use supremums so that you get RHS <= something?

(I'm just throwing out ideas not knowing if they carry any value or not!)

Also the fact that it is Cauchy means that if we have your sequence {f} then by definition |fm - fn| with your conditions will always converge (thats one of the definitive things that the "Cauchy" part of the definition brings to the table).

This is just an idea, but my guess is that you might need to look at the definition of Cauchy-sequence and see if it implies a particular bound for the parameters you have stated.

If it does then you can set your epsilon to that value and complete the proof.
 
  • #30
I think I found the solution. It was much easier than I thought. Since ##\langle f_n\rangle_{n=1}^\infty## is an increasing sequence, we have (if n>m)
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
\int |f_n-f_m|\dmu=\int(f_n-f_m)\dmu=\int f_n\dmu-\int f_m\dmu.$$
Since ##f_n\to f## pointwise, the monotone convergence theorem tells us that
$$\lim\int f_n\dmu=\int f\dmu.$$
So ##\langle\int f_n\rangle## is a convergent sequence of real numbers, and is therefore a Cauchy sequence. So we can choose N such that for all n,m≥N
$$\int f_n\dmu-\int f_m\dmu<\varepsilon.$$
 
  • #31
Fredrik, there's another approach to integration that may interest you, the one that is developed in Wheeden and Zygmund. In this approach you essentially get the definition of the integral for "free", and a lot of the theorems that you have to struggle with in other approaches also come for "free", but then you have to prove some theorems that are pretty trivial in the other approaches, like the dominated convergence theorem and the result that the integral is equal to the limit of the integrals of simple functions. It goes like this: for any nonnegative function [itex]f:ℝ^{n}\rightarrowℝ[/itex] we define the integral of f over a region [itex]E\subsetℝ^{n}[/itex] to be the Lebesgue measure of the region under the graph of the function, considered as a subset of [itex]ℝ^{n+1}[/itex]. That's it! (And of course you have to do the usual definition of the integral for arbitrary functions using positive and negative parts, but that's trivial.) Altogether, I like this geometric approach better, and of course you ultimately get the same theorems and properties of the integral as from the step-function approach.
 
  • #32
Thanks for the tip. That sounds like a good approach if we're only interested in integrals on ##\mathbb R^n##. If we want to work with other measures than the Lebesgue measure, or do integrals on something else, like topological groups, then I think we need one of the approaches discussed above.
 
  • #33
Fredrik said:
Thanks for the tip. That sounds like a good approach if we're only interested in integrals on ##\mathbb R^n##. If we want to work with other measures than the Lebesgue measure, or do integrals on something else, like topological groups, then I think we need one of the approaches discussed above.
Actually, I think this approach can be easily generalized to abstract integration: if [itex](X,\mu)[/itex] is a measure space, just use the product measure on the space [itex]X\timesℝ[/itex]. Now as for other things you might want to do, like integrals with respect to operator-valued measures, that might be a bit more iffy because I don't know whether you can meaningfully form a product measure out of an operator-valued measure.
 
  • #34
Anyone see how to prove that the limit definition "implies" the supremum definition?

Limit definition: An a.e. real-valued measurable function f is said to be integrable if there's a sequence ##\langle f_n\rangle## of integrable simple functions that's Cauchy in the mean and such that ##f_n\to f## a.e. The integral of an integrable function is defined by
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
\int f\dmu=\lim_n\int f_n\dmu,$$ where the ##f_n## are the terms of any sequence of integrable simple functions that's Cauchy in the mean and such that ##f_n\to f## a.e.

Supremum definition: An a.e. real-valued measurable function f is said to be integrable if the sets
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
I_f^\pm=\bigg\{\int\phi\dmu\bigg|\phi\text{ is simple, }0\leq\phi\leq f^\pm\bigg\}$$ are bounded from above. The integral of an integrable function is defined by
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
\int f\dmu=\sup I_f^+-\sup I_f^-.$$
I was able to prove that if f is integrable according to the supremum definition, then it's integrable according to the limit definition, and the two definitions agree about the value of the integral. But I don't see how to prove that if f is integrable according to the limit definition, then it's integrable according to the supremum definition. I have proved that f is integrable according to the limit definition if and only if f+ and f- are integrable according to the limit definition. I'm thinking that I should try to derive a contradiction from the assumption that one of ##I_f^+## and ##I_f^-## is not bounded from above. If ##I_f^+## isn't bounded from above, there's a sequence ##\langle\phi_n\rangle## of integrable simple functions such that for all n, ##0\leq\phi_n\leq f^+## and ##\newcommand{\dmu}{\ \mathrm{d}\mu}\int\phi_n\dmu>\lim_n\int f^+_n\dmu+1##, making it impossible for the sequence ##\newcommand{\dmu}{\ \mathrm{d}\mu}\langle\int\phi_n\dmu\rangle## to have the right limit, but unless I can find such a seqeuence that's Cauchy in the mean, this doesn't prove anything.

Reminder: A sequence ##\langle f_n\rangle## of a.e. real-valued measurable functions is said to be Cauchy in the mean, or to be a Cauchy sequence in the mean, if for all ##\varepsilon>0## there's an ##N\in\mathbb Z^+## such that for all ##n,m\in\mathbb Z^+##,
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
n,m\geq N\ \Rightarrow\ \int|f_n-f_m|\dmu<\varepsilon.$$
 
Last edited:
  • #35
Fredrik said:
Reminder: A sequence ##\langle f_n\rangle## of a.e. real-valued measurable functions is said to be Cauchy in the mean, or to be a Cauchy sequence in the mean, if for all ##\varepsilon>0## there's an ##N\in\mathbb Z^+## such that for all ##n,m\in\mathbb Z^+##,
$$\newcommand{\dmu}{\ \mathrm{d}\mu}
n,m\geq N\ \Rightarrow\ \int|f_n-f_m|\dmu<\varepsilon.$$
Is that the same as Cauchy in the L1 norm?
 
Back
Top