Prove f(x) is zero in range [-1,1]

  • Thread starter jaus tail
  • Start date
  • Tags
    Range Zero
In summary: What could be false? I only thought continuity invites to be integrated to ##F'(x)=f(x)## with ##F(0)=F(-1)=F(1)=0## and ##F=F\circ g## if I didn't make a mistake. These are easier to handle.In summary, the homework statement is saying that a function is continuous in the range [-1, 1] and that it equals 0 everywhere within that range. However, the function is not always zero and it may be helpful to use integration to find the solution to the equation.
  • #36
even for x neighbouring to ##\xi## , ##2^np_n(x)## is totally different than ##2^n(\xi \pm \epsilon)^n##.
 
Last edited:
Physics news on Phys.org
  • #37
Delta² said:
even for x neighbouring to ##\xi## , ##2^np_n(x)## is totally different than ##2^n(\xi+-\epsilon)^n##. I don't think the limit ##\lim_{x->\xi}p_n(x)## for x any but not ##\xi## is equal to ##2^n(\xi+-\epsilon)^n##.
I'm not saying it won't work, only that it requires more care. It's not obvious, at least to me, why it will work. I assume that some topological argument might be needed to make the step from iterations to the continuum. I would prefer a theorem before cases, but that's a personal view.
 
  • #38
Delta² said:
even for x neighbouring to ##\xi## , ##2^np_n(x)## is totally different than ##2^n(\xi \pm \epsilon)^n##.

If you are using that to approximate ##f ## in ##[0,1]## then the limit should be finite, as continuous function in compact is bounded. But then you have a limit of the type ##\infty . \epsilon ## which should be dominated by ##\epsilon ## in order for the limit to be finite.
 
  • #39
Part of the problem is that there are continuous functions on compact intervals with uncountably many zeros and even with the added condition that there are non-zero values between any two zeros, i.e., without f being zero in (a,b) , e.g. the distance function d(x,C) , for C a Cantor set. So it is difficult to discount many choices on these grounds alone. Basically, any closed subspace can be the zero set of a continuous function. This seems to allow a great amount of functions with uncountably-many zeros in a compact subspace; I thought such functions would be rare.
 
  • #40
It seems like a lot more brain power has gone into this than perhaps the question wanted...

Another approach would be to show that for any any ##x \in (-1,0)##, after a large enough finite number of iterations, the resulting x becomes 'trapped' in ##\{(-1, 0.99], [0.01, 0)\}## and with respect to the magnitude of the function the effect of the x scaling over two iterations is a multiple less than one fourth, which overwhelms the scaling of ##2^2 = 4##, and hence the infinite product decays toward zero. This can be shown for 3 iterations at a time as well. It's sufficient to show this for 10 iterations for starting x = 0.5.

With respect to the ##2x## factor. Another way to think about it is how it relates to ##GM \leq AM## where the inequality is strict unless all values are the same. If we wanted to have some fun, look at the magnitude of the scalars ##\big \vert \approx 0\big \vert, \big \vert \approx -1\big \vert##, and see that the arithmetic mean is ##\frac{1}{2}## which is strictly larger than the geometric mean. Hence the effect of scaling by ##x## in our infinite product must overwhelm the scaling by 2. The infinite product decays toward zero.
 
  • #41
We've only been taught of limits and derivatives and integration.
I tried using derivatives.
f(x) = 2x(f(x2-a))
Let f(x^2 - 1) = g(x)
So
f(x) = 2x g(x)
f'(x) = 2xg'(x) + 2g(x)
f'(x) = 2xf'(x)(2x) + 2f(x^2 - 1)
f'(x) (1-4x^2) = 2f(x^2 - 1)
So we get f'(x) = 2f(x^2 - 1)/(1-4x^2)

Now we got that f(-1) = f(1) = 0.
So if f(x) is non zero at any point between 1 and -1, then f'(x) must be positive some place and negative some place.
But above we see that f'(x) is an even function.
So f'(x) is not satisfying condition of positive and negative. It's either positive or negative and if it's one of them then it can have only 1 zero.
And so we can say that f(x) is constant at zero.

Aren't all continuous functions derivative? I'm not sure what's the difference. Sine and consine are continuous and differentiable. So i guess it should be true.
 
  • #42
jaus tail said:
Aren't all continuous functions derivative? I'm not sure what's the difference. Sine and consine are continuous and differentiable. So i guess it should be true.

Not all continuous functions are differentiable. One example is the modulus ##|x|##, which is not differentiable at ##x = 0##.

There is, in fact, the Weierstrass function, which is continuous but nowhere differentiable!

https://en.wikipedia.org/wiki/Weierstrass_function
 
  • #43
So I guess we can take two conditions.
1) f(x) is differentiable then we can go with my solution.
and 2) f(x) is not differentiable. Any idea how to prove in this case?
 
  • #44
jaus tail said:
We've only been taught of limits and derivatives and integration.
I tried using derivatives.
f(x) = 2x(f(x2-a))
Let f(x^2 - 1) = g(x)
So
f(x) = 2x g(x)
f'(x) = 2xg'(x) + 2g(x)
f'(x) = 2xf'(x)(2x) + 2f(x^2 - 1)
f'(x) (1-4x^2) = 2f(x^2 - 1)
So we get f'(x) = 2f(x^2 - 1)/(1-4x^2)

Now we got that f(-1) = f(1) = 0.
So if f(x) is non zero at any point between 1 and -1, then f'(x) must be positive some place and negative some place.
But above we see that f'(x) is an even function.
So f'(x) is not satisfying condition of positive and negative. It's either positive or negative and if it's one of them then it can have only 1 zero.
And so we can say that f(x) is constant at zero.

First:

##g'(x) = 2xf'(x^2-1) \ne 2xf'(x)##

But also, I don't understand how you get that an even function can't take positive and negative values.

I wonder where you got this question, because it seems to be a little beyond your level at the moment.
 
  • Like
Likes jaus tail and Delta2
  • #45
jaus tail said:
So I guess we can take two conditions.
1) f(x) is differentiable then we can go with my solution.
and 2) f(x) is not differentiable. Any idea how to prove in this case?

jaus tail said:
2) f(x) is not differentiable. Any idea how to prove in this case?

See @Ray Vickson's solution in post #14.
 
  • Like
Likes jaus tail
  • #46
If you are allowed to use integral and derivatives and all that, follow Ray Vickson solution at post #14. It is mostly correct, and it uses the theorem that ##F(x)=\int_0^x{f(y)dy}## is differentiable and it is ##F'(x)=f(x)## (that is the fundamental theorem of calculus). it doesn't need the assumption that f(x) is differentiable.

If you are not allowed to use integral and derivatives then you can follow my approach explained in post #15, #20 and #28 which I believe is also correct.
 
  • Like
Likes jaus tail
  • #47
WWGD said:
If you are using that to approximate ##f ## in ##[0,1]## then the limit should be finite, as continuous function in compact is bounded. But then you have a limit of the type ##\infty . \epsilon ## which should be dominated by ##\epsilon ## in order for the limit to be finite.

##2^np_n(x)## is dominated by ##p_n(x)## for any ##x\neq\xi## its ##\lim_{n->\infty}{p_n(x)}=0## and the convergence of ##p_n(x)## to 0 is such that it overcomes ##2^n##. I believe I prove this at post #28.
 
  • Like
Likes jaus tail and PeroK
  • #48
Delta² said:
##2^np_n(x)## is dominated by ##p_n(x)## for any ##x\neq\xi## its ##\lim_{n->\infty}{p_n(x)}=0## and the convergence of ##p_n(x)## to 0 is such that it overcomes ##2^n##. I believe I prove this at post #28.

If you take an interval close enough to -1 then you can generate a sequence of function values of increasing magnitude without bound in that interval. Which contradicts continuity, unless the function is zero on the interval. After that your argument simplifies for other points.
 
  • Like
Likes jaus tail

Similar threads

Back
Top