How Do We Understand Differentiability and Gradients in Multivariable Calculus?

In summary, Lang discusses the definition of the derivative for functions of several variables by extending the concept from single variable derivatives. He suggests using the limit of the function g(h) as h approaches 0, where g(h) satisfies the equation f(x + h) - f(x) = f(x)h + |h|g(h) with lim h-->0 g(h) = 0. He then redefines everything in terms of vectors, f(X + H) - f(X) = ∇f(X)•H + |H|g(H), for the case of functions of several variables.
  • #1
sponsoredwalk
533
5
For a function ƒ defined on an open set U having the point X:(x1,x2,...,xn)
and the point ||H|| such that the point X + H lies in the set we try to
define the meaning of the derivative.

[tex]\frac{f(X \ + \ H) \ - \ f(X)}{H}[/tex] is an undefined quantity, what does it mean
to divide by a vector?

The idea is to use

[tex] f'(x) \ = \ \lim_{h \to 0} \frac{f(x \ + \ h) \ - \ f(x)}{h}[/tex]

and define

[tex] g(h) \ = \ \lim_{h \to 0} \frac{f(x \ + \ h) \ - \ f(x)}{h} \ - \ f'(x)[/tex]

for h ≠ 0 but notice that [tex] \lim_{h \to 0} \ g(h) \ = \ 0 [/tex].

If we then write

[tex] g(h) \ + \ f'(x) \ = \ \frac{f(x \ + \ h) \ - \ f(x)}{h} [/tex]

[tex] g(h) \cdot h \ + \ f'(x) \cdot h \ = \ f(x \ + \ h) \ - \ f(x) [/tex]

this holds true as long as h ≠ 0 because if h = 0 the derivative term, f'(x) makes no sense.

If the above relation holds for a function whether we've chosen h to be either positive or
negative we can then claim the function differentiable seeing as the limit exists
& the relation holds true. Therefore h can be rewritten as |h| and the above as

[tex] g(h) \cdot |h| \ + \ f'(x) \cdot h \ = \ f(x \ + \ h) \ - \ f(x) [/tex]

(My book only gives the h indicated above with |h| & I'm not sure why Lang doesn't give
the h beside the f'(x) absolute value signs one too? I think it's because he divides in a
minute by h & wants to get rid of the h beside the f'(x) & leave the g(h)•h one but it seems
to me like picking & choosing things as you please...)


So, if the function is differentiable there exists a function g such that

[tex] g(h) \cdot |h| \ + \ f'(x) \cdot h \ = \ f(x \ + \ h) \ - \ f(x) \ , \ \lim_{h \to 0} \ g(h) \ = \ 0 [/tex]

so we can define the following

[tex] g(h) \cdot \frac{|h|}{h} \ + \ f'(x) \ = \ \frac{f(x \ + \ h) \ - \ f(x)}{h} [/tex]

and in the limit we find equality above as the g(h) term makes the fraction of |h|/h
dissappear.

Is all of this correct? What is with the absolute value situation? Why are we picking &
choosing our definitions with |h| when it's convenient? (Or, more accurately - what am I not getting? :-p)

Thanks for taking the time to read that! :approve:
 
Physics news on Phys.org
  • #2
This exposition is very confusing; I think they're trying to give an "intuitive" reason why we would choose the definition of derivative in a particular way, rather than a more rigorous argument.


IMO, differential approximation is one of the cleanest ways to deal with these things. We want:
[tex]f(\vec{x} + \vec{y}) \approx f(\vec{x}) + f'(\vec{x}) \vec{y}[/tex]​

And so we insist that the error term vanish as [itex]\vec{y} \mapsto 0[/itex] more rapidly than [itex]\vec{y}[/itex] itself. So:
[tex]\lim_{\vec{y} \mapsto 0} \frac{f(\vec{x} + \vec{y}) - f(\vec{x}) - f'(\vec{x}) \vec{y}}{|\vec{y}|} = 0[/tex]​
or equivalently
[tex]\lim_{|\vec{y}| \mapsto 0} \frac{|f(\vec{x} + \vec{y}) - f(\vec{x}) - f'(\vec{x}) \vec{y}|}{|\vec{y}|} = 0[/tex]​

([itex]f'(\vec{x})[/itex] should be a "covector" -- something that maps vectors to numbers in a special way -- but any such thing can be written instead as a dot product with a vector, which is why you usually see vectors and dot products instead)



Now that I looking over again the expression from the book, it's not so bad. Everything before your parenthetical makes sense for real-valued functions of a single variable. His goal is to find a characterization of the derivative that would also make sense for the multi-variate case. Of course, he has the benefit of knowing the "right" answer beforehand. :smile: Note that for univariate functions,
[tex]\lim_{h \rightarrow 0} \frac{g(h)}{h} = 0[/tex]​
is true if and only if
[tex]\lim_{h \rightarrow 0} \frac{g(h)}{|h|} = 0[/tex]​
is true. Also, these are true if and only if
[tex]\lim_{h \rightarrow 0} \frac{|g(h)|}{|h|} = 0[/tex]​
is true. So the three formulations of derivative are equivalent.

The only other reasonable guess I see would be to define the error term g(h) as a vector, and have g(h).h as a dot product. But if you think about it, you'll see this is unsatisfactory because this says we could have massive error in our differential approximation, as long as it's perpendicular to h.
 
  • #3
I think I get what you're talking about, I'm reading Lang's Multivariable calculus book & it's a
bit unconventional in that vectors are written as X instead of [tex]\overline{x}[/tex] and
other little things I've noticed. I'll read a more modern book later :wink:

I take it that [tex]\overline{y}[/tex] in your post corresponds with H in mine & that you're
method is a vector version of single variable differential approximation.

I like what Lang did here, I've read less rigorous expositions on multivariable derivatives & they
confused a little but I'm liking this, he's extended the concept of the single variable
derivative for vectors by doing all of what I posted in my first post then redefined things.

When I take the limit in the last latex frame of my first post this is just Lang giving a
justification that it does exist in parallel to the single variable method, but then he says

"
Therefore, the existence of a number a=f'(x) and a function g satisfying
f(x + h) - f(x) = f(x)h + |h|g(h) with lim h-->0 g(h) = 0
could have been used as the definition of differentiability in the case
of functions of one variable. The great advantage of this is that no h
appears in the denominator. It is this relation which will suggest to us
how to define differentiability for functions of several variables, and how
to prove the chain rule for them. "

Lang then redefines everything in terms of vectors
f(X + H) - f(X) = ∇f(X)•H + |H|g(H)

So the idea is that the |H|g(H) dissappears in the limit is it? This would answer your
worry about the dot product masking the huge error in the differential approximation.
I mean he hasn't explicitly said it yet but it seems to be the case.
 
  • #4
sponsoredwalk said:
For a function ƒ defined on an open set U having the point X:(x1,x2,...,xn)
and the point ||H|| such that the point X + H lies in the set we try to
define the meaning of the derivative.

[tex]\frac{f(X \ + \ H) \ - \ f(X)}{H}[/tex] is an undefined quantity, what does it mean
to divide by a vector?
It doesn't mean anything to divide by a vector. You should be dividing by the length of the vector:
[tex]\frac{f(x+ H)- f(x)}{||H||}[/tex]

The idea is to use

[tex] f'(x) \ = \ \lim_{h \to 0} \frac{f(x \ + \ h) \ - \ f(x)}{h}[/tex]

and define

[tex] g(h) \ = \ \lim_{h \to 0} \frac{f(x \ + \ h) \ - \ f(x)}{h} \ - \ f'(x)[/tex]

for h ≠ 0 but notice that [tex] \lim_{h \to 0} \ g(h) \ = \ 0 [/tex].

If we then write

[tex] g(h) \ + \ f'(x) \ = \ \frac{f(x \ + \ h) \ - \ f(x)}{h} [/tex]

[tex] g(h) \cdot h \ + \ f'(x) \cdot h \ = \ f(x \ + \ h) \ - \ f(x) [/tex]

this holds true as long as h ≠ 0 because if h = 0 the derivative term, f'(x) makes no sense.
yes, the derivative is defined as the limit as h goes to 0, not the value when h= 0.

If the above relation holds for a function whether we've chosen h to be either positive or
negative we can then claim the function differentiable seeing as the limit exists
& the relation holds true. Therefore h can be rewritten as |h| and the above as

[tex] g(h) \cdot |h| \ + \ f'(x) \cdot h \ = \ f(x \ + \ h) \ - \ f(x) [/tex]

(My book only gives the h indicated above with |h| & I'm not sure why Lang doesn't give
the h beside the f'(x) absolute value signs one too? I think it's because he divides in a
minute by h & wants to get rid of the h beside the f'(x) & leave the g(h)•h one but it seems
to me like picking & choosing things as you please...)


So, if the function is differentiable there exists a function g such that

[tex] g(h) \cdot |h| \ + \ f'(x) \cdot h \ = \ f(x \ + \ h) \ - \ f(x) \ , \ \lim_{h \to 0} \ g(h) \ = \ 0 [/tex]

so we can define the following

[tex] g(h) \cdot \frac{|h|}{h} \ + \ f'(x) \ = \ \frac{f(x \ + \ h) \ - \ f(x)}{h} [/tex]

and in the limit we find equality above as the g(h) term makes the fraction of |h|/h
dissappear.

Is all of this correct? What is with the absolute value situation? Why are we picking &
choosing our definitions with |h| when it's convenient? (Or, more accurately - what am I not getting? :-p)

Thanks for taking the time to read that! :approve:
You have to define "differentiable" before you can define the "derivative" of a
vector valued function.

In general, we say that function f(x), from Rn to Rm is differentiable at x= a if and only if there exist a linear function L from Rn To Rm and a function [itex]\epsilon(x)[/itex] from Rn to Rm such that:
1) [itex]f(x)= f(a)+ L(x- a)+ \epsilon(x- a)[/itex]
and
2) [itex]\lim_{x\to a}\frac{\epsilon(x-a)}{|x-a|}= 0[/itex]

Then we show that the L defined above is unique and say that it is the derivative of f at x= a.
 

FAQ: How Do We Understand Differentiability and Gradients in Multivariable Calculus?

What is differentiability?

Differentiability is a mathematical concept that describes the smoothness of a function. A function is considered differentiable if it has a well-defined derivative at every point in its domain. This means that the function is continuous and has a unique slope at each point.

What is the difference between differentiability and continuity?

Continuity and differentiability are closely related concepts, but they are not the same. A function is continuous if it has no breaks or gaps in its graph, while a function is differentiable if it has a well-defined derivative at every point in its domain. A function can be continuous without being differentiable, but a differentiable function must also be continuous.

What is a gradient?

The gradient is a vector that represents the direction and magnitude of the steepest slope of a function at a given point. It is a generalization of the concept of a derivative to functions with multiple variables. The components of the gradient vector are the partial derivatives of the function with respect to each variable.

How is the gradient used?

The gradient is used in many areas of mathematics and science, including optimization, physics, and machine learning. It can be used to find the maximum and minimum values of a function, determine the direction of fastest change, and solve systems of equations. In machine learning, the gradient is used in the process of gradient descent to minimize a cost function and find the optimal values for a set of parameters.

What is the relationship between differentiability and the gradient?

The gradient of a differentiable function is a measure of its rate of change. If a function is differentiable, its gradient exists at every point and can be used to determine the direction of maximum increase and decrease. In other words, differentiability is a necessary condition for the existence of a gradient, but not all functions that are differentiable have a gradient that can be easily calculated.

Similar threads

Back
Top