Cookies   I display ads to cover the expenses. See the privacy policy for more information. You can keep or reject the ads.

Video thumbnail
- [Narrator] In the last video we were
looking at this particular function.
It's a very non linear function.
And we were picturing it as a transformation
that takes every point x, y in space to the point
x plus sign y, y plus sign of x.
And moreover, we zoomed in on a specific point.
And let me actually write down what
point we zoomed in on, it was (-2,1).
That's something we're gonna want to record here (-2,1).
And I added couple extra grid lines around it
just so we can see in detail what the transformation
does to points that are in the neighborhood of that point.
And over here, this square shows the zoomed
in version of that neighborhood.
And what we saw is that even though the
function as a whole, as a transformation,
looks rather complicated, around that one point,
it looks like a linear function.
It's locally linear so what I'll show you here
is what matrix is gonna tell you the linear
function that this looks like.
And this is gonna be kind of two by two matrix.
I'll make a lot of room for ourselves here.
It'll be a two by two matrix and the way
to think about it is to first go back
to our original setup before the transformation.
And think of just a tiny step to the right.
What I'm gonna think of as a little, partial x.
A tiny step in the x direction.
And what that turns into after the transformation
is gonna be some tiny step in the output space.
And here let me actually kind of draw on
what that tiny step turned into.
It's no longer purely in the x direction.
It has some rightward component.
But now also some downward component.
And to be able to represent this in a nice way,
what I'm gonna do is instead of writing the
entire function as something with
a vector valued output, I'm gonna go ahead
and represent this as a two separate scalar value functions.
I'm gonna write the scalar value functions f1 of x, y.
So I'm just giving a name to x plus sign y.
And f2 of x, y, again all I'm doing is
giving a name to the functions we already have written down.
When I look at this vector, the
consequence of taking a tiny d, x step
in the input space that corresponds to
some two d movement in the output space.
And the x component of that movement.
Right if I was gonna draw this out
and say hey, what's the x component of that movement.
That's something we think of as a little
partial change in f1, the x component of our output.
And if we divide this, if we take you know
partial f1 divided by the size of that
initial tiny change, it basically scales
it up to be a normal sized vector.
Not a tiny nudge but something that's more
constant that doesn't shrink as we
zoom in further and further.
And then similarly the change in the y direction,
right the vertical component of that step
that was still caused by the dx.
Right, it's still caused by that initial
step to the right, that is gonna be
the tiny, partial change in f2.
The y component of the output cause
here we're all just looking in the output space
that was caused by a partial change in the x direction.
And again I kind of like to think about this
we're dividing by a tiny amount.
This partial f2 is really a tiny, tiny nudge.
But by dividing by the size of the initial
tiny nudge that caused it, we're getting
something that's basically a number.
Something that doesn't shrink when
we consider more and more zoomed in versions.
So that, that's all what happens when
we take a tiny step in the x direction.
But another thing you could do, another thing you can
consider is a tiny step in the y direction.
Right cause we wanna know, hey, if
you take a single step some tiny unit upward,
what does that turn into after the transformation.
And what that looks like is this vector
that still has some upward component.
But it also has a rightward component.
And now I'm gonna write its components
as the second column of the matrix.
Because as we know when you're representing
a linear transformation with a matrix,
the first column tells you where the first
basis vector goes and the second column
shows where the second basis vector goes.
If that feels unfamiliar, either
check out the refresher video or
maybe go and look at some of the linear algebra content.
But to figure out the coordinates of this guy,
we do basically the same thing.
Let's say first of all, the change in the x direction
here, the x component of this nudge vector.
That's gonna be given as a partial change to f1, right,
to the x component of the output.
Here we're looking in the outputs base.
We're dealing with f1, f1 and f2
and we're asking what that change was
that was caused by a tiny change in the y direction.
So the change in f1 caused by some tiny step in the y
direction divided by the size of that tiny step.
And then the y component of our output here.
The y component of the step in the outputs base
that was caused by the initial tiny
step upward in the input space.
Well that is the change of f2,
second component of our output as caused by dy.
As caused by that little partial y.
And of course all of this is very specific
to the point that we started at right.
We started at the point (-2,1).
So each of these partial derivatives
is something that really we're saying,
don't take the function, evaluate it at the point (2,-1),
and when you evaluate each one of these
at the point (2,-1) you'll get some number.
And that will give you a very
concrete two by two matrix that's gonna
represent the linear transformation that this
guy looks like once you've zoomed in.
So this matrix here that's full of all
of the partial derivatives has a very special name.
It's called as you may have guessed, the Jacobian.
Or more fully you'd call it the Jacobian Matrix.
And one way to think about it is that it
carries all of the partial differential information right.
It's taking into account both of these components
of the output and both possible inputs.
And giving you a kind of a grid of
what all the partial derivatives are.
But as I hope you see, it's much
more than just a way of recording
what all the partial derivatives are.
There's a reason for organizing it
like this in particular and it really
does come down to this idea of local linearity.
If you understand that the Jacobian Matrix
is fundamentally supposed to represent
what a transformation looks like when you zoom
in near a specific point, almost everything else
about it will start to fall in place.
And in the next video, I'll go ahead
and actually compute this just to
show you what the process looks like.
And how the result we get kind of
matches with the picture we're
looking at, see you then.