0708-1300/Class notes for Tuesday, November 13: Difference between revisions

From Drorbn
Jump to navigationJump to search
 
(8 intermediate revisions by 3 users not shown)
Line 64: Line 64:




Now, <math>dx_2(\frac{\partial}{\partial x}) = \frac{\partial}{\partial x_i}x_2 =\delta_{i2}</math> and hence we get a basis.
Now, <math>dx_j(\frac{\partial}{\partial x}) = \frac{\partial}{\partial x_i}x_j =\delta_{ij}</math> and hence we get a basis.




Line 92: Line 92:




Now differential forms pull back, i.e., for <math>\phi:M\rightarrow N</math> then <math>\phi_*(\lambda)\in\Omega^k(M)\leftarrow\lambda\in\Omega^k(N)</math>
Now differential forms pull back, i.e., for <math>\phi:M\rightarrow N</math> then <math>\phi^*(\lambda)\in\Omega^k(M)\leftarrow\lambda\in\Omega^k(N)</math>
via
via
<math>\phi_*(\lambda)(v_1,\ldots,v_k)=\lambda(\phi_* v_1,\ldots \phi_* v_k)</math>
<math>\phi^*(\lambda)(v_1,\ldots,v_k)=\lambda(\phi_* v_1,\ldots \phi_* v_k)</math>




Line 103: Line 103:




Given M, <math>\exists</math> ! linear map <math>d:\Omega^*(M)\rightarrow\Omega^{k*+1}(M)</math> satisfies
Given M, <math>\exists</math> ! linear map <math>d:\Omega^k(M)\rightarrow\Omega^{k+1}(M)</math> satisfies


1) If <math>f\in\Omega^0(M)</math> then <math>df(X) = X(f)</math> for <math>X\in TM</math>
1) If <math>f\in\Omega^0(M)</math> then <math>df(X) = X(f)</math> for <math>X\in TM</math>
Line 140: Line 140:




Returning, we thus get <math>d(\sum_I f_I dx_I) = \sum_{i,J} \frac{\partial f_I}{\partial x_j} dx_k\wedge dx_I</math>
Returning, we thus get <math>d(\sum_I f_I dx_I) = \sum_{I,j} \frac{\partial f_I}{\partial x_j} dx_j\wedge dx_I</math>


Thus our d takes functions to vector fields by <math>f\mapsto (\frac{\partial f}{\partial x_1},\frac{\partial f}{\partial x_2},\frac{\partial f}{\partial x_3})</math>
Thus our d takes functions to vector fields by <math>f\mapsto (\frac{\partial f}{\partial x_1},\frac{\partial f}{\partial x_2},\frac{\partial f}{\partial x_3})</math>
Line 147: Line 147:




Now let us compute <math>d(h_1 dx_2\wedge dx_3+h_2 dx_3\wedge dx_1+h_3 dx_1\wedge dx_2) =\frac{\partial h_1}{\partial x_1} dx_1\wedge dx_2\wedge dx_3 + \frac{\partial h_1}{\partial x_2} dx_3\wedge dx_2\wedge dx_3 + \frac{\partial h_1}{\partial x_3} dx_3\wedge dx_2\wedge dx_3</math> + 6 more terms representing the 3 partials of each of the last 2 terms.
Now let us compute <math>d(h_1 dx_2\wedge dx_3+h_2 dx_3\wedge dx_1+h_3 dx_1\wedge dx_2) =\frac{\partial h_1}{\partial x_1} dx_1\wedge dx_2\wedge dx_3 + \frac{\partial h_1}{\partial x_2} dx_2\wedge dx_2\wedge dx_3 + \frac{\partial h_1}{\partial x_3} dx_3\wedge dx_2\wedge dx_3</math> + 6 more terms representing the 3 partials of each of the last 2 terms.


As each <math>dx_i\wedge dx_i</math> term vanishes we are left with just,
As each <math>dx_i\wedge dx_i</math> term vanishes we are left with just,


<math>= \frac{\partial h_1}{\partial x_1} + \frac{\partial h_2}{\partial x_2} + \frac{\partial h_3}{\partial x_3})dx_1\wedge dx_2\wedge dx_3</math>
<math>= (\frac{\partial h_1}{\partial x_1} + \frac{\partial h_2}{\partial x_2} + \frac{\partial h_3}{\partial x_3})dx_1\wedge dx_2\wedge dx_3</math>


I.e., d takes <math>(h_1,h_2,h_3)\mapsto \sum_i\frac{\partial h_i}{\partial dx_i}</math>
I.e., d takes <math>(h_1,h_2,h_3)\mapsto \sum_i\frac{\partial h_i}{\partial x_i}</math>


this is just the div operator from calculus and appropriately takes vector fields to functions and represents the d from <math>\Omega^2(M)</math> to <math>\Omega^3(M)</math>.
this is just the div operator from calculus and appropriately takes vector fields to functions and represents the d from <math>\Omega^2(M)</math> to <math>\Omega^3(M)</math>.




We are left with computing d from <math>\Omega^2(M)</math> to <math>\Omega^3(M)</math>
We are left with computing d from <math>\Omega^1(M)</math> to <math>\Omega^2(M)</math>


Computing, <math>d(g_1 dx_1 + g_2 dx_2 + g_3 dx_3) = (\frac{\partial g_3}{\partial x_2} - \frac{\partial g_2}{\partial x_3})dx_2\wedge dx_3 + (\frac{\partial g_1}{\partial x_3} - \frac{\partial g_3}{\partial x_1})dx_3\wedge dx_1 + (\frac{\partial g_2}{\partial x_1} - \frac{\partial g_1}{\partial x_2})dx_1\wedge dx_2</math>
Computing, <math>d(g_1 dx_1 + g_2 dx_2 + g_3 dx_3) = (\frac{\partial g_3}{\partial x_2} - \frac{\partial g_2}{\partial x_3})dx_2\wedge dx_3 + (\frac{\partial g_1}{\partial x_3} - \frac{\partial g_3}{\partial x_1})dx_3\wedge dx_1 + (\frac{\partial g_2}{\partial x_1} - \frac{\partial g_1}{\partial x_2})dx_1\wedge dx_2</math>
Line 172: Line 172:
2) In a say compressible fluid, the divergence corresponds to the difference between in the inflow and outflow of fluid in some small epsilon box around a point.
2) In a say compressible fluid, the divergence corresponds to the difference between in the inflow and outflow of fluid in some small epsilon box around a point.


3) The curl corresponds to the rotation vector for a ball. Ie consider a ball (of equal density to the liquid about it) going down a river. In the x_2, x_1 plane the tenancy for it to rotate clockwise would be given by <math>\frac{\partial g_2}{\partial x_1} - \frac{\partial g_1}{\partial x_2}</math>
3) The curl corresponds to the rotation vector for a ball. Ie consider a ball (of equal density to the liquid about it) going down a river. In the <math>x_2</math>, <math>x_1</math> plane the tenancy for it to rotate clockwise would be given by <math>\frac{\partial g_2}{\partial x_1} - \frac{\partial g_1}{\partial x_2}</math>

Latest revision as of 22:37, 8 February 2008

Announcements go here

Typed Notes

The notes below are by the students and for the students. Hopefully they are useful, but they come with no guarantee of any kind.


First Hour

We begin with a review of last class. Since no one has typed up the notes for last class yet, I will do the review here.

Recall had an association which was the "k forms on M" which equaled

where

which is

1) Multilinear

2) Alternating

We had proved that :

1) is a vector space

2) there was a wedge product via that is

a) bilinear

b) associative

c) supercommutative, i.e.,


From these definitions we can define for and that with the same properties as above.


Claim


If is a basis of then is a basis of and


If and a basis of then any can be written as where are smooth.


The equivalence of these is left as an exercise.


Example

Let us investigate (the * just means "anything").


Now, where and so


Hence,


Now, and hence we get a basis.


So, {vector fields on }


where the are smooth.



This is because to each point p we associate something that takes zero copies of the tangent space into the real numbers. Thus to each p we associate a number.


{functions} where again the k is just a smooth function from to .


{vector fields}


Aside


Recall our earlier discussion of how points and things like points (curves, equivalence classes of curves) pushfoward while things dual to points (functions) pullback and that things dual to functions (such as derivations) push forward. See earlier for the precise definitions.


Now differential forms pull back, i.e., for then via


The pullback preserves all the properties discussed above and is well defined. In particular, it is compatible with the wedge product via


Theorem-Definition


Given M,  ! linear map satisfies

1) If then for

2) . I.e. if and then .

3)

Second Hour

Some notes about the above definition:

1) When we restrict our d to functions we just get the old meaning for d.

2) Philosophically, there is a duality between differential forms and manifolds and that duality is given by integration. In this duality, d is the adjoint of the boundary operator on manifolds. For manifolds, the boundary of the boundary is empty and hence it is reasonable that on differential forms.

3) To remember the formula in 3 given above and others like it, it helps to keep in mind what objects are "odd" and what are "even" and thus when commuting such operators we will get the signs as you would expect from multiplying objects that are either odd or even.


Example

Let us aim for a formula for d on .

Lets compute where and

Then,

The last term vanishes because of (2) in the theorem (proving uniqueness!)


Now, as an aside, we claim that for

Indeed, we know

However, which is the same.


Returning, we thus get

Thus our d takes functions to vector fields by

This is just the grad operator from calculus and we can see that the d operator appropriately takes things from to .


Now let us compute + 6 more terms representing the 3 partials of each of the last 2 terms.

As each term vanishes we are left with just,

I.e., d takes

this is just the div operator from calculus and appropriately takes vector fields to functions and represents the d from to .


We are left with computing d from to

Computing,

I.e., we just have the curl operator.

Note that the well known calculus laws that curl grad = 0 and div curl = 0 are just the expression that .

To provide some physical insight to the meanings of these operators:

1) The gradient represents the direction of maximum descent. I.e. if you had a function on the plane the graph would look like the surface of a mountain range and the direction that water would run would be the gradient.

2) In a say compressible fluid, the divergence corresponds to the difference between in the inflow and outflow of fluid in some small epsilon box around a point.

3) The curl corresponds to the rotation vector for a ball. Ie consider a ball (of equal density to the liquid about it) going down a river. In the , plane the tenancy for it to rotate clockwise would be given by