Early in differential geometry, texts typically reformalize our usual gradient, divergence and curl operators as covariant tensors rather than vectors. This is primarily motivated by the observation that they transfer covariantly rather than contravariantly, and therefore need to be assigned a new tensorial structure to govern the transformation rules.
However, in assigning additional structure to these operators, it lends them additional interpretation that does not seem very physically useful.
Take, for example, the curl. Observing that it transforms covariantly, we can assign it a covariant tensor structure. When we want to take the curl of a vector field, we first map the vectors to covectors, take an exterior derivative, and then remap the resulting 2-forms back to vectors.
However, while this achieves our goal of defining the curl in a way that transforms covariantly, we've unintentionally given it a new interpretation to its codomain as maps from 2 vectors to a real number. And, even worse, any vector field is now a map from 2 vectors to a real number via hodge dual. While defining new structures has given us something more useful, it seems to have equivocated our familiar vectors with uninteresting structural baggage in return.
It seems like I'm forced to either, 1) find a useful interpretation of the codomain of the curl operator of the form $(a_x \ dy \wedge dz + a_y \ dx \wedge dz + a_z dx \wedge dy)(v_1, v_2)$ as a useful real number, so that I can be at peace with thinking of the curl as a map from covectors to covectors and not just a vector operator with covariant transformation rules, or 2) accept that sometimes in making physically useful things formal with tensor calculus and the hodge dual, we've created some uninteresting interpretations that simply weren't possible beforehand.
What's a physicist to do?
Answer
This answer takes into account the elaboration on the question which takes place in the comments.
Let's build a tensor space. I won't bother with the definition of a tangent vector as a "directional derivative" operator - we'll just start from the fairly intuitive notion of tangent vectors as little "arrows" which live in the tangent space at a point $p$ in some manifold $\mathcal{M}$.
Vectors
A vector $X$ defined at a point $p\in M$ can be expanded in terms of a vector basis $\hat e_i$. That is, $$X = X^i \hat e_i$$ where we follow the Einstein summation convention.
Covectors
A covector is an $\mathbb{R}-$linear map which eats vectors and returns real numbers. While the vector space is spanned by the vector basis $\hat e_i$, the covector space is spanned by the covector basis $\epsilon_i$. In principle, these bases may be chosen independently of one another, but one typically chooses the covector basis such that $$\epsilon_i(\hat e_j) = \cases{+1 & $i=j$ \\ 0 & $i\neq j$}$$
The action of a covector $\omega = \omega_i \epsilon^i$ on a vector $X = X^i \hat e_i$ is as follows: $$\omega(X) = \omega_i \epsilon^i(X^j \hat e_j) = \omega_i X^j \epsilon^i(\hat e_j) = \omega_i X^i$$
We can similarly define the action of a vector on a covector: $$X(\omega) \equiv \omega(X)$$
Therefore, we can regard vectors and covectors as objects which eat each other and return real numbers.
Tensors
A $(p,q)-$tensor $T$ is an $\mathbb{R}-$multilinear map which eats $p$ covectors and $q$ vectors and returns a real number. From this standpoint, functions on the manifold are $(0,0)-$tensors, vectors are $(1,0)-$tensors, and covectors are $(0,1)-$tensors. As an example, consider a $(1,1)-$tensor $T$. Its action on a covector $\omega$ and a vector $X$ is then $$ T(\omega,X) = T(\omega_i \epsilon^i,X^j \hat e_j) = \omega_i X^j T(\epsilon^i,\hat e_j) = \omega_i X^j T^i_{\ j}$$ where $$T^i_{\ j} \equiv T(\epsilon^i,\hat e_j)$$ are called the components of $T$ in the chosen basis.
Metric Tensors
A metric tensor $g$ is a symmetric, positive-definite $(0,2)$-tensor. Its action on two vectors $X$ and $Y$ can be written $$g(X,Y) = g(X^i\hat e_i,Y^j \hat e_j) = X^i Y^j g(\hat e_i,\hat e_j) = X^i Y^j g_{ij}$$ where the components $$g_{ij}$$ can be put in the form of a symmetric, invertible matrix. This generalizes the notion of the familiar dot product between vectors.
Flat
We can define an operation called flat which takes a vector $X$ to a covector $X^\flat$ by plugging it into the first slot of the metric but leaving the second slot empty. In other words, $$X^\flat \equiv g(X,\bullet)$$ Its action on a vector $Y$ is given by $$X^\flat(Y) = (X^\flat)_i \epsilon^i(Y^j \hat e_j) = (X^\flat)_i Y^j \epsilon^i(\hat e_j) = (X^\flat)_i Y^i$$ but $$X^\flat(Y) = g(X,Y) = X^i Y^j g_{ij}$$ so it follows that $$(X^\flat)_i = g_{ij} X^j$$
Inverse Metric
We can define a $(2,0)-$tensor $\Gamma$ which acts like a metric on covectors rather than vectors. Its defining characteristic would be the fact that
$$\Gamma(X^\flat,Y^\flat) = g(X,Y)$$ Therefore, $$(X^\flat)_m (Y^\flat)_n \Gamma^{mn} = X^i Y^j g_{mi}g_{nj}\Gamma^{mn} = X^i Y^j g_{ij}$$
It follows that the components $\Gamma^{ij}$ are the components of the inverse of the metric tensor written as a matrix. We sometimes write $\Gamma = g^{-1}$, despite the fact that the inverse of a $(2,0)-$tensor is not a meaningful notion. We also often write the components of $\Gamma$ as $g^{ij}$ - that is, the components of the metric tensor with "raised indices." This is a fairly severe abuse of notation, but such is life.
Sharp
Now that the inverse metric has been defined, we can define a sharp operation which takes a covector $\omega$ to a vector $\omega^\sharp$: $$ \omega^\sharp \equiv \Gamma(\omega,\bullet)$$
Musical Isomorphism
Notice that because of our previous definitions, $$(X^\flat)^\sharp = X$$ and $$(\omega^\sharp)^\flat = \omega$$ This correspondence is often called the musical isomorphism.
Differential Forms
A differential $k$-form is totally asymmetric $(0,k)-$tensor. Functions on $\mathcal{M}$ are 0-forms, while covectors are 1-forms. The asymmetry condition starts to be important with 2-forms, which are spanned by the basis $$\epsilon^i \wedge \epsilon^j$$. If $\mathcal{M}$ has dimension 2, then we could have
- 0-forms: $f$
- 1-forms: $\alpha = \alpha_1 \epsilon^1 + \alpha_2 \epsilon^2$
- 2-forms: $\beta = \beta_{12} (\epsilon^1\wedge \epsilon^2)$
where $$\epsilon^1\wedge\epsilon^2 = -\epsilon^2\wedge\epsilon^1$$
No higher forms are possible (or rather, all higher forms are trivially 0) because of the asymmetry condition.
Hodge Star
Notice that on a manifold of dimension $d$, the set of $k-$forms has dimension ${d\choose{k}}$. Therefore, we can set up a correspondence between $k-$forms and $(d-k)-$forms. We call this operation $\star$ - the Hodge star.
In our example of a 2D manifold, the Hodge star takes a 0-form $f$ and returns a 2-form $$\star f = f (\epsilon^1 \wedge \epsilon^2)$$ and vice-versa. The action of the Hodge star on a 1-form is the identity operation.
For a 3D manifold, the Hodge star takes a 0-form $f$ and returns a 3-form $$\star f = f (\epsilon^1\wedge\epsilon^2\wedge\epsilon^3)$$ and vice-versa. It takes a 1-form $$\alpha = a \epsilon^1 + b\epsilon^2 + c\epsilon^3$$ and returns a 2-form $$\star \alpha = a (\epsilon^2\wedge\epsilon^3) + b (\epsilon^3 \wedge \epsilon^1) + c (\epsilon^1 \wedge \epsilon^2)$$ and vice-versa.
So on and so forth.
Exterior Derivative
We can define an operation called exterior differentiation which takes a $k-$form $\alpha$ and returns a $(k+1)-$form $d(\alpha)$. It is defined by the following properties:
- $df=(\partial_i f)\epsilon^i$ for all smooth functions $f$, and
- $d(d(f))\equiv d^2(f) = 0$ for all smooth functions $f$, and
- $d(\alpha \wedge \beta) = d\alpha \wedge \beta + (-1)^p \alpha \wedge d\beta$ where $\alpha$ is a $p-$form and $\beta$ is a $k-$form ($k$ is unspecified)
In our 2D example, given a 0-form $f$, we have that $$d(f) = (\partial_1 f)\epsilon^1 + (\partial_2 f)\epsilon^2$$ and given a 1-form $$\alpha = \alpha_1 \epsilon^1 + \alpha_2 \epsilon^2$$ we have that $$d(\alpha) = (\partial_2 \alpha_1)\epsilon^2\wedge\epsilon^1 + (\partial_1 \alpha_2) \epsilon^1\wedge\epsilon^2=(\partial_1 \alpha_2 - \partial_2 \alpha_1)\epsilon^1\wedge\epsilon^2$$
Notice that $$d(d(\alpha))=0$$ This is no coincidence - $$d^2\alpha = 0$$ for all $k-$forms $\alpha$.
Grad,Curl,Div
We now define the operations $grad$ (which maps functions to vector fields), $curl$ (which maps vector fields to vector fields), and $div$ (which maps vector fields to functions). Note that we restrict our attention to manifolds of dimension 3.
Given a function (aka a $(0,0)-$tensor, or a $0-$form) $f$
- $df$ is a 1-form (a covector field), and
- $(df)^\sharp$ is a vector field
so we let $$grad(f) = (df)^\sharp$$
Given a vector field $X$,
- $X^\flat$ is a covector field (a 1-form), and
- $d(X^\flat)$ is a 2-form, and
- $\star d(X^\flat)$ is a 1-form (a covector field), and
- $[\star d(X^\flat)]^\sharp$ is a vector field
so we let $$curl(X) = [\star d(X^\flat)]^\sharp$$
Lastly, given a vector field $X$,
- $X^\flat$ is a covector field (a 1-form), and
- $\star X^\flat$ is a 2-form, and
- $d(\star X^\flat)$ is a 3-form, and
- $\star d(\star X^\flat)$ is a 0-form (a function)
so we let $$div(X) = \star d(\star X^\flat)$$
We have therefore defined the standard vector calculus operations in the language of differential forms/differential geometry. Notice that it follows immediately that
$$curl(grad(f)) = [\star d([df^\sharp]^\flat)]^\sharp = [\star d^2 f]^\sharp = 0$$ and $$div(curl(f)) = 0$$ (that one is messy due to the symbols we used, but it's pretty straightforward).
No comments:
Post a Comment