There is a lot of confusion around the determinant, and that’s because it isn’t taught properly. To begin talking about volume, you first need to really understand what space is. The key is that points in space like (x1,x2,…,xn) aren’t the thing you actually care about—it’s the values you assign to those points. Suppose you have some generic function, fiber, you-name-it, that takes in points and spits out something else. The function may vary continuously along some dimensions, or even vary among multiple dimensions at the same time. To keep track of this, we can attach tensors to every point:
dx1⊗dx2⊗⋯⊗dxn
Or maybe a sum of tensors:
dx1⊗dx2⊗⋯⊗dxn+dx2⊗dx1+⋯
The order in that second tensor is a little strange. Why didn’t I just write it like
dx1⊗dx2?
It’s because sometimes the order matters! However, what we can do is break up the tensors into a sum of symmetric pieces. For example,
To find all the symmetric pieces in higher dimensions, you take the symmetric group Sn and compute its irreducible representations (re: the Schur-Weyl duality). Irreducible representations are orthogonal, so each of these symmetric pieces don’t really interact with each other. If you only care about one of the pieces (say the antisymmetric one) you only need to keep track of the coefficient in front of it. So,
and this is where the determinant comes from! It turns out that our physical world seems to only come from the antisymmetric piece, so when we talk about volumes, we’re talking about summing stuff like
Strongly disagree that this is the right way to think about the determinant. It’s pretty easy to start with basic assumptions about the signed volume form (flat things have no volume, things that are twice as big are twice as big) and get everything. Or we might care about exterior algebras, at which point the determinant pops out naturally. Lie theory gives us a good reason to care about the exterior algebra—we want to be able to integrate on manifolds, so we need some volume form, and if we differentiate tensor products of differentiable things by the Leibniz rule, then invariance becomes exactly antisymmetry (in other words, we want our volume form to be a Haar measure for SL(V)). Or if we really want to invoke representation theory, the determinant is the tensor generator of linear algebraic characters of GL(V), and the functoriality that implies is maybe the most important fact about the determinant. Schur-Weyl duality is extremely pretty but I usually think of it as saying something about Sn more than it says something about GL(V). (Fun to think about: we often write the determinant, as you did, as a sum over elements of Sn of products of matrix entries with special indexing, times some function from Sn to C. What do properties of the determinant (say linearity, alternativity, functoriality) imply about that function? Does this lead to a more mechanistic understanding of Schur-Weyl duality?)
The terms you’re invoking already assume you’re living in antisymmetric space.
Signed volume / exterior algebra (literally the space).
Derivatives and integrals come from the boundary operator ∂[12…n]=∑kρ(1↔k)[12…k−1,k+1…n], and the derivative/integral you’re talking about is ρ(σ)=(−1)sgn(σ). That is why some people write their integrals as ∫f(x)∧dx.
It is a nice propery that det:GL(V)→F happens to be the only homomorphism (because sgn is one-dimensional), but why do you want this property? My counterintuition is, what if we have a fractal space where distance shouldn’t be ℓ2 and volume is a little strange? We shouldn’t expect the volume change from a series of transformations to be the same as a series of volume changes.
The terms you’re invoking already assume you’re living in antisymmetric space.
I think this is the other way around. We could decide to only care about the alternating tensors, sure, but your explanation for why we care about this component of the tensor algebra in particular is just “it turns out” that the physical world works like this. I’m trying to explain that we could have known this in advance, it’s natural to expect the alternating component to be particularly nice.
I think my answers to your other two questions are basically the same so I’ll write a longer paragraph here:
We can think of derivatives and integrals as coming from a boundary operator, sure, but this isn’t “the” boundary operator, because it’s only defined up to simplicial structure and we might choose many simplicial structures! In practice we don’t depend on a simplicial structure at all. We care a lot about doing calculus on (smooth) manifolds—we want to integrate over membranes or regions of spacetime, etc. With manifolds we get smooth coordinate charts, local coordinates with some compatibility conditions to make calculus work, and this makes everything nice! To integrate something on a manifold, you integrate it on the coordinate charts via an isomorphism to Rn (we could work over complex manifolds instead, everything is fine), you lift this back to the manifold, everything works out because of the compatibility. Except that we made a lot of choices in this procedure: we chose a particular smooth atlas, we chose particular coordinates on the charts. Our experience from Rn tells us that these choices shouldn’t matter too much, and we can formalize how little they should matter.
There’s a particularly nice smooth atlas, where every point gets local coordinates that are essentially the projection from its tangent space. Now what sorts of changes of coordinates can we do? The integral maps nice functions on our manifold to our base field (really we want to think of this as a map of covectors). That should give us a linear map on tangent spaces, and this map should respect change of basis—that is, change-of-basis should look like some map GLR(n)→R, and functoriality comes from the universal property of the tangent space. We’re not just looking at transformations, locally we’re looking at change of basis, and we expect the effect of changing from basis A to B to C to just be the same as changing from A to C. All we’re asking for is a special amount of linearity, and linearity is exactly what we expect when working with manifolds.
All this to say, we want a homomorphism because we want to do calculus on smooth manifolds, because physics gives us lots of smooth manifolds and asks us to do calculus on them. This is “why” the alternating component is the component we actually care about in the tensor algebra, which is a step you didn’t motivate in your original explanation, and motivating this step makes the rest of the explanation redundant.
(Now you might reasonable ask, where does the fractal story fit into this? What about non-Hausdorff measures? These integrals aren’t functorial, sure. They’re also not absolutely continuous, they don’t respect this linear structure or the manifold structure. One reason measure theory is useful is because it can simultaneously formalize these two different notions of size, but that doesn’t mean these notions are comparable—measure theory should be surprising to you, because we shouldn’t expect a single framework to handle these notions simultaneously. Swapping from one measure to another implies a huge change in priorities, and these aren’t the priorities we have when trying to define the sorts of integrals we encounter in physics.)
(You might also ask, why are we expecting the action to be local? Why don’t we expect GL(n) to do different things in different parts of our manifold? And I think the answer is that we want these transformations to leave the transitions between charts intact and gauge theory formalizes the meaning of “intact” here, but I don’t actually know gauge theory well enough to say more.)
(Third aside, it’s true that derivatives and integrals come from the double differential being 0, but we usually don’t use the simplicial boundary operator you describe, we use the de Rham differential. The fact that these give us the same cohomology theory is really extremely non-obvious and I don’t know a proof that doesn’t go through exterior powers. You can try to define calculus with reference to simplicial structures like this, but a priori it might not be the calculus you expect to get with the normal exterior power structure, it might not be the same calculus you would get with a cubical structure, etc.)
(Fourth aside, I promise this is actually the last one, the determinant isn’t the only homomorphism. It tensor generates the algebraic homomorphisms, meaning det−1,det2,det3,… are also homomorphisms, and there are often non-algebraic field automorphisms, so properly we should be looking at φ∘detk for any integer k and any field automorphism φ. But in practice we only care about algebraic representations and the linear one is special so we care about the determinant in particular.)
To begin talking about volume, you first need to really understand what space is.
No, stop it, this is a terrible approach to math education. “Ok kids, today we’re learning about the area of a circle. First, recall the definition of a manifold.” No!!
There is a lot of confusion around the determinant, and that’s because it isn’t taught properly. To begin talking about volume, you first need to really understand what space is. The key is that points in space like (x1,x2,…,xn) aren’t the thing you actually care about—it’s the values you assign to those points. Suppose you have some generic function, fiber, you-name-it, that takes in points and spits out something else. The function may vary continuously along some dimensions, or even vary among multiple dimensions at the same time. To keep track of this, we can attach tensors to every point:
dx1⊗dx2⊗⋯⊗dxnOr maybe a sum of tensors:
dx1⊗dx2⊗⋯⊗dxn+dx2⊗dx1+⋯The order in that second tensor is a little strange. Why didn’t I just write it like
dx1⊗dx2?It’s because sometimes the order matters! However, what we can do is break up the tensors into a sum of symmetric pieces. For example,
dx2⊗dx1=12[dx1⊗dx2+dx2⊗dx1]symmetric−12[dx1⊗dx2−dx2⊗dx1]antisymmetric.To find all the symmetric pieces in higher dimensions, you take the symmetric group Sn and compute its irreducible representations (re: the Schur-Weyl duality). Irreducible representations are orthogonal, so each of these symmetric pieces don’t really interact with each other. If you only care about one of the pieces (say the antisymmetric one) you only need to keep track of the coefficient in front of it. So,
dx2⊗dx1=−12dx1∧dx2antisymmetric+other pieces.We could also write the antisymmetric piece as
+12dx2∧dx1reverse order!and this is where the determinant comes from! It turns out that our physical world seems to only come from the antisymmetric piece, so when we talk about volumes, we’re talking about summing stuff like
dx1∧dx2∧⋯∧dxn.If we have vectors
a1=a11dx1+a12dx2+⋯+a1ndxna2=a21dx1+a22dx2+⋯+a2ndxn ⋮an=an1dx1+an2dx2+⋯+anndxnthen the volume between them is
a1∧a2∧⋯∧an.Note that
dxi∧dxi=−dxi∧dxi⟹dxi∧dxi=0so we’re only looking for terms where no dxi overlap, or equivalently terms with every dxi. These are
∑σ∈Sn(a1σ1dxσ1)∧(a2σ2dxσ2)∧⋯∧(anσndxσn)or rearranging so they show up in the same order,
∑σ∈Sn(−1)sgn(σ)n∏i=1aiσin⋀i=1dxi.Strongly disagree that this is the right way to think about the determinant. It’s pretty easy to start with basic assumptions about the signed volume form (flat things have no volume, things that are twice as big are twice as big) and get everything. Or we might care about exterior algebras, at which point the determinant pops out naturally. Lie theory gives us a good reason to care about the exterior algebra—we want to be able to integrate on manifolds, so we need some volume form, and if we differentiate tensor products of differentiable things by the Leibniz rule, then invariance becomes exactly antisymmetry (in other words, we want our volume form to be a Haar measure for SL(V)). Or if we really want to invoke representation theory, the determinant is the tensor generator of linear algebraic characters of GL(V), and the functoriality that implies is maybe the most important fact about the determinant. Schur-Weyl duality is extremely pretty but I usually think of it as saying something about Sn more than it says something about GL(V). (Fun to think about: we often write the determinant, as you did, as a sum over elements of Sn of products of matrix entries with special indexing, times some function from Sn to C. What do properties of the determinant (say linearity, alternativity, functoriality) imply about that function? Does this lead to a more mechanistic understanding of Schur-Weyl duality?)
The terms you’re invoking already assume you’re living in antisymmetric space.
Signed volume / exterior algebra (literally the space).
Derivatives and integrals come from the boundary operator ∂[12…n]=∑kρ(1↔k)[12…k−1,k+1…n], and the derivative/integral you’re talking about is ρ(σ)=(−1)sgn(σ). That is why some people write their integrals as ∫f(x)∧dx.
It is a nice propery that det:GL(V)→F happens to be the only homomorphism (because sgn is one-dimensional), but why do you want this property? My counterintuition is, what if we have a fractal space where distance shouldn’t be ℓ2 and volume is a little strange? We shouldn’t expect the volume change from a series of transformations to be the same as a series of volume changes.
I think this is the other way around. We could decide to only care about the alternating tensors, sure, but your explanation for why we care about this component of the tensor algebra in particular is just “it turns out” that the physical world works like this. I’m trying to explain that we could have known this in advance, it’s natural to expect the alternating component to be particularly nice.
I think my answers to your other two questions are basically the same so I’ll write a longer paragraph here:
We can think of derivatives and integrals as coming from a boundary operator, sure, but this isn’t “the” boundary operator, because it’s only defined up to simplicial structure and we might choose many simplicial structures! In practice we don’t depend on a simplicial structure at all. We care a lot about doing calculus on (smooth) manifolds—we want to integrate over membranes or regions of spacetime, etc. With manifolds we get smooth coordinate charts, local coordinates with some compatibility conditions to make calculus work, and this makes everything nice! To integrate something on a manifold, you integrate it on the coordinate charts via an isomorphism to Rn (we could work over complex manifolds instead, everything is fine), you lift this back to the manifold, everything works out because of the compatibility. Except that we made a lot of choices in this procedure: we chose a particular smooth atlas, we chose particular coordinates on the charts. Our experience from Rn tells us that these choices shouldn’t matter too much, and we can formalize how little they should matter.
There’s a particularly nice smooth atlas, where every point gets local coordinates that are essentially the projection from its tangent space. Now what sorts of changes of coordinates can we do? The integral maps nice functions on our manifold to our base field (really we want to think of this as a map of covectors). That should give us a linear map on tangent spaces, and this map should respect change of basis—that is, change-of-basis should look like some map GLR(n)→R, and functoriality comes from the universal property of the tangent space. We’re not just looking at transformations, locally we’re looking at change of basis, and we expect the effect of changing from basis A to B to C to just be the same as changing from A to C. All we’re asking for is a special amount of linearity, and linearity is exactly what we expect when working with manifolds.
All this to say, we want a homomorphism because we want to do calculus on smooth manifolds, because physics gives us lots of smooth manifolds and asks us to do calculus on them. This is “why” the alternating component is the component we actually care about in the tensor algebra, which is a step you didn’t motivate in your original explanation, and motivating this step makes the rest of the explanation redundant.
(Now you might reasonable ask, where does the fractal story fit into this? What about non-Hausdorff measures? These integrals aren’t functorial, sure. They’re also not absolutely continuous, they don’t respect this linear structure or the manifold structure. One reason measure theory is useful is because it can simultaneously formalize these two different notions of size, but that doesn’t mean these notions are comparable—measure theory should be surprising to you, because we shouldn’t expect a single framework to handle these notions simultaneously. Swapping from one measure to another implies a huge change in priorities, and these aren’t the priorities we have when trying to define the sorts of integrals we encounter in physics.)
(You might also ask, why are we expecting the action to be local? Why don’t we expect GL(n) to do different things in different parts of our manifold? And I think the answer is that we want these transformations to leave the transitions between charts intact and gauge theory formalizes the meaning of “intact” here, but I don’t actually know gauge theory well enough to say more.)
(Third aside, it’s true that derivatives and integrals come from the double differential being 0, but we usually don’t use the simplicial boundary operator you describe, we use the de Rham differential. The fact that these give us the same cohomology theory is really extremely non-obvious and I don’t know a proof that doesn’t go through exterior powers. You can try to define calculus with reference to simplicial structures like this, but a priori it might not be the calculus you expect to get with the normal exterior power structure, it might not be the same calculus you would get with a cubical structure, etc.)
(Fourth aside, I promise this is actually the last one, the determinant isn’t the only homomorphism. It tensor generates the algebraic homomorphisms, meaning det−1,det2,det3,… are also homomorphisms, and there are often non-algebraic field automorphisms, so properly we should be looking at φ∘detk for any integer k and any field automorphism φ. But in practice we only care about algebraic representations and the linear one is special so we care about the determinant in particular.)
No, stop it, this is a terrible approach to math education. “Ok kids, today we’re learning about the area of a circle. First, recall the definition of a manifold.” No!!
I always liked this way of looking at the determinant.