Matrix trace isn't just summing the diagonal | Lie groups, algebras, brackets #5

  Рет қаралды 194,424

Mathemaniac

Mathemaniac

Күн бұрын

Part 6: • Lie algebras visualize...
Can we visualise this algebraic procedure of adding diagonal entries? What is really happening when we add them together? By visualising it, it is possible to almost immediately see how the different properties of trace comes about.
Files for download:
Go to www.mathemaniac.co.uk/download and enter the following password: traceisdiv
The concept of the whole video starts from one line the Wikipedia page about trace, and I am surprised this isn't on KZfaq: "A related characterization of the trace applies to linear vector fields. Given a matrix A, define a vector field F on R^n by F(x) = Ax. The components of this vector field are linear functions (given by the rows of A). Its divergence div F is a constant function, whose value is equal to tr(A)."
Actually, this is one of the last concepts in linear algebra that I really wanted a visualisation for, the other being transpose, but this is already on the channel: • The deeper meaning of ...
Chapters:
00:00 Introduction
00:48 Matrix as vector field
02:24 Divergence
04:50 Connection between trace and divergence
10:12 Trace = sum of eigenvalues
13:32 Determinant and matrix exponentials
15:15 Trace is basis-independent
18:10 Jacobi's formula
Further reading:
Trace (the origin of the whole video): en.wikipedia.org/wiki/Trace_(...
Divergence (more qualitative, and subtly different from the video): en.wikipedia.org/wiki/Diverge...
Jacobi's formula (more formal proof): en.wikipedia.org/wiki/Jacobi%...
Other than commenting on the video, you are very welcome to fill in a Google form linked below, which helps me make better videos by catering for your math levels:
forms.gle/QJ29hocF9uQAyZyH6
If you want to know more interesting Mathematics, stay tuned for the next video!
SUBSCRIBE and see you in the next video!
If you are wondering how I made all these videos, even though it is stylistically similar to 3Blue1Brown, I don't use his animation engine Manim, but I use PowerPoint, GeoGebra, and (sometimes) Mathematica to produce the videos.
Social media:
Facebook: / mathemaniacyt
Instagram: / _mathemaniac_
Twitter: / mathemaniacyt
Patreon: / mathemaniac (support if you want to and can afford to!)
Merch: mathemaniac.myspreadshop.co.uk
Ko-fi: ko-fi.com/mathemaniac [for one-time support]
For my contact email, check my About page on a PC.
See you next time!

Пікірлер: 187
@mathemaniac
@mathemaniac 4 ай бұрын
EDIT: It should be adjugate matrix, not adjoint. Can't understand why I called it adjoint... There probably wouldn’t be another video until (maybe) Easter, and for some reason I have been quite efficient with this video, and so here you have another video within a week. I am surprised that this intuition of trace hasn’t been picked up on KZfaq, because this is on the Wikipedia page.
@GiovannaIwishyou
@GiovannaIwishyou 4 ай бұрын
Thank you and best of luck with your studies
@enterprisesoftwarearchitect
@enterprisesoftwarearchitect 4 ай бұрын
It’ll take us that long to absorb it in spite of you brilliant elucidation!!
@Stebanoid
@Stebanoid 4 ай бұрын
Wikipedia's math articles are the most abstract and general, so they are not really useful for a non-mathematician. You go to look up what a triangle is and get overwhelmed by non-Abelian algebras, Schwartzshild-Horowitz manifolds and such. Maybe nobody even bothered to look it up.
@mke344
@mke344 4 ай бұрын
​@@Stebanoid that's why it is better to learn from books
@epictaters3738
@epictaters3738 4 ай бұрын
@@Stebanoid I'm just impressed you wrote non-Abelian algebras and Schwartzshild-Horowitz manifolds (terms I just copied)!
@eigenchris
@eigenchris 4 ай бұрын
Very cool. I've never seen any of these interpretations before. Gonna have to watch this one again later. It's interesting how we're both studying Lie groups/algebras, and yet you're going down roads I've never seen or heard of.
@motherisape
@motherisape 4 ай бұрын
Your relativity videos are best on internet. And you are the only one who understands twin paradox clearly thanks for videos
@mathemaniac
@mathemaniac 4 ай бұрын
The visualisation of trace (and another concept of transpose) is something that I have been dying for ever since 3blue1brown makes the linear algebra series. So I have actually been thinking about this for a very long time - but trace relates to Lie algebra (like why does det = 1 condition for SU(n) means that the matrices in the Lie algebra needs to be traceless), and that's why I put it in this series.
@misterlau5246
@misterlau5246 4 ай бұрын
@@mathemaniac heh heh I just commented the Special Unitary group 🤓😅🍜🌶️
@misterlau5246
@misterlau5246 4 ай бұрын
@@motherisape "the only one who understands twin paradox" 🥺 I thought I was doing well 😳 😟 🤓😁😁😁🌶️🍜🉐🈹
@motherisape
@motherisape 4 ай бұрын
@@misterlau5246 I mean he is only teacher on youtube who explain it with full derivation . all other videos are very loose or have few errors .
@Stebanoid
@Stebanoid 4 ай бұрын
I cannot understand why we never were given an explanation of what Tr actually is. We were given its definition and properties and that's it. Until few minutes ago I thought it is some absolutely arbitrary useless thing that was given a name for some weird historical reason. Thank you for your service, good person who made this video!
@YaofuZhou
@YaofuZhou 3 ай бұрын
It is such a blessing that we live in an era when math and physics are explained so clearly, from the fundamentals all the way to quantum field theory.
@7th808s
@7th808s 4 ай бұрын
This is amazing. There was always so much mysticism around the trace. Just a seemingly random choice to study which happened to have all these nice properties. Now it's much clearer why it is important.
@orfeasliossatos
@orfeasliossatos 4 ай бұрын
Super cool video, I've never seen this interpretation of a trace before
@mathemaniac
@mathemaniac 4 ай бұрын
That's why I need to share this! In fact, this intuition is in the Wikipedia page, but somehow not on KZfaq.
@grayjphys
@grayjphys 4 ай бұрын
Love these videos. They make infinitely more sense than reading about the concepts. Especially with the asides along the way to say why a certain concept is important. A lot of texts just barrel forward without that explanation. it makes things less sticky.
@IliaOlkhovskii
@IliaOlkhovskii 4 ай бұрын
Thank you so much for the explanation! I have sought "physical intuition" behind trace for very long during my studies and eventually gave up. It is so satisfying to finally find this beautiful geometrial interpretation =)
@apolloandartemis4605
@apolloandartemis4605 4 ай бұрын
This is beautiful. I once satisfied my curiosity with geometric interpretation of the trace as the directional derivative of a matrix A along the identity matrix, but I now see how much more beautiful and intuitive the trace can be. Thanks!
@alexboche1349
@alexboche1349 4 ай бұрын
Interested to hear more about "directional derivative of a matrix A along the identity matrix" if convenient. Thanks
@AllenKnutson
@AllenKnutson 4 ай бұрын
@@alexboche1349 lim_{t -> 0} (det(Id + tA) - det(Id))/t = trace(A) which I would call instead "derivative along the direction A at the identity"
@horstheinemann2132
@horstheinemann2132 3 ай бұрын
​@@alexboche1349Look up Jacobi's formula if you are still looking for that.
@kkanden
@kkanden 4 ай бұрын
i'm really enjoying this series, keep up the good work!
@derickd6150
@derickd6150 4 ай бұрын
Haven’t had my breath taken away like this since 3blue1browns linear algebra videos all those years ago😢
@inciaradible7144
@inciaradible7144 4 ай бұрын
Awesome video, I never knew these connections and visualisations; really helped me understand all of this better.
@InverseTachyonPulse
@InverseTachyonPulse 3 ай бұрын
Very good explanation. It made me realize how much I like linear algebra but how much I still need to learn and understand (I haven't used it since around 20 years ago)
@jamesknapp64
@jamesknapp64 4 ай бұрын
amazing animations, for the first time I really understood why we found Trace.
@TagRLCS
@TagRLCS 4 ай бұрын
This is absolutely brilliant. I once searched every mathstackexchange topic about trace in the hope of findind an interpretation like this, without success. The fact that you not only give it but also derive all of trace's classical properties seemingly "easily" with it really gives the impression that the true meaning of this operation lies here. I've tried interpreting trace as related to the "perimeter" of the unit square under a matrix (in the case of diagonalizable matrices, the formula itself lead me there, by analogy with tthe determinant) , but it never was completely successful. This video really feels like the missing piece of the puzzle. Thank you so much.
@mathemaniac
@mathemaniac 4 ай бұрын
This is also what I have been thinking for a long time, and decide to put it into this series about Lie theory because it is related. Actually, this intuition is in Wikipedia: "A related characterization of the trace applies to linear vector fields. Given a matrix A, define a vector field F on R^n by F(x) = Ax. The components of this vector field are linear functions (given by the rows of A). Its divergence div F is a constant function, whose value is equal to tr(A)."
@TagRLCS
@TagRLCS 4 ай бұрын
​@@mathemaniac Wow, crazy that I never saw it on Wikipedia. Trace now checks off my list of "interpretation needed" concepts. The last one regarding linear algebra where I feel my interpretation isnt complete are complex eigenvalues. I guess when "scaling" is really scaling and rotating, things get messier...
@jcferreira03
@jcferreira03 4 ай бұрын
​@@TagRLCSI think that complex eigenvalues says to us that the eigenvector is in another dimension, implying that all the original plane is affected by the portion of the linear transformation related to rotation.
@mathemaniac
@mathemaniac 4 ай бұрын
Complex eigenvectors I can provide an intuition for, but I am not sure about complex eigenvalues (it would be a lot more complicated). That will certainly be on my video idea list, and maybe after this video series about Lie theory.
@TagRLCS
@TagRLCS 4 ай бұрын
@@mathemaniac Would truly be incredible. Looking forward to the next Lie theory videos too tho !
@hannahnelson4569
@hannahnelson4569 4 ай бұрын
Good video! I did not previously know the relation between trace and divergence!
@quintonpierre
@quintonpierre 4 ай бұрын
Very nice video, thank you. Looking forward to the next one.
@portport
@portport 4 ай бұрын
Super good content. I didn't know most of these
@pullrequest1296
@pullrequest1296 4 ай бұрын
Incredibly fantastic video!
@ccc3001
@ccc3001 4 ай бұрын
AWESOME!!! Never see this kind of thinking way
@rarebeeph1783
@rarebeeph1783 4 ай бұрын
Representing the divergence as the volume rate of change per volume is interesting. I'm used to thinking of it as the flux per unit volume, that is, the flow rate through the boundary per volume enclosed. They're naturally very related concepts; a net flow out of the boundary naturally results in net expansion of a volume undergoing said flow, and a net flow into the boundary naturally results in net contraction of the volume, and in the limiting case approaching 0 flow time, those amounts must intuitively be equal. Algebraically, the former can be expressed as the determinant of (1+εa, εb; εc, 1+εd)/ε, which as you showed in the video is a+d as ε tends to 0, while the latter can be expressed as the integral of (Av dot dB*) divided by the volume enclosed by B, where dB is a differential boundary element and so dB* is a normal vector of the boundary. In 2d, for a square, basis-vector-aligned choice of boundary of side length ε, this gives ((-(ε^2)c/2) + ((ε^2)a + (ε^2)b/2) + ((ε^2)d + (ε^2)c/2) + (-(ε^2)b/2))/(ε^2) -- intentionally leaving it unsimplified to show the process of finding the integral on each side of the square independently -- which then simplifies to a+d as expected. Edit: I had briefly thought this method has a benefit over your method of not needing to take the limit as ε approaches 0; however I realized that I used ε for different quantities in each case. The limit as flow time goes to 0 was already performed implicitly in my expressing of the instantaneous flow rate at v as Av, so I repurposed ε to represent a parameter of the boundary, which is a property you had already covered by that the divergence of a field generated by a matrix is the same at all locations in the field. So I think your method is pretty much better in every way for thinking about divergence, in that framing it your way implies a much simpler calculation.
@mathemaniac
@mathemaniac 4 ай бұрын
Actually the main reason why I am thinking of rate of change in area / area rather than flux is for the properties described in the video, especially the matrix exponential part. It would be difficult to use the flux intuition of divergence.
@vonneumann6161
@vonneumann6161 4 ай бұрын
I wish the video showed the two interpretations were equivalent. I was confused initially
@adamboussif8035
@adamboussif8035 3 ай бұрын
i'm just getting started with this trying to build a good intuition , and i must say i've had no previous background in calculating flux or divergence , so may i ask what would the dot product be explicitly here ? and what's the normal vector ? Thanks
@rarebeeph1783
@rarebeeph1783 3 ай бұрын
@@adamboussif8035 the normal vector of a portion of a surface is perpendicular to the surface, and often is chosen to have a length corresponding to the area of the portion of the surface in question. the dot product of a vector field with the normal vector of a surface at a point is the product of the parallel components of the field and the normal vector, which, to phrase it more intuitively, means that it expresses how strongly the field is pointing directly through the surface at that point. summing up all those strengths over a closed surface gives how much the field tends to exit or enter the enclosed region overall, which is sorta like the average divergence in the enclosed region.
@loppy1243
@loppy1243 4 ай бұрын
An interpretation of trace I don't see often enough: It is the dimension times the average amount that a linear transformation L preserves direction, i.e. it is dimension*(1/(unit sphere surface area))*(integral of v.L(v) over v in the unit sphere)
@TagRLCS
@TagRLCS 4 ай бұрын
Do you any reading material related to this interpretation ?
@loppy1243
@loppy1243 4 ай бұрын
@@TagRLCS No, but you can derive it from the connection between divergence and trace using the integral definition of divergence: en.m.wikipedia.org/wiki/Divergence#Definition
@TheJara123
@TheJara123 4 ай бұрын
Man, what a surprise!! After your phd story i was saden that there wont be many videos!!
@mathemaniac
@mathemaniac 4 ай бұрын
Please don't expect another video next week - the next video might be somewhere during Easter.
@lerssilarsson6414
@lerssilarsson6414 2 ай бұрын
Yet another channel worth to be subscribed.
@alexboche1349
@alexboche1349 4 ай бұрын
Great! At 8:05 I found it a little hard to understand, but I think what you're saying is the following. Algebraically: adding columns doesn't change the determinant. Geometrically: This can be seen by just drawing a parallelogram and showing it's area equals that of a square by repositioning the outer triangles to form a square. More abstractly, when you move in the horizontal component, you're moving the vertical basis vector in the direction of the horizontal basis vector, so that the angle of the resulting vectors starts to become less than 90 degrees (closer to 45 degrees). But area of the parallelogram (i.e. the determinant) between vectors u,v (which equals |u||v|sinθ) is maximized when the vectors are perpendicular (so θ=90 degrees) so this is bad for the area. So on the one hand you're making the vector norm |u| longer (which increases area ), but on the other hand you're decreasing the angle θ so on balance the effect is zero.
@ianmcewan8851
@ianmcewan8851 4 ай бұрын
Really nicely done!
@angeldude101
@angeldude101 4 ай бұрын
Yet another time when thinking about matrices as grids of numbers leads one astray. I'd heard that the trace is the number of entries on the diagonal, but no explanation for why that should even be a meaningful operation. Thank you for actually explaining trace in a way that actually makes it seem important. The idea of trace as divergence I hadn't considered, but makes a surprising amount of sense considering the matrix forms of elements from Clifford algebras. There, the "geometric" product of two vectors can be seen as a sum of their inner and exterior products. The exterior product product is easy to see in terms of matrices as it's ½(uvᵀ - vuᵀ) (or maybe the other way around), which neatly parallels how it's defined in Clifford algebra as ½(uv - vu) (with the geometric product rather than the matrix outer product). The inner product, which is what's more traditionally used to find divergences, however is harder to connect with the matrix representations. adapting ½(uv + vu) as ½(uvᵀ + vuᵀ) gives a diagonal matrix, but not a scaled identity matrix like you'd expect a scalar to be. However, if you add up the diagonal (and remove the factor of a half), you _do_ get the scalar that you'd normally get from the dot product. Adding up the diagonal in this case giving the trace. So algebraically, it does make sense that trace would be related to divergence, since the inner product is also related to divergence and trace appeared to be related to the inner product.
@evandrofilipe1526
@evandrofilipe1526 4 ай бұрын
I was looking for the geometric algebra interpretation, somehow I knew it would be here. Thanks
@kelly4187
@kelly4187 2 ай бұрын
This always felt more intuitive after I studied statistics, where for a covariance matrix the main diagonal elements are the variances of the data, so describing how "spread out" data is from that variable.
@Dr.HarshTruth
@Dr.HarshTruth 4 ай бұрын
Such a high quality content!
@GUTS-vw7rs
@GUTS-vw7rs 4 ай бұрын
wow i am on this topic in my course and you just made a vid on it thanks
@ffs55
@ffs55 3 ай бұрын
Amazing video. It would be a prayer answered if you would cover the meaning of successive powers of a [square] matrix, particularly the trace of each successive power when the matrix is a graph adjacency matrix (and therefore triangular-symmetric). Big tie-in with quantum theory!!
@gijsbartholomeus1973
@gijsbartholomeus1973 4 ай бұрын
Really cool and daring that you go into these topics which are generally a little bit more advanced than say 3b1b content
@bacon_boat2641
@bacon_boat2641 3 ай бұрын
I love this playlist!
@7th808s
@7th808s 4 ай бұрын
I think a better visualization of the rate of area change might be the following: First create two parallellograms generated from A(t) and A(t +dt) with areas det(A(t)) and det(A(t + dt)). From the definition of the derivative and making dt very small, the rate of change becomes: d/dt(det(A(t))). Then go back to the parallellograms. The vectors [A(t +dt) - A(t)](1,0) and [...](0,1) describe how the first parallellogram transforms into the second. Again, through the definition of the derivative and making dt very small we get: d/dt(A(t))(1,0) and d/dt(A(t))(0,1). These vectors don't act on the unit square but on the parallellogram generated by A(t). So we should first normalize - so to say - the area of the A(t) parallellogram to get the effect of the vertical and horizontal changes which generate the trace. So we can act A^-1(t) on it, but we shouldn't forget to multiply the result with det(A(t)) if we still want to measure the same thing. Here the fact is used that multiplying by A^-1(t) constitutes a factor det(A^-1(t)) on any area, which equals det(A(t))^-1. So we find that the same rate of change in area is det(A(t))*Tr(A^-1(t)d/dt(A(t))). And since we already established that the rate of change of this area is equal to d/dt(det(A(t))), we get the desired result.
@mathemaniac
@mathemaniac 4 ай бұрын
That is basically the last bit of the video!
@7th808s
@7th808s 4 ай бұрын
​@@mathemaniac No, my point is that you don't explain the change of area in terms of the literal change of one parallellogram towards the other, but keep the story in terms of vectors and formulas, which makes the explanation very algebraic. You point to the formula and say "this term means this and that term means that, and if we cleverly rearrange it, you get the desired result". This is the opposite of an intuitive visualization; this is closing yourself off from reality, playing around with a formula, getting a result, and then afterwards checking back in with reality to see what you just did, i.e.: Algebra. My method is that there are two ways to calculate the rate of area change: The difference in area of before and after; and the effect of an infinitesimal change in A on the unit square applied to the before parallellogram. And then just equate the two. This way the explanation is kept on a very fundamental - dare I say high school level - understanding of dealing with areas of shapes.
@Iamfafafel
@Iamfafafel 4 ай бұрын
​@@7th808s can you update with some pictures? you probably have a specific visualization in mind which is easier to convey than with text
@schobihh2703
@schobihh2703 4 ай бұрын
I wonder, how many watching this could really follow. Some of the underlying math was gently skipped to follow from one to the other step of argumentation. Took me some thoughts to fiddle the pieces together. But a nice visualization.
@arduous222
@arduous222 4 ай бұрын
Great video! It would probably help the viewers if you made a blue dot in the (0,0), because some visual arguments you make seem to rely on (0,0) being mapped to itself. Or you could just draw axes for x and y.
@mathemaniac
@mathemaniac 4 ай бұрын
Yeah, probably a blue dot here would help. The axes would have been very distracting from the already crowded picture of vectors.
@maurocruz1824
@maurocruz1824 3 ай бұрын
The last two videos are more entangled than the first ones. I couldn't follow a lot of details. But i gonna watch the whole series.
@rodrigoappendino
@rodrigoappendino 4 ай бұрын
8:40 It's not clear to me that the horizontal component wouldn't change the area. For me, it looks like the both components would not change the area. 8:25 The red vertex I understand that it's changing to the right because of that vectors horizontal component, but the vertex above shouldn't change to the left, since the vector in that vertex has a horizontal component pointing to the left?
@alexboche1349
@alexboche1349 4 ай бұрын
Regarding 8:40, it's a "second-order" effect (i.e. the effect is only zero at the margin, then it becomes non-zero). When you move in the horizontal component, you're moving the vertical basis vector in the direction of the horizontal basis vector, so that the angle of the resulting vectors starts to become less than 90 degrees (closer to 45 degrees). But area of the parallelogram between vectors u,v (which equals |u||v|sinθ) is maximized when the vectors are perpendicular (so θ=90 degrees) so this is bad for the area. So on the one hand you're making the vector norm |u| longer (which increases area ), but on the other hand you're decreasing the angle θ so on balance the effect is zero (at the margin).
@baronvonbeandip
@baronvonbeandip Ай бұрын
I'm working on module theory right now and some of these results are paying the lighting bills in my head
@zacklee5787
@zacklee5787 2 ай бұрын
What do you mean by the vertical and horizontal components not mattering? I see they wouldn't matter for just a small area around the unit vectors, but we are considering the whol unit square.
@VeteranVandal
@VeteranVandal 3 ай бұрын
Isn't the trace a particular case of a convolution of a diagonal? Because you could use a more general convolution instead, but, since you generally want to have a square diagonalizable matrix, you usually don't need to, and this tool does enough. It's slightly circular, but the argument is: all info is in the vectors and values, and all else you need to store is the map that takes you from the old to the new matrix. Otherwise, you actually don't have cozy stuff like eigenvalues and eigenvectors and you have to settle on different tools, yes? And those can be defined like the sum of other diagonals and are occasionally useful. Or, at least, that's what I thought it was the thing we did it for, but I might have misunderstood it.
@miketaylor3947
@miketaylor3947 4 ай бұрын
Well done!
@eofirdavid
@eofirdavid 4 ай бұрын
I have been around mathematics most of my life, and it was really hard understanding what you are trying to do. Then I saw the "Lie groups, algebras" in the title and everything fell into place. I think that you should really start such explanations with the simplest non trivial example, instead of some abstract numbers (maybe diagonal, or unipotent matrix).
@drdca8263
@drdca8263 4 ай бұрын
At 5:25 , what you are saying is, comparing x + epsilon A x, to x + d + epsilon A (x + d) , that the difference is d + epsilon A d, correct? Edit: wow, ok, so it seems like something pretty similar to “the trace is the derivative of the determinant” can be used pretty well as a definition of the trace? I guess specifically, tr(A) := (d/dt)(det(1 + t A))|_{t=0} (And then later getting the Jacobi formula from the properties you first get from that)
@mathemaniac
@mathemaniac 4 ай бұрын
Yes.
@cmilkau
@cmilkau 4 ай бұрын
Might want to mention why the evolution of the unit square will remain a parallelogram with one corner at the origin (it all directly follows from linearity of A and the equation of evolution, but the same could be said for everything in the video).
@TheJara123
@TheJara123 4 ай бұрын
In that case we can wait!!
@samtux762
@samtux762 4 ай бұрын
I studied linear algebra for a semester. Noone explained us why we care about traces, determinants or eigenvalues/iegenvectors. "Shut up and compute" was the approach. My University was good, but I wasn't in math department. Curious to know why these metrics are important.
@vinbo2232
@vinbo2232 4 ай бұрын
Compare this explanation to the formal def, the key is to recogonize the "flux" as a "generalized area changing rate per unit area".
@cbbuntz
@cbbuntz 2 ай бұрын
I'm self-taught, so bear with my poor grasp on terminology. I suspect there's a relationship to this and the characteristic polynomial. Let K be a Krylov subspace of A. (K^-1 A K) yields a companion matrix C of A. C describes the matrix transformation of A as a recurrence relation defined by the characteristic polynomial. There's an implied leading 1 of the characteristic polynomial, so it's already in monic form. The first non-trivial coefficient is the second one, which happens to also be the trace of the matrix A. (It also happens to be the only nonzero term on the diagonal of C). For example a companion matrix with characteristic polynomial x^n - r x^(n-1) will generate a geometric progression with a common ratio r. The trace of that companion matrix is r. So in (over)simplified terms, the trace acts like a common ratio/exponential base. There's might be a better way to describe this using the matrix exponential. This is actually my first time seeing that det(exp(t A)) = exp(t tr A) equation, but that appears that it might be describing what I'm struggling to explain. It would appear that this is the same phenomenon described in a different way
@apteropith
@apteropith 4 ай бұрын
i'll have to look at these again at some point, because i had difficulty following halfway through (various local distractions compounding other factors, such as my dislike of matrix notation) i've understood most properties of the trace, in the past, through observing that the trace can also be the delayed completion of a scalar product between two vectors, of which the matrix is the outer product; i have not checked how generalizable this is to square matrices larger than 2x2, but i have doubts i'm also not positive how to compare it to my understanding of the determinant, the dimensionality of which is locked to n for an nxn matrix, but has more flexible equivalents in algebras i *do* like (particularly geometric algebra), but that's mostly down to severe executive dysfunction at this point
@japedr
@japedr 4 ай бұрын
For the third property you could have used the fact that eignevalues are basis-independent, so their sum (trace) must also be basis-independent.
@mathemaniac
@mathemaniac 4 ай бұрын
Yes, but I want to use another, perhaps more direct perspective, which would be useful for showing the fourth property, and would be useful in developing intuition for what AB-BA actually means (but that's a story for another time).
@japedr
@japedr 4 ай бұрын
I see, that makes sense BTW, thanks for your work!
@peterwaksman9179
@peterwaksman9179 4 ай бұрын
A smart take.
@sinecurve9999
@sinecurve9999 4 ай бұрын
That moment when you realize why the trace of the density matrix matters in statistical mechanics...
@mathemaniac
@mathemaniac 4 ай бұрын
I would love to know more about statistical mechanics - but I have definitely seen trace of an operator somewhere in QM.
@maestraccivalentin316
@maestraccivalentin316 4 ай бұрын
For the first property you only proved it when A is diagonalisable but then it's kinda obvious from Prop 3 right? Or did I miss something?
@ecdavek230
@ecdavek230 3 ай бұрын
Wow ! ! Thank you
@TheFallenTitan
@TheFallenTitan 4 ай бұрын
Pretty cool seeing the trace as a divergence! Is there an equivalent intuition for the curl of the vector field generated by the matrix?
@mathemaniac
@mathemaniac 4 ай бұрын
Curl is a very 3-dimensional concept. You can actually compute yourself what the curl of such a vector field would be, but I'm not sure if the resulting vector has a particularly well-known meaning.
@smolboi9659
@smolboi9659 4 ай бұрын
The reason why curl is a 3D concept is because it is the cross product of the grad operator with a vector. Cross product is a 3D concept. This is because in 3D there is only 1 direction orthogonal to both the vectors you are crossing. In 4D for example you have 2 directions orthogonal to the vectors you are crossing. The Hodge Star operator sort of generalizes the cross product by saying let the "cross product" be the whole orthogonal (n-2)D hyperplane. For example in 4D the "cross product" is the whole orthogonal plane. This object is a bivector. You can use this Hodge Star Operator to generalize the curl but I'm not even sure how to begin visualizing it.
@alexboche1349
@alexboche1349 2 ай бұрын
@@smolboi9659 I was recently learning the Hodge star operator. I wonder why it is not more widely taught? I think physicists have screwed up math ed by making everything 3-d when it really should be n-d.
@smolboi9659
@smolboi9659 2 ай бұрын
@@alexboche1349 yea u really want to start with learning geometric calculus instead of standard multivariable calculus with curl and div. It makes so much more pedagogical sense. I linked some motivation videos for why we should start with geometric algebra/calculus. kzfaq.info/get/bejne/ntqJm5CF1p2-cn0.htmlsi=Ht06EnTdo6Ef5VWJ kzfaq.info/get/bejne/bJaqksuhqqfYdGw.htmlsi=ppzdznrbtTf-zCTK
@smolboi9659
@smolboi9659 2 ай бұрын
@@alexboche1349 yea it really makes more pedagogical sense to start with geometric calculus than vector calculus. Checkout "why can't you multiply vectors" by freya holmer and "a swift introduction to geometric algebra" by sudgylacmoe on youtube.
@yNot-uu2uz
@yNot-uu2uz 5 күн бұрын
very interesting
@DynestiGTI
@DynestiGTI 4 ай бұрын
8:05 to anyone confused as to why the area is changing like that, remember that at the point (0,0) the vector *Ax* is zero, i.e it’s fixed.
@cmilkau
@cmilkau 4 ай бұрын
15:30 Actually, the whole divergence argument was basis independent, so is a basis change argument even necessary?
@rodrigoappendino
@rodrigoappendino 4 ай бұрын
Where do you learn stuff like that idea of divergent? The books I normally use don't give me intuitions like that, so I have to search for videos like this one, but you probably know this because you read a book, right?
@mathemaniac
@mathemaniac 4 ай бұрын
This intuition is actually on the Wikipedia page about trace: "A related characterization of the trace applies to linear vector fields. Given a matrix A, define a vector field F on R^n by F(x) = Ax. The components of this vector field are linear functions (given by the rows of A). Its divergence div F is a constant function, whose value is equal to tr(A)." In most cases (apart from complex analysis), my videos come from online resources or actually my own thoughts, not books.
@larryfitzgerald7538
@larryfitzgerald7538 3 ай бұрын
Love this video, I am investing eigenvalues in my research, I would love to get some insight from you.
@yashtrivedi9403
@yashtrivedi9403 2 ай бұрын
Hey, I am planning to make educational modern physics videos. I'm amazed by your work. I have read your channel's description and I liked that you produce 3b1b quality animation just by power point, etc. please help by guiding the approach that you take on an animating diagram or maths part.
@146fallon9
@146fallon9 4 ай бұрын
I have an idea. A linear transformation A can be written as (so the matrix of A is [a b; c d]). Then, tr(A) is a + d and div(A) = d/dx (ax + by) + d/dy (cx + dy) = a + d so tr(A) = div(A)
@Iamfafafel
@Iamfafafel 4 ай бұрын
no this doesn't type check. the divergence of a (0,2) tensor is a (0,1) tensor, but trace gives a (0,0) tensor
@98danielray
@98danielray 2 ай бұрын
​@@Iamfafafelthats because they are doing div(Av)
@naratipmath
@naratipmath 4 ай бұрын
As a teacher, I would love to learn how to create a video like this! Any guidance about where to start?
@mathemaniac
@mathemaniac 4 ай бұрын
You can download the files following the link in the description.
@jonetyson
@jonetyson 4 ай бұрын
Thanks!
@glynnec2008
@glynnec2008 4 ай бұрын
Is the proper name for adj(A) "adjoint" or "adjugate" ?
@mathemaniac
@mathemaniac 4 ай бұрын
Yes, can't believe I called it adjoint. I have pinned the correction to it.
@zoetropo1
@zoetropo1 4 ай бұрын
If trace is the divergence of a matrix's vector field, then what is its curl?
@mathemaniac
@mathemaniac 4 ай бұрын
As I replied to another comment, curl is a very 3-dimensional concept, and even for a 3-dimensional matrix A, I don't think the curl of Ax (which is a vector) connects to a well-known thing in linear algebra.
@loulephille
@loulephille 3 ай бұрын
you're a real e^x person
@joshuagrumski7459
@joshuagrumski7459 4 ай бұрын
I have a question. Does your proof that tr(A)= sum of eigenvectors work for nondiagonalizable matrices? Because if the matrix isn’t diagonalizable, then there does not exist 2 eigenvectors, but there is an eigenvalue with multiplicity 2, so trace of the matrix will just be twice of that eigenvalue. I know I’ve done the proof that tr(A) is sum of eigenvalues. I think a nice one involves the characteristic polynomial, iirc, but I do wanna understand this visualization. It’s not like nondiagonalizable matrices have to be singular and/or have no meaningful vector field. Just take the classic example of the matrix with eigenvalue 1 and a 2D Jordan block.
@mathemaniac
@mathemaniac 4 ай бұрын
I mean, even for matrices like A = (0,-1; 1,0), which describes a rotational vector field, the argument does not work, and you would have to think about how complex eigenvalues / complex eigenvectors work there. However, it would still be true that tr(A) is the divergence, but I am not entirely sure how to say in general tr(A) = sum of eigenvalues in those weird cases. One thing you might be interested in is to look at a dynamical systems text. The vector fields as illustrated here can be thought as a linearisation of even more general vector fields (and the matrix that generates the linear vector field is actually the Jacobian matrix), and these are the objects of study in dynamical systems. For non-diagonalisable matrices, these are known to be quite difficult to deal with. In short, I don't know how to interpret eigenvalues in those contexts, at least in this vector field interpretation. They are in some sense more "algebraic" in nature.
@joshuagrumski7459
@joshuagrumski7459 4 ай бұрын
@@mathemaniac Yeah, I know from like diff geo that most of the time, Riemannian manifolds are a bit easier to deal with than symplectic manifolds, where the metric no longer needs to be Hermitian, and therefore diagonalizable. I have a professor who does research in dynamical and control systems, so I may ask him for text recommendations. One thing that may help with understanding this is det(exp(tA))=exp(tr(tA)), so in particular, det(exp(tA))=exp(t*tr(A)), and you may be able to do a proof with this that is more "general" by taking some derivative? Idk, I don't have thaaaat much experience with dynamical systems, but this gives me the vibes of a less visual proof but one which I may see in a text on like diff geo. And yeah, it makes sense that the Jacobian would be the the matrix that generates the linear vector field. We have that v(x)=Ax, so simply dv=A.
@angeldude101
@angeldude101 4 ай бұрын
​@@mathemaniac The matrix i has imaginary eigenvalues‽ :O So multiplying by i is like... multiplying by i‽ It still feels awkward for me to use ℂomplex scalars in matrices, when ℂomplex numbers themselves are already rotation matrices, and pure 2D rotation matrices are themselves ℂomplex numbers.
@joshuagrumski7459
@joshuagrumski7459 4 ай бұрын
@@angeldude101sometimes it’s not useful to think of imaginary numbers as “rotation matrices,” but rather as “elements of the algebraic completion of the reals.” So, although the space of complex numbers is, for example, isomorphic to the field of skew symmetric matrices, it’s sometimes not useful to think of it in those terms. Sometimes we want to think of it in a more abstract sense. For example, real numbers can be understood to actually be sets (Dedekin Cuts), but I have no problem putting them in sets because I’m not really thinking of them as sets. We can use matrices to show that there is a nice construction for the complex numbers in terms of the real numbers, then say “ok, now, suppose you have some ‘representation’ of the complex numbers… what is possible?”
@cbbuntz
@cbbuntz 2 ай бұрын
@@joshuagrumski7459I just typed out a response explaining how complex numbers operations are just matrix operations and that imaginary eigenvalues show up when rotation is involved, just like with complex numbers, but then I read that your comment basically said the same thing except written better
@dsolis7532
@dsolis7532 Ай бұрын
My quantum mechanics classes would have made way more sense with this explanation and just “Welp, it’s trace because Von Neumann said so”
@dr.kraemer
@dr.kraemer 4 ай бұрын
cool stuff, but I lost the thread during the last part. maybe I'll take another look after my brain has had time to chew on the rest.
@Zealot0630
@Zealot0630 4 ай бұрын
it feels that trA is the partial derivative of Ax w.r.t x, where x is an manifold of space A.
@mathemaniac
@mathemaniac 4 ай бұрын
Yes in a vague sense. More precisely, the divergence of Ax (which is in some sense "partial derivative w.r.t. x") is the trace of A.
@jaakjpn
@jaakjpn 4 ай бұрын
The eigenvectors should be orthogonal to each other. I wonder why in the "Trace = sum of eigenvalues" visualization the eigenvectors are not orthogonal. Thanks for the nice video!
@mathemaniac
@mathemaniac 4 ай бұрын
Eigenvectors are not necessarily orthogonal to each other - in fact they might not even exist (in real numbers).
@Zealot0630
@Zealot0630 4 ай бұрын
Such matrix (whose eigenvalues are orthogonal) are called hermition matrix for complex matrix, or symmetric matrix for real matrix.
@joshuagrumski7459
@joshuagrumski7459 4 ай бұрын
@@Zealot0630not to be the “um actually” guy, but your matrix doesn’t need to be Hermitian or symmetric to have orthogonal eigenvectors. They need to be “normal,” which means that A^T commutes with A (choose dagger instead of T for a vector space over the complex numbers). Unitary matrices, for example, are also normal, and not all Unitary matrices are Hermitian. And now I think about it, I can’t remember if it’s a biconditional statement or if normal implies orthogonal eigenvectors in one direction only.
@planaritytheory
@planaritytheory 4 ай бұрын
@@joshuagrumski7459 If we can't "um actually" in mathematics, then when can we?
@misterlau5246
@misterlau5246 4 ай бұрын
Not "orthogonal" per se, but normal. Commuting for example. And, there can be orthogonality with a vector in Z too, and then we extend to planes, surfaces and we get to folds, 🤓😁🌶️🍜
@user-ug3gq8bk9t
@user-ug3gq8bk9t 4 ай бұрын
I know the formula 'tr(AB)=tr(BA)' is valid even if B is not invertible. But I think this video couldn't explain this case. Is there any intuition which is helpful to understand this?
@mathemaniac
@mathemaniac 4 ай бұрын
As said in the video, we will have to wait until the end of the video series as I need to explain the notion of Lie brackets. In short, the explanation would be demonstrating tr(AB - BA) = 0. AB - BA is what is called the Lie bracket of A and B, but what it really means would be the topic of a future video.
@user-ug3gq8bk9t
@user-ug3gq8bk9t 4 ай бұрын
Thank you!❤
@drdca8263
@drdca8263 4 ай бұрын
Here’s an argument (I don’t know whether to say it “provides intuition”) : let c be a number which is not an eigenvalue of A. Then, A - c I is invertible (where I is the identity matrix) Then, tr((A - c I) B) = tr(B (A - c I)) But also, tr((A - c I) B) = tr(A B - c B) (by distributing and I B = B), and this is equal to tr(A B) - c tr(B) , because the trace is linear. meanwhile, on the other side, we have that tr(B (A - c I)) = tr(B A - c B) = tr(B A) - c tr(B) And so, we have that tr(AB) - c tr(B) = tr(BA) - c tr(B) So, adding a c tr(B) to both sides, we obtain tr(AB) = tr(BA), regardless of whether A is invertible. And, I suppose here is an intuition motivating the argument: a generic matrix is invertible, and so the claim holds for “almost all” matrices A, and like, on a dense subspace of the space of matrices. And, taking the trace is a continuous function. And, if an equation holds on a dense subspace, and the equation uses only continuous functions, then it holds throughout. Like, the equation tr(A B) = tr(B A) is equivalent to tr(A B) - tr(B A) = 0 and, tr(A B) - tr(B A) is a continuous function of A, and is zero whenever A is invertible, which is in a dense subspace. And, because the function is continuous, the preimage of {0} must be a closed set, and because it contains a set which is dense in the space of matrices, it must contain this entire space. Oops, that was another argument, perhaps not quite an intuition either. The commonality between the arguments being: it is true for invertible A, and most A are invertible, specifically, you can always change A by just a little bit in order to make it invertible, and this lets you get that it works for all A.
@__christopher__
@__christopher__ 4 ай бұрын
@@drdca8263 Actually the first argument also works over arbitrary fields, as long as A has not all values of the fields as eigenvalue (which could happen with finite fields), while the second argument obviously only works on non-discrete topological fields (there might be additional conditions on the topology for the argument to work; obviously whatever conditions are needed are fulfilled by the real numbers).
@drdca8263
@drdca8263 4 ай бұрын
@@__christopher__ Nice point, thanks!
@CraigFalls1
@CraigFalls1 4 ай бұрын
At 5:20 you say "after a very short time epsilon", but that is the first mention of time. There have been no "t" variables or anything up to that point, just matrices and vectors, with the matrices being visualized as vector fields. I feel like you need to explain how time plays into this first, before we can know what is meant by "after a short time".
@bumsupark3074
@bumsupark3074 4 ай бұрын
Great
@user-nb6el9te7m
@user-nb6el9te7m 4 ай бұрын
at 5:25 , you gave displacement after time epsilon formula without proof. And used it to say that the divergence is independent of position x. This is circular reasoning if you don’t explain why you define displacement to be independent of x in the first place.
@user-nb6el9te7m
@user-nb6el9te7m 4 ай бұрын
Are all these comments bots? There is an obvious hole in the explanation to anyone who studied maths at college level.
@mathemaniac
@mathemaniac 4 ай бұрын
The displacement after time epsilon d + epsilon Ad pretty much directly follows from the definition of velocities. The argument is that if initially the displacement was d (and this is regardless of the position), then the final displacement is also independent of the position. Hope this clarifies.
@chonchjohnch
@chonchjohnch 4 ай бұрын
The trace and determinant appear pretty naturally if you try solving a 2x2 matrix by hand :)
@alexboche1349
@alexboche1349 2 ай бұрын
I think he meant *adjugate* matrix not *adjoint* at 18:33 (tho the latter term used to mean that in old terminology). en.wikipedia.org/wiki/Adjugate_matrix en.wikipedia.org/wiki/Jacobi%27s_formula
@lovishnahar1807
@lovishnahar1807 4 ай бұрын
sir do u self study to know this much?
@MasterHigure
@MasterHigure 4 ай бұрын
Regarding trace being basis-independent, that just follows automatically from the fact that the vector field is basis-independent (the actual arrows drawn in the plane are basis-independent, not their coordinate components), and therefore also its divergence. There is no need to actually transform the plane and the arrows, because that's not what a change-of-basis is. It just looks that way if you insist on drawing both the old basis and the new basis using a horizontal first axis and a vertical second axis. But the whole idea of basis change is that the plane is unchanged and you just put new axes on it. The moment you start to turn and squish the arrows at 15:30, you have made everything a lot more complicated than it has to be. You proved that trace is the sum of eigenvalues, the proof of basis independence is basically identical, you can even use the exact same drawings. Heck, I would even consider proving them in the opposite order, and appeal to a diagonalization to show that trace is the sum of eigenvalues, at least for diagonalizable linear transformations (which, to be fair, was also assumed in your proof). That being said, how have I never heard that trace is divergence? This is awesome! One more thing I can now freely talk about in linear algebra without ever appealing to matrices. Which is how it should be done in my opinion.
@mathemaniac
@mathemaniac 4 ай бұрын
Yes, I briefly thought about doing it your way, but quickly abandoned this idea: the problem is that, at least for me, an active change is way easier to understand than a passive change (which you might call "change of basis"). This is also to establish the QAQ^(-1) order (active change) rather than P^(-1)AP (normally in change of basis, passive change) because it would be a lot more useful later on in Lie brackets.
@MasterHigure
@MasterHigure 4 ай бұрын
@@mathemaniac I personally disagree about which is easier to understand (but then again, I have never actually *taught* linear algebra, so I don't actually know). But since you had a thought behind it and also since you want to use it later, I can respect your choice. Tradeoffs must be made some times.
@Tom-qz8xw
@Tom-qz8xw 3 ай бұрын
I guess this also means that the trace of the Jacobian of a vector field is equal to its divergence.
@minJuunee
@minJuunee 4 ай бұрын
형 돌아와...
@AlexRodriguez-bt5jb
@AlexRodriguez-bt5jb 4 ай бұрын
You are incredible! I'm a math Phd student right now and I'm blown away by this series
@xephyr417
@xephyr417 2 ай бұрын
The geometric argument with the square doesnt quite work because you arent explaining all of the constraints youre assuming. You are assuming the 3rd and 4th corners of the parallelogram have a fixed distance and direction from the red anchor points.
@gregoryfenn1462
@gregoryfenn1462 4 ай бұрын
I got lost so early in this :( i used to understand this stuff
@goodlearnerbadstudent756
@goodlearnerbadstudent756 4 ай бұрын
Honestly...i am pretty lost with this video. That is more due to my weak foundation in math rather than your video i believe. 1 thing i don't understand is if trace is the sum of eigenvalues, trace is also the sum of diagonal. When exchange rows, it is said that eigenvalues are not affected, but given a square matrix, wouldn't that change the trace? since (e.g) row 7's 7i7j entry might not be the same as its 7i-2j entry if it exchanges with row 2, and may not be the same(?) as row 2's 2i-7j entry. the same/converse can be said of row 2's 2i-2j and 2i-7j entries? So, the trace should be different? then how does the eigenvalues remain the same. and it gets confusing(to me at least) because a matrix can represent a system of linear equations, where the rows are the corresponding equations, it can also be like(not sure the word for it) collection(?) of column vectors? and then it can also be a linear transformation. I am really lost.
@mathemaniac
@mathemaniac 4 ай бұрын
I am not really sure where you got exchanging rows does not affect eigenvalues, because it is not true. Pick a matrix, say A = (1,0; 0,1) the identity matrix, then exchanging rows gives B = (0,1;1,0). The eigenvalue of A is 1 (repeated), and eigenvalues of B are -1 and 1.
@a21imanadhikary83
@a21imanadhikary83 4 ай бұрын
As an Indian,I already learnt it in my high school and during Jee advanced Preparation😊
@misterlau5246
@misterlau5246 4 ай бұрын
Cool! So, we have here unitary generators and the specific matrices which comply certain characteristics so an special unitary group is mandatory 🤓🤓🤓🤓🤓🌶️🌶️🌶️🌶️🌶️🌶️😈😈😈😈
@larsdebrabander3613
@larsdebrabander3613 4 ай бұрын
first :)
@ucngominh3354
@ucngominh3354 4 ай бұрын
hi
@NavigatEric
@NavigatEric 3 ай бұрын
following math is not about numbers, but alphabets + flamboyant vocabulary = simple ideas made complex.
@LorenzoClemente
@LorenzoClemente 4 ай бұрын
there is nothing about Lie groups, algebras, or brackets...
@mathemaniac
@mathemaniac 4 ай бұрын
I did make a very brief mention of Lie brackets in the video - but it is a part of a video series on Lie algebra, and this video is a requisite for the things ahead because I want to use the intuition in the video (mainly thinking about matrices as vector fields or flows of vector fields) to talk about Lie algebras and brackets afterwards.
@BleachWizz
@BleachWizz 4 ай бұрын
it's so stupid. seriously. adding these diagonal entries should have no meaning at all. it's like... wat? I'm breaking.
@Checkthisontube
@Checkthisontube 4 ай бұрын
I apologize in advance if I offend you or make you feel insecure with the comment. I have a difficult time following your voice, you have some upspeak. Your voices raises at the end of many sentences and it is too irritating for me. It feels like someone is interrupting your monolog at the end of every sentence for me, which makes it really hard to focus on what you are saying. Your channel is quite successful so I was not sure should I give you feedback or would I just put unnecessary thoughts in your head. Nothing to complain about the video but I just can watch it on mute and read the subtitles because else it is way to distracting for me.
@shanathered5910
@shanathered5910 4 ай бұрын
hi
She’s Giving Birth in Class…?
00:21
Alan Chikin Chow
Рет қаралды 6 МЛН
ДЕНЬ РОЖДЕНИЯ БАБУШКИ #shorts
00:19
Паша Осадчий
Рет қаралды 5 МЛН
Indian sharing by Secret Vlog #shorts
00:13
Secret Vlog
Рет қаралды 55 МЛН
How AI Discovered a Faster Matrix Multiplication Algorithm
13:00
Quanta Magazine
Рет қаралды 1,3 МЛН
How (and why) to raise e to the power of a matrix | DE6
27:07
3Blue1Brown
Рет қаралды 2,7 МЛН
Feynman’s Integration Technique is Overpowered…
19:33
Jago Alexander
Рет қаралды 3,5 М.
The weirdest paradox in statistics (and machine learning)
21:44
Mathemaniac
Рет қаралды 1 МЛН
What does the second derivative actually do in math and physics?
15:19
Why can't you multiply vectors?
51:16
Freya Holmér
Рет қаралды 397 М.
The Most Useful Curve in Mathematics
23:43
Welch Labs
Рет қаралды 263 М.
Visualization of tensors  - part 1
11:41
udiprod
Рет қаралды 551 М.
She’s Giving Birth in Class…?
00:21
Alan Chikin Chow
Рет қаралды 6 МЛН