Okay, if anyone is like me who is interested in Linear Algebra and not necessarily Differential Equations, but still wants to understand this lecture for the sake of completeness in this course, here is the list of all the Differential Equations theory you need for this lecture and where you can find content to learn it: 1. Separable differential equations - Khan Academy - this is for v(t)=exp(Lambda t) v(0) 2. 2nd order linear homogeneous equations - Khan Academy - you need this if you want to understand one calculation in the next item on this list (quick to learn anyways) 3. Systems of differential equations - MIT OCW 18.03, lectures 24 and 25 - basically does the first example in this lecture, but actually goes through the steps of explaining what is happening, highly recommended to watch both videos As someone who is unlikely to apply a whole lot of differential equations in the future, I still found this lecture useful to reinforce the properties of Eigenstuff from the previous lecture, also the very last bit of the lecture is very cool
to whomever think that it is complicated: you are missing essentials from 3blue1brown. His explanations will give you a solid basis to watch MIT lectures with joy and pleasure.
For those confused by this lecture, don't worry, you should be, it's no a failure of yours, it's a syllabus description mistake. I'm happy I see these videos as a review for things I ALREADY LEARNED years ago, because obviously if I were first exposed to this lecture with the prerequisites the course mentions, it wouldn't have been possible to understand this lecture. While the course specifies 18.02 as a prereq, and then immediately dismisses in its syllabus as not important, this lecture requires both 18.02 and 18.03, the latter is not mentioned at all in the syllabus. Universities usually parallelize these kind of courses with the parallel requisites such as 18.03, but OCW does not hint that at their site or here. While for me it was nice, I recognize that if you don't know diff. eqs this would not be possible for you to learn. The good news is that you can skip 1 or two lectures and you'd be right back on track, as this is just a nice implementation demonstration, but does not block Lin. Alg. learning.
The textbook actually says: "This section is mostly linear algebra, but to read it you need one fact from calculus: The derivative of e^(lambda)(t) is (lambda)e^(lambda)t" which I have learnt before in junior college. However I have not yet lernt differential equations at the university level and this lecture has been especially hard for me. The textbook also does not follow the same flow or use the same examples as the video. At the same time I'm hesistant to skip a whole section with the fear that I might actually already know the prerequisites but merely stumbled by how the lecture is conducted.
I think i did this in first year. I have seen this kind of solution of e^t terms for multiple ODE constraints given. Are these that only but just more elegantly solved using matrices. Because I remember we used give differentials some sort of variable then form a polynomial out of it solve them and then move to creating a solution. Is the matrix method doing the same but more elegantly just wanted to know..... It's bee years I have studied those.
Most unis have a whole course dedicated to diff eq. However, most engineering programs throw students into a statics course at the same time as they're taking diff eq. Having elementary exposure to diff eq prior to statics much quicker, and they will be quicker to move toward numerical methods for non-linear diff eq.
@@gordonlim2322 unfortunately the textbook doesn't follow the lectures very well at all... Best to watch the lectures first (both 18.03 and 18.06), take notes, then use that to study the textbook. It's taken me about a year to tackle all that on my own, but I was taking a full time course load at my university as well. Now I'm reviewing select lectures for the second time and it's helping solidify everything nicely
An awesome lecture indeed! Studied elementary differential equations at University 40 years ago. About two weeks of lectures to present the same materail. Linear algebra clarifies the process very well. The information relayed here conveys the cumulative effort of my lifetime study and concentration..as a hobby....yet some students will soak this up in minutes...work problems for a couple of weeks and have it down.....Awesome!!!!
Dr. Strang effortlessly incorporates little nuggets of wisdom in his lectures to keep the more advanced students interested, at the same time keeping most of the lecture simple enough so other students don't get lost. Bravo Sir!
(note for myself) 23:46) if the two eigenvalues are complex numbers, they'll be conjugates of each other(sum becomes a real number). 31:47, 33:26) Search what 'matrix exponential' is. It's *NOT* just a simple notation of element-wise exponential operation. * Also note that this lecture does *not* cover the entire rigorous mathematical proofs, as well as other lectures of Prof. Strang (At least this series of Linear Algebra lectures so far). * The solution using "matrix exponential" notation (31:55) can be understood if you see 44:30, given the Taylor series of the matrix is the _'definition'_ of the matrix exponential notation.
Ok guys don't complain about the difficulty of this section. You're gonna need some differential equation knowledge in order to fully understand this one, and the next one (FFT). E.g. solving a system of differential equation. Prof Strang did a fantastic job on that which I've never used a linear algebra point of view to understand system of differential equations and stabilities.
It helps to have Strang's textbook. But yes, you would need to have taken a course in Diff. equations. Knowing a little bit about control systems theory would help, too.
Amazing lecture! It will be better to learn this course with the textbook also by Strang. The videos provide big pictures and kind of whet your appetite for reading the book. The book gives detailed explanation and just satisfies your curiosity.
Although the audio is low in this video, DR. Strang relay the information on differential equations and exp(At) very well. Matrix exponential is very important in linear systems.
For all of you who are lost in this conference, don't worry, if you are just starting your career and have not seen calculus or differential equations, this will be almost impossible to understand, I think this conference seeks to show the potential of linear algebra in other fields, and I don't think it is necessary to fully understand it if you are just starting out, don't get frustrated by not understanding it, the level is much higher than the previous ones. By the way, if you develop an interest in dynamic systems (chaos theory), this is definitely a conference that you will remember in the future.
These lectures are awsome. Strang makes Linear Algebra seem so easy - completely oppisite of his "colleagues" at my university. I've stopped attending their lectures in favour of these. Thanks MIT (W. Gilbert Strang)!
if you are not familiar with laplace transform, especially coming from a signal and system background, you do get lost in here, it's not because of your poor understanding. But for EE, this is an essential and magic piece!
This was a great lecture as usual. My only complaint would be that the audio is only present on the right channel. I am sure many others would also appreciate if that could possibly be fixed. I would not mind doing that myself if there is a contact that I could send it to. Cheers
Yeah, I watch the lecture twice and it all gels for me. There's a beautiful thing lurking in here, I think, which is the situation where only one eigenvalue is relevant.
For that first example, I don’t think assuming the solution was the Iinear combination of exponentials was the right way to go. Of course, it’s correct, but it’s not very instructive. I solved this example problem in my own, and this is what I did: Broke the system of odes into a system of difference equations (finite element approximations of the odes). Then I solved for the vector one node away in terms of the first node by reapplying something like u_(n+1)=A*u_n over and over. The computational hack here is that A being middle diagonalizable matrix doesn’t actually need to be multiplied many, many times. Once you get your answer, take the limit of your solution such that the difference equations converge to the original odes, and boom, you have your solution. It blew my mind a lil bit when I got the right answer lol 😅.
@hypnoticpoisons Use Euler’s formula with x =6t: e^ix = cos x + i sin x |e^ix| = |cos x + i sin x| = sqrt((cos x + i sin x)(cos x - i sin x)) = sqrt(cos^2 x + sin^2 x) = sqrt(1) = 1
@sbhdgr8 No. He's trying to write the second order differential equation as a system of first order linear equations, or basically you are just substituting y'=some variable, basically just manipulating the expressions so the 2nd order equation is "transformed" into equations of first order. The matirx of coeffiecients represents the coefficients of the "transformed" resulting system of 1ST order linear equations. It is the same idea as substitution in quadratic equations like ax^4+bx^2+c, and
Seeing this lecture.....i say it's enough in linear algebra.....i am done with it.....most of the points are unclear in this lecture.....bt before this,every lecture was extraordinary...
@@Sandyy10101 18.06 , has 18.01 (Single variable calculus ) , 18.02 ( Multivariate Calculus) and 18.03 (Differential Equations) as prerequisites. But my guess is , even if you dont udnerstand the differential equations stuff , it's still ok , you can just skip any video with things like "fourier transforms , diff eqs " and so on.You should definitely watch the positive defineteness lectures
Can anyone explain 32:18 why u(t)= s e^(lambda* t) s(inverse) * u(0) = e^A*t *u(0)? From 9:33, for example u(0)= c1* 1*x1, c2*1*x2 (t=0, so e^lambda 1 or 2 * t, are 0)
My understanding: earlier part of the lecture showed u(t) is some form of exponential function that u(t)=e^(At), and u(0) is just some initial condition to position u(t)'s initial position, so u(t)=e^(At)*u(0); Professor set u=Sv, where S's basis are eigenvectors of A. We have (u)'=(Sv)'=Au=ASv, S here is a constant matrix, so take it out, (u)'=(Sv)'=S(v)'=ASv, to write this in another way as S(dv/dt)=ASv, so we have dv/dt=S^(inverse)ASv. We know that S^(inverse)AS is the diagonal matrix with lambdas, so dv/dt=Lambda*v, and this shows that v is some form of exponential function as well, so we can write v(t)=e^(Lambda*t)v(0); given u=Sv, so substitute u(0)=Sv(0) here (which leads to v(0)=S^(inverse)*u(0)), so we have u(t)=Sv(t)=S*e^(Lambda*t)*S^(inverse)*u(0).
@26:00 shouldn't it be det >=0 to account for the solutions with steady state? And I suppose yes, this lecture is clear if you have been previously exposed to at least basic differential equations. But hey! OCW offers that course for free too! ;)
You are confusing stability with steady state. A system is said to be stable when equilibrium is achieved i.e the system goes to 0. For that to happen, the strict condition is that all the eigenvalues MUST be negative. If in case one or more of the eigenvalues become 0, the system is MAY OR MAY NOT be stable depending upon the multiplicity of the eigenvalues.
in 30.41 he writes "set u=Sv" he says that S is the eigenvector Matrix but he does not say nothing for WHAT IS THIS "v small" ,and why did he do that trick..it is disheartening
+Andrea Gjoka I understand how that could frustrate you. Here's what he's doing. He's decomposing "u" into a linear combination of the eigenvectors, which are arranged in the matrix "S". The elements of the vector "v" are the weights in that linear combination. He performed that "trick" so he could get rid of "A" (in which the variables of the system are coupled together), and replace it with capital Lambda, which is a diagonal matrix (uncoupled variables). The fact that capital Lambda is diagonal is important because the exponential of a diagonal matrix turns out to be trivial. So he's setting up an identity that relates the exponential of a non-diagonal matrix to the exponential of a diagonal one (multiplied as needed by the eigenvector matrix and its inverse). This is the foundation for the series-based definition of matrix exponentials he presents a little later in the lecture. Hope this helps.
He is merely trying to express u as a linear combination of eigen vectors. Vector v is not composed of constants as S contains constants. All components of V are time (or t) dependent. Hope it helped.
In case anyone had the same question, here's how I figured it out: S spans the space (it is a set of basis vectors) so you can transform S to any vector in the space with the right combination of its column vectors. It is similar to what was done in the last lecture transforming u_0 to Sc
I just finished watching 3Blue1Brown’s first video on matrix exponentials, it’s so well made. Even though I feel like I haven’t fully understood every aspect of this lecture, I do believe that with the help of 3Blue1Brown, this lecture might turn into intuition for me someday.
If anyone manages to see this, should I have done differential equations by now; or not? Also, before differential equations, should I do calculus 3/multivariate calculus; or differential equations first? Thanks to anyone who responds and I hope you have had a lovely day so far!
To be clear that I'm not missing something in the lecture, he assumes exp(lambda*t) is a solution to the matrix differential equation BEFORE he has explained what a matrix exponential is or shown how it might be derived, correct? I'm asking because I want to make sure I'm not missing some section of the video. I went back and watched parts of it multiple times and I think he is skipping around at some places in a way that makes it hard to follow, even if you're familiar with differential equations and the material up to this point. I generally like his broad and conceptual approach but in this video I believe there were too many jumps from one line of thought to another. Nonetheless, once I got past some of the excessive handwaviness, the overall approach to decoupling differential equations is excellent and I learned a lot.
Prof. Strang is amazing but you should skip this lecture if you are NOT familiar with differential equations. It is too fast paced, and avoids multiple non-distinct eigenvalues. I didn't find the book section very helpful as well. The good thing is skipping this "applications" lecture doesn't affect the series. I would recommend Edwards-Penney differential equations Chapter 5 if you really want to understand what is going on and where these equations are coming from (i.e. the link between eigenvectors, exponential matrices and different equations). You can look at ocw 18.03 section 4 as well, which will explain this at a reasonable pace. or you can just leave thinking aside and put the following things into memory. A system of differential equations has a solution in the eigenvectors and eigenvalues of A. Exponential of a matrix is a taylor series with scalar replaced by a matrix and follows the same derivative. For stability, real part of eigenvalues has to be negative.
It's the initial condition of the function v(t). Since at time=0 the e^λt equals 1, then it's just v(0). It's practically the C(which represents the initial condition) as it is referred in the exponential models and differential equations.
I have a question if someone could answer it; right at the end of the lecture where a second order diff. eqn is converted into a 2x2 matrix called A; while defining the matrix A, shouldn't element A22 be= -k/b since on rearranging the given eqn and solving for y' we get y'= - (1/b) *y"/b - (k/b)* y; so since the coefficient of y is -k/b ; shouldn't this coefficient correspond to element A22 of the 2x2 Matrix A?
No, because as Professor explained in the last seconds of the lecture in order to convert a 2nd order system to a first order we have to introduce a redundant equation hence the elements A21 and A22 must be 1 and 0 to complete the matrix.
At 49:30 can someone please explain me how in matrix -b and -k came, I tried hard but couldn't understood, or am I missing some sort of pre requisite here?
he made y^|| the subject of the formula which took b and k to the other end of the equation and then used the first row of the matrix to get the equation and the second row was used to get the second equation which is trivial, hope this helps
In the end why does he say that matrix converts 5th order differential equation to 1st order differential equation! there are higher orders in the L.H.S...
on the book there is a part I did not get at all-namely starting"to display a circle on screen,replace y"=-y' with finite finite difference equation" from end of 315 page till 317, anyone who has understood that part?
start with dv/dt = Lambda v. This says the derivative of a (vector) variable is a constant times the variable. From differential equations, you know the solution is an exponential function (multiplied by the initial condition). This is just expressing it as a vector (i.e., solving several single-variable cases simultaneously). Thus, v(t)=e^Lambda t v(0). But v was a vector that he got by taking u, and finding what combination of columns of S would be necessary to produce u. v is the vector of those coefficients. Since there are n independent columns of S, it forms a basis, so any vector u can be written as a combination of those columns. Thus, u=Sv. So the next step is u(t) = S v(t) = S e^Lambda v(0). But since u(0)=Sv(0), v(0) = S^-1 u(0). Substitute that, and you're done: u(t) = S e^Lambda S^-1 u(0).
The notes for this lecture on ocw.mit.edu mixed up u(0) and v(0) in a bunch of places... (like u(t) = S e^Lambda S^-1 V(0)). Spent hours trying to figure what's going on before I saw this comment XD. Thanks!
You need to know first order linear differential equations. You might learn this somewhere towards the end of a single-variable integral calculus course, and of course you learn it in a first course on differential equations.
From the fact that dv(t)/dt = Λ * v(t), try to derive formula for v(t). You'll get v(t) = c1 * e^(λ1t) * [1; 0] + c2 * e^(λ2t) * [0; 1]. When you set t = 0, you'll get c1 = v1(0) and c2 = v2(0). So it becomes the "natural notation".
I'm so gonna forget all the homeworks I've done in this course (because it makes me learn how to compute rather than actual knowledge, I'll still remember the "knowledge" in the lecture though) Confusing part: 0:00 what does that linear problem really mean? (Why is the solution only u(t) when there're 2 variables u1 and u2? Shouldn't it be u1(t) and u2(t) Maybe that some calculus that I don't know yetf) 9:55 (Au popping off like a wild Pokemon) 18:48 (What is that value, sine and cosine thingy and why? Maybe that's some calculus problem that I should just agree that it is for now) I heard about Taylor series Is (1-x)^-1 = 1 + x + x^2 +... true with every x? Oh, here's the answer : mathforum.org/library/drmath/view/61847.html But then how can we judge if At < I, monkaHmm
I think ill leave this and the next lecture for some other day, coz i dont have enough solid background with diffs to understand these 2 lectures. those who has completed course are this two lectures a must for understanding the rest of the course? thanks
Not quite. But I was doing this through OCW so it came in quiz 2. I don't think you'll need it later until you want to solve high order differential equations. But this is as important as the applications of Linear algebra get so, you should study it nevertheless.
Okay, if anyone is like me who is interested in Linear Algebra and not necessarily Differential Equations, but still wants to understand this lecture for the sake of completeness in this course, here is the list of all the Differential Equations theory you need for this lecture and where you can find content to learn it:
1. Separable differential equations - Khan Academy - this is for v(t)=exp(Lambda t) v(0)
2. 2nd order linear homogeneous equations - Khan Academy - you need this if you want to understand one calculation in the next item on this list (quick to learn anyways)
3. Systems of differential equations - MIT OCW 18.03, lectures 24 and 25 - basically does the first example in this lecture, but actually goes through the steps of explaining what is happening, highly recommended to watch both videos
As someone who is unlikely to apply a whole lot of differential equations in the future, I still found this lecture useful to reinforce the properties of Eigenstuff from the previous lecture, also the very last bit of the lecture is very cool
the comment I needed, thank you! xD
gracias
Thank you, much appreciated
Thanks a lot buddy! This was really helpful.
Thanks for the really helpful resources!
This lecture and the one before are some of the hardest in the whole course. Stay the course, everyone.
@Sirin Kalapatuksice probably people in MIT learns ODE's at the same time of this lecture in their analysis course
to whomever think that it is complicated: you are missing essentials from 3blue1brown. His explanations will give you a solid basis to watch MIT lectures with joy and pleasure.
@@snowy0110 Thank you so much, I'll check that out latter. I've done calculus years ago but forgot all of it, and I'm struggling with the course.
But both seem to me interesting 👍
We are not weak in math; we just didn't have Gilbert Strang as a teacher. 😊
For those confused by this lecture, don't worry, you should be, it's no a failure of yours, it's a syllabus description mistake.
I'm happy I see these videos as a review for things I ALREADY LEARNED years ago, because obviously if I were first exposed to this lecture with the prerequisites the course mentions, it wouldn't have been possible to understand this lecture.
While the course specifies 18.02 as a prereq, and then immediately dismisses in its syllabus as not important, this lecture requires both 18.02 and 18.03, the latter is not mentioned at all in the syllabus. Universities usually parallelize these kind of courses with the parallel requisites such as 18.03, but OCW does not hint that at their site or here.
While for me it was nice, I recognize that if you don't know diff. eqs this would not be possible for you to learn. The good news is that you can skip 1 or two lectures and you'd be right back on track, as this is just a nice implementation demonstration, but does not block Lin. Alg. learning.
Thanks for stating that man.
Although I figured that out, reading this reassured me.
The textbook actually says: "This section is mostly linear algebra, but to read it you need one fact from calculus: The derivative of e^(lambda)(t) is (lambda)e^(lambda)t" which I have learnt before in junior college. However I have not yet lernt differential equations at the university level and this lecture has been especially hard for me. The textbook also does not follow the same flow or use the same examples as the video. At the same time I'm hesistant to skip a whole section with the fear that I might actually already know the prerequisites but merely stumbled by how the lecture is conducted.
I think i did this in first year. I have seen this kind of solution of e^t terms for multiple ODE constraints given. Are these that only but just more elegantly solved using matrices. Because I remember we used give differentials some sort of variable then form a polynomial out of it solve them and then move to creating a solution. Is the matrix method doing the same but more elegantly just wanted to know..... It's bee years I have studied those.
Most unis have a whole course dedicated to diff eq. However, most engineering programs throw students into a statics course at the same time as they're taking diff eq. Having elementary exposure to diff eq prior to statics much quicker, and they will be quicker to move toward numerical methods for non-linear diff eq.
@@gordonlim2322 unfortunately the textbook doesn't follow the lectures very well at all... Best to watch the lectures first (both 18.03 and 18.06), take notes, then use that to study the textbook. It's taken me about a year to tackle all that on my own, but I was taking a full time course load at my university as well. Now I'm reviewing select lectures for the second time and it's helping solidify everything nicely
This lecture was all about flexing the computational power of eigenstuffs. I'm definitely impressed!
The greatest instructor I have ever met!! Thank you, Dr. Strang!
An awesome lecture indeed! Studied elementary differential equations at University 40 years ago. About two weeks of lectures to present the same materail. Linear algebra clarifies the process very well. The information relayed here conveys the cumulative effort of my lifetime study and concentration..as a hobby....yet some students will soak this up in minutes...work problems for a couple of weeks and have it down.....Awesome!!!!
I watched this lecture maybe three times. and still screaming. thank you Prof. Gilbert Strang
he just keeps pushing the boundaries!!
it's not that i don't understand but i never thought that u can use such tools this way!
Dr. Strang effortlessly incorporates little nuggets of wisdom in his lectures to keep the more advanced students interested, at the same time keeping most of the lecture simple enough so other students don't get lost. Bravo Sir!
this one lecture covers most of my math and control engineering module. guys a genius
+dunsacc Same here. You really know someone got it if he can explain something others were referring to as hard in actually simple words.
The two dimensional sound channel matrix has only one independent vector. Bad one.
You can use projection Matrix Mono Sound to solve the problem
@@puneetgarg4731 You seem like a smart man. Been following your comments around for a while here.
@@danishji2172 thankyou
(note for myself) 23:46) if the two eigenvalues are complex numbers, they'll be conjugates of each other(sum becomes a real number).
31:47, 33:26) Search what 'matrix exponential' is. It's *NOT* just a simple notation of element-wise exponential operation.
* Also note that this lecture does *not* cover the entire rigorous mathematical proofs, as well as other lectures of Prof. Strang (At least this series of Linear Algebra lectures so far).
* The solution using "matrix exponential" notation (31:55) can be understood if you see 44:30, given the Taylor series of the matrix is the _'definition'_ of the matrix exponential notation.
Ok guys don't complain about the difficulty of this section. You're gonna need some differential equation knowledge in order to fully understand this one, and the next one (FFT). E.g. solving a system of differential equation. Prof Strang did a fantastic job on that which I've never used a linear algebra point of view to understand system of differential equations and stabilities.
It helps to have Strang's textbook. But yes, you would need to have taken a course in Diff. equations. Knowing a little bit about control systems theory would help, too.
Amazing lecture! It will be better to learn this course with the textbook also by Strang. The videos provide big pictures and kind of whet your appetite for reading the book. The book gives detailed explanation and just satisfies your curiosity.
This lecture is amazing! Its amazing how well he communicates these concepts
My right ear learned a lot
turn on mono sound in ease of access if you use windows
@@AmolDeometaphys thanks for your advice :)
You can also do the same on Android
hahahahahahha awesome comment
Thanks for your comment, I can give my left ear a chance to study!😎
For anyone looking for some more visual intuition, 3blue1Brown has an amazing video on raising matrices to powers in his DE series.
a note for those considering [-1 1]^T as the second eigenvector, we get c1 = 1/3 and c2 = -1/3.
Awesome lecture. Neat and consistent. The stability part is much better than what i have learned in linear system class.
Although the audio is low in this video, DR. Strang relay the information on differential equations and exp(At) very well. Matrix exponential is very important in linear systems.
i think this lecture is probably the most difficult one in this series, but as usual well explained!
For all of you who are lost in this conference, don't worry, if you are just starting your career and have not seen calculus or differential equations, this will be almost impossible to understand, I think this conference seeks to show the potential of linear algebra in other fields, and I don't think it is necessary to fully understand it if you are just starting out, don't get frustrated by not understanding it, the level is much higher than the previous ones.
By the way, if you develop an interest in dynamic systems (chaos theory), this is definitely a conference that you will remember in the future.
I think it's better to first watch the amazing TA Linan Chen video then come back and watch this lecture
These lectures are awsome. Strang makes Linear Algebra seem so easy - completely oppisite of his "colleagues" at my university. I've stopped attending their lectures in favour of these. Thanks MIT (W. Gilbert Strang)!
high five broo (same situation even after 7 years lol)
Seems like he had an extra cup of coffee this morning
probably u were watching it on 1.5x , too LOL
@@nimaabbasi1927 i m on 2x speed. he is shivering lol.
haha, that is true
if you are not familiar with laplace transform, especially coming from a signal and system background, you do get lost in here, it's not because of your poor understanding. But for EE, this is an essential and magic piece!
Professor Strang is the Columbo of Linear Algebra
i have never been so lost in my entire life.....
Keep learning, this is worth it
This was a great lecture as usual. My only complaint would be that the audio is only present on the right channel. I am sure many others would also appreciate if that could possibly be fixed. I would not mind doing that myself if there is a contact that I could send it to. Cheers
Took me almost 3 hours to get close to "fully understand" the lecture.
Same.
Yeah, I watch the lecture twice and it all gels for me. There's a beautiful thing lurking in here, I think, which is the situation where only one eigenvalue is relevant.
me 2, but at third try
For that first example, I don’t think assuming the solution was the Iinear combination of exponentials was the right way to go. Of course, it’s correct, but it’s not very instructive. I solved this example problem in my own, and this is what I did:
Broke the system of odes into a system of difference equations (finite element approximations of the odes). Then I solved for the vector one node away in terms of the first node by reapplying something like u_(n+1)=A*u_n over and over. The computational hack here is that A being middle diagonalizable matrix doesn’t actually need to be multiplied many, many times. Once you get your answer, take the limit of your solution such that the difference equations converge to the original odes, and boom, you have your solution. It blew my mind a lil bit when I got the right answer lol 😅.
Wow, for a video 14 years ago is better than that of my math teacher today.
these are my favorite lessons
@hypnoticpoisons Use Euler’s formula with x =6t:
e^ix = cos x + i sin x
|e^ix| = |cos x + i sin x|
= sqrt((cos x + i sin x)(cos x - i sin x))
= sqrt(cos^2 x + sin^2 x)
= sqrt(1)
= 1
Thank you, sir. It was very valuable for those of us who isn't yet too comfortable with imaginary stuff yet.
For those who have struggled with the where does -k and -b came from: See that y^|| = dy^| / dt = -by^| - ky
30:30 i guess this should be at the beginning of the lecture
but it was a very nice lecture thank you for ur efforts
life saver for my final
Ahhh!!! Mind blown near the end!!!
@sbhdgr8 No. He's trying to write the second order differential equation as a system of first order linear equations, or basically you are just substituting y'=some variable, basically just manipulating the expressions so the 2nd order equation is "transformed" into equations of first order. The matirx of coeffiecients represents the coefficients of the "transformed" resulting system of 1ST order linear equations. It is the same idea as substitution in quadratic equations like ax^4+bx^2+c, and
If I need to understand this lecture to proceed with the rest of the course, I’d better just give up maths for life...
This lecture is amazing...
Seeing this lecture.....i say it's enough in linear algebra.....i am done with it.....most of the points are unclear in this lecture.....bt before this,every lecture was extraordinary...
yeah!! Everything is clear until this lecture... Maybe it's because he assumes we have taken some courses of 18.01-05 before watching this video
@@Sandyy10101 18.06 , has 18.01 (Single variable calculus ) , 18.02 ( Multivariate Calculus) and 18.03 (Differential Equations) as prerequisites. But my guess is , even if you dont udnerstand the differential equations stuff , it's still ok , you can just skip any video with things like "fourier transforms , diff eqs " and so on.You should definitely watch the positive defineteness lectures
Can anyone explain 32:18 why u(t)= s e^(lambda* t) s(inverse) * u(0) = e^A*t *u(0)?
From 9:33, for example u(0)= c1* 1*x1, c2*1*x2 (t=0, so e^lambda 1 or 2 * t, are 0)
My understanding: earlier part of the lecture showed u(t) is some form of exponential function that u(t)=e^(At), and u(0) is just some initial condition to position u(t)'s initial position, so u(t)=e^(At)*u(0); Professor set u=Sv, where S's basis are eigenvectors of A. We have (u)'=(Sv)'=Au=ASv, S here is a constant matrix, so take it out, (u)'=(Sv)'=S(v)'=ASv, to write this in another way as S(dv/dt)=ASv, so we have dv/dt=S^(inverse)ASv. We know that S^(inverse)AS is the diagonal matrix with lambdas, so dv/dt=Lambda*v, and this shows that v is some form of exponential function as well, so we can write v(t)=e^(Lambda*t)v(0); given u=Sv, so substitute u(0)=Sv(0) here (which leads to v(0)=S^(inverse)*u(0)), so we have u(t)=Sv(t)=S*e^(Lambda*t)*S^(inverse)*u(0).
@@vivianhuang5647 appreciate your help.
29:50 Why he can set u= Sv? Should s have n independent eigenvectors for 2 by 2 A ?
I watch on 2x and understood everything and now i am feeling like i need to watch again in 0.25x
The comments are always much consolation to see...
@26:00 shouldn't it be det >=0 to account for the solutions with steady state? And I suppose yes, this lecture is clear if you have been previously exposed to at least basic differential equations. But hey! OCW offers that course for free too! ;)
You are confusing stability with steady state. A system is said to be stable when equilibrium is achieved i.e the system goes to 0. For that to happen, the strict condition is that all the eigenvalues MUST be negative. If in case one or more of the eigenvalues become 0, the system is MAY OR MAY NOT be stable depending upon the multiplicity of the eigenvalues.
even the beauty gilbert strang cannot make me love that freaking algebra
if the formula for (I - At)-1 is correct at 38:49, if At = 2I, rhs must blow up but clearly (I - 2I)-1 does exists and is -I. Am I missing something?
As a non English native speaker university student, the first minute of the video is like the professor speaking alien language to me :|
S contains a bunch of numbers independent of t (the eigenvectors), so d/dt and S would commute.
in 30.41 he writes "set u=Sv" he says that S is the eigenvector Matrix but he does not say nothing for WHAT IS THIS "v small" ,and why did he do that trick..it is disheartening
+Andrea Gjoka I understand how that could frustrate you. Here's what he's doing. He's decomposing "u" into a linear combination of the eigenvectors, which are arranged in the matrix "S". The elements of the vector "v" are the weights in that linear combination. He performed that "trick" so he could get rid of "A" (in which the variables of the system are coupled together), and replace it with capital Lambda, which is a diagonal matrix (uncoupled variables). The fact that capital Lambda is diagonal is important because the exponential of a diagonal matrix turns out to be trivial. So he's setting up an identity that relates the exponential of a non-diagonal matrix to the exponential of a diagonal one (multiplied as needed by the eigenvector matrix and its inverse). This is the foundation for the series-based definition of matrix exponentials he presents a little later in the lecture. Hope this helps.
+L.G. Lucid Thank you for the explanation.I have to see it again now.
you seem to focus better than I do ..
He is merely trying to express u as a linear combination of eigen vectors. Vector v is not composed of constants as S contains constants. All components of V are time (or t) dependent. Hope it helped.
In case anyone had the same question, here's how I figured it out:
S spans the space (it is a set of basis vectors) so you can transform S to any vector in the space with the right combination of its column vectors. It is similar to what was done in the last lecture transforming u_0 to Sc
@hypnoticpoisons Because e^(6it) = cos(6t)+i sin(6t), which always has a modulus of 1.
can i call this statement "kadkaesque explanation of |z|=z(~z)"?
Great teacher.
This is the best lecture series of Linear algebra ❤️
Great! What a gifted Professor, thank you Dr. Strang!
This is genius! Now I want to go to MIT, but not enough money... guess I will stay at Dalhousie
I just finished watching 3Blue1Brown’s first video on matrix exponentials, it’s so well made.
Even though I feel like I haven’t fully understood every aspect of this lecture, I do believe that with the help of 3Blue1Brown, this lecture might turn into intuition for me someday.
For a moment I thought my left ear had given up.
Best one Yet ! !! !
If anyone manages to see this, should I have done differential equations by now; or not? Also, before differential equations, should I do calculus 3/multivariate calculus; or differential equations first?
Thanks to anyone who responds and I hope you have had a lovely day so far!
very satisfactory explanation!
ERROR: stability regions are meesed up. Unit circle is for degrees, and left half-plane is for exponentials.
To be clear that I'm not missing something in the lecture, he assumes exp(lambda*t) is a solution to the matrix differential equation BEFORE he has explained what a matrix exponential is or shown how it might be derived, correct?
I'm asking because I want to make sure I'm not missing some section of the video.
I went back and watched parts of it multiple times and I think he is skipping around at some places in a way that makes it hard to follow, even if you're familiar with differential equations and the material up to this point. I generally like his broad and conceptual approach but in this video I believe there were too many jumps from one line of thought to another.
Nonetheless, once I got past some of the excessive handwaviness, the overall approach to decoupling differential equations is excellent and I learned a lot.
Yes
Beautiful
@hypnoticpoisons I think x is limited to numbers between 0 and 1
@sbhdgr8 you substitute z=x^2 so the higher power "quadratic" equation can be written as a actual quadratic equation az^2+bz+c.
When you dont hear anything and then realise you have to put in the other earphone.
Prof. Strang is amazing but you should skip this lecture if you are NOT familiar with differential equations. It is too fast paced, and avoids multiple non-distinct eigenvalues. I didn't find the book section very helpful as well.
The good thing is skipping this "applications" lecture doesn't affect the series. I would recommend Edwards-Penney differential equations Chapter 5 if you really want to understand what is going on and where these equations are coming from (i.e. the link between eigenvectors, exponential matrices and different equations).
You can look at ocw 18.03 section 4 as well, which will explain this at a reasonable pace.
or you can just leave thinking aside and put the following things into memory.
A system of differential equations has a solution in the eigenvectors and eigenvalues of A. Exponential of a matrix is a taylor series with scalar replaced by a matrix and follows the same derivative. For stability, real part of eigenvalues has to be negative.
Gilbert "The Wizard" Strang!
Can anyone explain, at 31:39, how do we know that v = e^lambda t * v(0)? Where did the v(0) come from?
It's the initial condition of the function v(t). Since at time=0 the e^λt equals 1, then it's just v(0). It's practically the C(which represents the initial condition) as it is referred in the exponential models and differential equations.
Superb!
The corresponding lecture form 18.03 by MIT OCW that bridges this module is Lecture 25.
ua-cam.com/video/heBvViSi9xQ/v-deo.html
e^At does not converge as t tends to infinity....why does he say that?
Great cource!Thank you a lot!
36:45 why ?
I have a question if someone could answer it;
right at the end of the lecture where a second order diff. eqn is converted into a 2x2 matrix called A;
while defining the matrix A, shouldn't element A22 be= -k/b since on rearranging the given eqn and solving for y' we get y'= - (1/b) *y"/b - (k/b)* y; so since the coefficient of y is -k/b ; shouldn't this coefficient correspond to element A22 of the 2x2 Matrix A?
No, because as Professor explained in the last seconds of the lecture in order to convert a 2nd order system to a first order we have to introduce a redundant equation hence the elements A21 and A22 must be 1 and 0 to complete the matrix.
At 49:30 can someone please explain me how in matrix -b and -k came, I tried hard but couldn't understood, or am I missing some sort of pre requisite here?
Prerequisites: Multivariable Calculus (18.02). See ocw.mit.edu/18-06S05 for more info. Best wishes on your studies!
he made y^|| the subject of the formula which took b and k to the other end of the equation and then used the first row of the matrix to get the equation and the second row was used to get the second equation which is trivial, hope this helps
In the end why does he say that matrix converts 5th order differential equation to 1st order differential equation! there are higher orders in the L.H.S...
Thank you.
Copyrighted in 2000, made in 2005...
Has MIT discovered time travel??
uploaded in 2009 also
+Andrea Gjoka The videos were recorded in 1999, published on MIT OpenCourseWare in 2005, and re-published at a higher resolution in 2009.
very satisfactory explanation! i wonder though if the curriculum has changed since 1999.
@@mitocw Oh my GOD, these lectures are so old... I didn't realize it. They are still the best.
@@mitocw Please fix this video's audio channel.
46:30, 46:45, 49:45
on the book there is a part I did not get at all-namely starting"to display a circle on screen,replace y"=-y' with finite finite difference equation" from end of 315 page till 317,
anyone who has understood that part?
why u(0)= [1,0].T, is it a dummy vector that help calcuate?
In 32.08, can anyone explain how does he derive v(t) and u(t)?
start with dv/dt = Lambda v. This says the derivative of a (vector) variable is a constant times the variable. From differential equations, you know the solution is an exponential function (multiplied by the initial condition). This is just expressing it as a vector (i.e., solving several single-variable cases simultaneously). Thus, v(t)=e^Lambda t v(0). But v was a vector that he got by taking u, and finding what combination of columns of S would be necessary to produce u. v is the vector of those coefficients. Since there are n independent columns of S, it forms a basis, so any vector u can be written as a combination of those columns. Thus, u=Sv. So the next step is u(t) = S v(t) = S e^Lambda v(0). But since u(0)=Sv(0), v(0) = S^-1 u(0). Substitute that, and you're done: u(t) = S e^Lambda S^-1 u(0).
The notes for this lecture on ocw.mit.edu mixed up u(0) and v(0) in a bunch of places... (like u(t) = S e^Lambda S^-1 V(0)). Spent hours trying to figure what's going on before I saw this comment XD. Thanks!
god~~ thank, man. you save my day.👍
thanks, you save my month lol
@@davidlovell729 lifesaver
Buy why is trace a sum of eige values and det product of it
Do I need to know calculus for this ;(. I will learn that after and watch this again.
Yes, you do.
You need to know first order linear differential equations. You might learn this somewhere towards the end of a single-variable integral calculus course, and of course you learn it in a first course on differential equations.
My puny mind can't handle this matrix exponential thing.
31:53 what is this "natural notation"?
I wish he derived that form from u(t) = c1e^(λt)x1
I'm also struggling with that part. Please explain...
From the fact that dv(t)/dt = Λ * v(t), try to derive formula for v(t). You'll get v(t) = c1 * e^(λ1t) * [1; 0] + c2 * e^(λ2t) * [0; 1]. When you set t = 0, you'll get c1 = v1(0) and c2 = v2(0). So it becomes the "natural notation".
48:14 you change it by switching from Lagrangian mechanics to Hamiltonian :D
Is zero always an Eigen value for 2 x 2 matrix?
Nice lecture. Easy to see why solutions are exponential in t. But, I did not get why solution is exponential in lambda*t.
He showed at 9:58
The first 5 minutes is like professor speaking alien language to me :'(
Dear God,
Thank you for Gilbert Strang
34:13 is a beauty
I'm so gonna forget all the homeworks I've done in this course (because it makes me learn how to compute rather than actual knowledge, I'll still remember the "knowledge" in the lecture though)
Confusing part:
0:00 what does that linear problem really mean? (Why is the solution only u(t) when there're 2 variables u1 and u2? Shouldn't it be u1(t) and u2(t) Maybe that some calculus that I don't know yetf)
9:55 (Au popping off like a wild Pokemon)
18:48 (What is that value, sine and cosine thingy and why? Maybe that's some calculus problem that I should just agree that it is for now)
I heard about Taylor series
Is (1-x)^-1 = 1 + x + x^2 +... true with every x?
Oh, here's the answer :
mathforum.org/library/drmath/view/61847.html
But then how can we judge if At < I, monkaHmm
I Really Like The Video From Your Differential Equations and exp(At)
TV 2 FILM Is It compulsory to put that same comment on every single lecture? 😅
I think ill leave this and the next lecture for some other day, coz i dont have enough solid background with diffs to understand these 2 lectures. those who has completed course are this two lectures a must for understanding the rest of the course? thanks
Not quite. But I was doing this through OCW so it came in quiz 2. I don't think you'll need it later until you want to solve high order differential equations. But this is as important as the applications of Linear algebra get so, you should study it nevertheless.
@0:00 "And, lift-off." LOL