Eigenvectors / eigenrooms (Question in comment)
So I need a bit of help understanding how to find the eigenroom. So I understand that you take the "independent" variables and describe them as a vector, of the dependent ones. So like \[A=\left[\begin{matrix} 1 & -1 & 0 \\ 0 & 0 & 1 \\ 0 & 0 & 0\end{matrix}\right]\] Would have \[u=\left[\begin{matrix} 1\\ 1 \\ 0\end{matrix}\right]\] Because I set up the equations \[x_1-x_2=0 -> x_1=x_2\]\[x_2=x_2\] But what if I have \[A=\left[\begin{matrix} 1 & 1 & -1 \\ 0 & 0 & 0 \\ 0 & 0 & 0\end{matrix}\right]\] How would i find the two vectors? that would give me the equations: \[x_1+x_2-x_3=0\]\[x_2=x_2\]\[x_3=x_3\] Right? And how do I find the two vectors from this?
what is the actual matrix ?
Don't you mean the matrix \(\left[\begin{matrix} 1 & 1 & -1 \\ 0 & 0 & 0 \\ 0 & 0 & 0\end{matrix}\right]\) represents \(A-\lambda I\) ?
First I have, \[A=\left[\begin{matrix} 5 & -1 & 1 \\ -1 & 5 & 1 \\ 0 & 0 & 6\end{matrix}\right]\] Then I have found the 3 different eigenvalues, by taking the determinant of \[A=\left[\begin{matrix} 5-l & -1 & 1 \\ -1 & 5-l & 1 \\ 0 & 0 & 6-l\end{matrix}\right]\] And I got \[l=4,6,6\] Now when I insert \[l=6\] I get \[A=\left[\begin{matrix} -1 & -1 & 1 \\ -1 & -1 & 1 \\ 0 & 0 & 0\end{matrix}\right]\] And then from gaus elimination I get: \[A=\left[\begin{matrix} 1 & 1 & -1 \\ 0 & 0 & 0 \\ 0 & 0 & 0\end{matrix}\right]\]
l is supposed to be "Lambda"
Now it makes sense... gimme a minute to go through your work :)
Yea, sure, thanks :)
you may use `\lambda` to write \(\lambda \)
Oh yea okay thx ;)
for \(\lambda = 6\), we do get two independent eigenvectors
Yes, I am having trouble finding out how to find these, because I have \[x_1+x_2-x_3=0\]\[x_2=x_2\]\[x_3=x_3\], and I dont know how to describe those into 2 different vectors.
Easy, simply let \((x_2, x_3) = (1, 0)\) to get one eigen vector
Again let \((x_2, x_3) = (0, 1)\) to get another eigen vector
Whats an Eigenroom?? the set of eigen vectors that define different paralleipeds?
\[x_1+x_2-x_3=0\] plugin \((x_2,x_3)=(1,0)\) in above equation and solve \(x_1\)
These special values guarantee that you get "independent" eigenvectors
\[x_1+x_2-x_3=0\] If I insert \[(x_2,x_3)=(1,0)\] do I then get \[x_1=-x_2\]?
plugin \(x_2 = 1\) and \(x_3=0\)
\[x_1+1=0\]\[x_1=-1\]
\(x_1+1-0 = 0\) \(x_1=?\)
Yep, so one eigen vector is \((-1,1,0)\)
Let \((x_2, x_3) = (0, 1)\) to get another independent eigen vector
Hmm. According to the test results (It is exam practice, where we have the answers) The answers are \[Span[(1,0,1),(0,1,1)]\]
I have 4 answers, the 4th says: Neither one of the above.
And it says the answer is [3]
Yes, answer is indeed [3]
Our answer is also going to match with [3]
Can you finish it off ?
find the other eigenvector, don't wry about the options yet
Yea, but wolframalpha / what we did gives 3 vectors?
Ok
So next I would insert \[(x_2,x_3)=(0,1)\] \[x_1+0-1=0\]\[x_1=1\]
wolfram's 3 eigenvectors correspond to ALL the eigenvalues your test answer is referring to just the eigenvectors corresponding to the eigenvalue 6
Yes, it was just before we got \[(-1,1,0)\] Which I assume comes from when \[\lambda =4\]
Yes. So, what are the eigenvectors corresponding to the eigenvalue 6 ?
So we get:\[(1,0,1)\] and \[(-1,1,0)\]?
Yes. that looks good.
Now lets look at options
Looking at options, using these vectors, I would choose [4], none of the above.
Not so fast. Recall that if \(v_1\) and \(v_2\) are any two eigenvectors, then any linear combination \(c_1v_1 + c_2v_2\) is also an eigenvector
\[(1,0,1)\] and \[(-1,1,0)\] Add these and look at options again
That gives \[(0,1,1)\]
Yes, \((0, 1, 1)\) is also an eigenvector
So yea, thats how they got it.
There is nothing special about the eigenvectors that we have got earlier. Any two "independent" eigenvectors will do the job
So would this also work in another case, inserting \[(x_1,x_2)=(0,1)\] and \[(x_1,x_2)=(1,0)\]?
Which case ?
Well I dont know, just like in general
If am to find to eigenvectors
two*
You plugin special values for "free variables"
In our case, \(x_2\) and \(x_3\) are free variables because they don't have pivots
Yeap, and these "special values" are they always (0,1) and (1,0)
Or would they correspond to the values found in the other lines, like in our we have nothing. So we only have the equations \[x_2=x_2\] and \[x_3=x_3\]
You could use anything you like. But there is no guarantee that you get "indepdent" eigenvectors when you use random values/
ohhh, so you insert a "smart values" that eliminates one of the free variables?
If we use the special values (0, 1) and (1, 0) etc for free variables, we always get "independent" eigenvectors
yea, okay think I got it :)
Thank you so much :)
np :) one sec, wait
sure
try going thru this video when free http://ocw.mit.edu/courses/mathematics/18-06-linear-algebra-spring-2010/video-lectures/lecture-7-solving-ax-0-pivot-variables-special-solutions/ it expliains in detail how to cookup special solutions
Thanks :)
finding eigenvectors require you to solve the system : \[\left[\begin{matrix} 5-l & -1 & 1 \\ -1 & 5-l & 1 \\ 0 & 0 & 6-l\end{matrix}\right]\begin{pmatrix}x_1\\x_2\\x_3\end{pmatrix}=0\] that video explains everythign about nullspace Enjoy...
Join our real-time social learning platform and learn together with your friends!