Using diagonalization, prove the matrix equals it's square

In summary: What do you mean, 0 and 1 are independent?? 0 and 1 are numbers, not vectors. Saying that they are independent makes no sense.
  • #1
PirateFan308
94
0

Homework Statement


Suppose that A is a 2x2 matrix with eigenvalues 0 and 1. Using diagonalization, show that A2 = A


The Attempt at a Solution


Let [tex]A=\begin{pmatrix}a&b\\c&d\end{pmatrix}[/tex]

Av=λv where [tex]v=\begin{pmatrix}x\\y\end{pmatrix}[/tex] and x,y≠0
If λ=0 then [tex]ax+by=0[/tex] and [tex]cx+dy=0[/tex]
If λ=1 then [tex]ax+by=1[/tex] and [tex]cx+dy=1[/tex]

so Av-λv=0, then Av-λIv=0, then (A-λI)v=0. Since v≠0, then (A-λI)=0
so for λ=0 [tex]\begin{pmatrix}a&b\\c&d\end{pmatrix}[/tex] and [tex]ax+by=0[/tex] and [tex]cx+dy=0[/tex]
For λ=1 [tex]\begin{pmatrix}a-1&b\\c&d-1\end{pmatrix}[/tex] and [tex]ax-x+by=1[/tex] and [tex]cx+dy-y=1[/tex]

We must find two lin. ind. vectors such that we can create X where the first column of X is the first vector, and the second column of X is the second vector.

[tex]X^{-1}AX= \begin{pmatrix}0&0\\0&1\end{pmatrix}[/tex]

If this is true, then [tex]X^{-1}A^{2}X= \begin{pmatrix}0&0\\0&1\end{pmatrix}[/tex]

The problem is, I'm not quite sure how to prove any of this
 
Physics news on Phys.org
  • #2
You don't need to specifically find your matrix X. Only knowing that the matrix X exists would suffice. So the only thing you need is that there exists a matrix X such that

[tex]XAX^{-1}[/tex]

is diagonal. You don't need the specific form of X.
 
  • #3
micromass said:
You don't need to specifically find your matrix X. Only knowing that the matrix X exists would suffice. So the only thing you need is that there exists a matrix X such that

[tex]XAX^{-1}[/tex]

is diagonal. You don't need the specific form of X.

In order for the matrix X to exist, it must be invertible (it must be nxn) and since it must be able to multiply by A (2x2), X must also be 2x2.

Since there are two distinct eigenvalues, there must be two linearly independant eigenvectors. These two distinct eigenvectors can for X.

I'm still a bit confused on how to prove exactly why there must be two linearly independent eigenvectors.
 
  • #4
PirateFan308 said:
In order for the matrix X to exist, it must be invertible (it must be nxn) and since it must be able to multiply by A (2x2), X must also be 2x2.

Since there are two distinct eigenvalues, there must be two linearly independant eigenvectors. These two distinct eigenvectors can for X.

I'm still a bit confused on how to prove exactly why there must be two linearly independent eigenvectors.

You can prove this directly. If v and w are eigenvectors belonging to distinct eigenvalues, then v and w are independent. Thus if [itex]Av=\lambda v[/itex] and if [itex]Aw=\mu w[/itex] and if [itex]\lambda =\mu[/itex], then v and w are independent.

Try to prove this. Prove it by contradiction. Assume that v and w are dependent. What do you know then??
 
  • #5
micromass said:
You can prove this directly. If v and w are eigenvectors belonging to distinct eigenvalues, then v and w are independent. Thus if [itex]Av=\lambda v[/itex] and if [itex]Aw=\mu w[/itex] and if [itex]\lambda =\mu[/itex], then v and w are independent.
You mean "if [itex]\lambda\ne \mu[/itex]".

Try to prove this. Prove it by contradiction. Assume that v and w are dependent. What do you know then??
 
  • #6
micromass said:
Try to prove this. Prove it by contradiction. Assume that v and w are dependent. What do you know then??

If v and w are dependent, then w=cv which could be put into [itex]Av=\lambda v[/itex] and [itex]Aw=\mu cv[/itex] so [itex]\lambda =\mu c[/itex] and they are dependent. But 0 and 1 are independent, so this is a contradiction.
 
  • #7
PirateFan308 said:
If v and w are dependent, then w=cv which could be put into [itex]Av=\lambda v[/itex] and [itex]Aw=\mu cv[/itex] so [itex]\lambda =\mu c[/itex] and they are dependent. But 0 and 1 are independent, so this is a contradiction.

What do you mean, 0 and 1 are independent?? 0 and 1 are numbers, not vectors. Saying that they are independent makes no sense.
 

Related to Using diagonalization, prove the matrix equals it's square

1. What is diagonalization?

Diagonalization is the process of finding a diagonal matrix that is similar to a given matrix. This is done by finding a set of eigenvectors and eigenvalues that can be used to transform the given matrix into a diagonal matrix.

2. How can diagonalization be used to prove that a matrix equals its square?

By diagonalizing the given matrix and its square, we can compare the eigenvalues and eigenvectors of both matrices. If they are the same, then the matrices are equal.

3. Can any matrix be diagonalized?

No, not all matrices can be diagonalized. Only square matrices with distinct eigenvalues can be diagonalized.

4. What is the significance of a matrix being equal to its square?

When a matrix is equal to its square, it is called a idempotent matrix. These matrices have many applications in real-world problems, such as modeling linear systems and finding steady-state solutions.

5. Are there other methods of proving a matrix equals its square?

Yes, there are other methods such as using the Cayley-Hamilton theorem or direct calculation. However, diagonalization is often the most efficient and straightforward method.

Similar threads

  • Calculus and Beyond Homework Help
Replies
2
Views
264
  • Calculus and Beyond Homework Help
Replies
6
Views
368
  • Calculus and Beyond Homework Help
Replies
2
Views
721
  • Calculus and Beyond Homework Help
Replies
6
Views
766
  • Calculus and Beyond Homework Help
Replies
12
Views
1K
  • Calculus and Beyond Homework Help
Replies
3
Views
937
  • Calculus and Beyond Homework Help
Replies
4
Views
906
  • Calculus and Beyond Homework Help
Replies
4
Views
1K
  • Calculus and Beyond Homework Help
Replies
3
Views
872
  • Calculus and Beyond Homework Help
Replies
6
Views
950
Back
Top