当前位置:网站首页>Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"
Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"
2022-06-26 23:40:00 【No card, no change】
The following mainly focuses on the following issues :
- The physical meaning of Schmidt orthogonalization
- The meaning of characteristic subspace
- How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”( a key !)
The physical meaning of Schmidt orthogonalization
The process of Schmidt orthogonalization :
Yes α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization ( Hereinafter referred to as ” Orthogonalization “)
- β 1 = α 1 \beta_1 = \alpha_1 β1=α1
- β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2−(β1,β1)(α2,β1)β1
Understand it as a whole , ( α 2 , β 1 ) (\alpha_2, \beta_1) (α2,β1) Express α 2 \alpha_2 α2 and β 1 \beta_1 β1 Inner product , Is a number ; similarly , ( β 1 , β 1 ) (\beta_1,\beta_1) (β1,β1) It's also a number ; β 1 \beta_1 β1 Namely α 1 \alpha_1 α1.
therefore β 2 \beta_2 β2 It's essentially α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The linear combination of .
Simplify β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2−(β1,β1)(α2,β1)β1 :
β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 = α 2 − ∣ α 2 ∣ ∣ β 1 ∣ c o s θ ∣ β 1 ∣ 2 β 1 = α 2 − ∣ α 2 ∣ c o s θ ∣ β 1 ∣ β 1 = α 2 − ∣ α 2 ∣ β 1 ∣ β 1 ∣ c o s θ \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 \\ =\alpha_2 - \frac{|\alpha_2||\beta_1|cos\theta}{|\beta_1|^2}\beta_1 \\ = \alpha_2 - \frac{|\alpha_2|cos\theta}{|\beta_1|}\beta_1 \\ = \alpha_2 - |\alpha_2|\frac{\beta_1}{|\beta_1|}cos\theta β2=α2−(β1,β1)(α2,β1)β1=α2−∣β1∣2∣α2∣∣β1∣cosθβ1=α2−∣β1∣∣α2∣cosθβ1=α2−∣α2∣∣β1∣β1cosθ
among θ \theta θ by α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The angle between .
After Visualization :

The meaning of characteristic subspace
Simple understanding : For a matrix A A A A characteristic value of λ \lambda λ , The linear independent eigenvector corresponding to the eigenvalue is α 1 \alpha_1 α1 、 α 2 \alpha_2 α2 、 …… 、 α s \alpha_s αs, this s s s The set of all vectors and zero vectors that can be formed by a linear combination of linearly independent eigenvectors is called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of .
Well defined :
Given n n n Order matrix A A A , Regulations V λ = { α ∈ C n ∣ A α = λ α } V_\lambda = \{\alpha∈C^n|A\alpha=\lambda \alpha\} Vλ={ α∈Cn∣Aα=λα}
(1) V λ V_\lambda Vλ Non empty : 0 ∈ V λ 0 ∈ V_\lambda 0∈Vλ
(2) V λ V_\lambda Vλ Close to addition :
α , β ∈ V λ ⇒ A α = λ α , A β = λ β ⇒ A ( α + β ) = λ ( α + β ) ⇒ α + β ∈ V λ \alpha,\beta∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha,A\beta=\lambda\beta \\ \Rightarrow A(\alpha+\beta)=\lambda(\alpha+\beta) \\ \Rightarrow \alpha+\beta∈V_\lambda α,β∈Vλ⇒Aα=λα,Aβ=λβ⇒A(α+β)=λ(α+β)⇒α+β∈Vλ
(3) V λ V_\lambda Vλ Logarithmic multiplication closed :
α ∈ V λ ⇒ A α = λ α ⇒ A ( k α ) = k ( A α ) = k ( λ α ) = λ ( k α ) ⇒ k α ∈ V λ \alpha∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha\\ \Rightarrow A(k\alpha)=k(A\alpha)=k(\lambda\alpha) = \lambda(k\alpha) \\ \Rightarrow k\alpha∈ V_\lambda α∈Vλ⇒Aα=λα⇒A(kα)=k(Aα)=k(λα)=λ(kα)⇒kα∈Vλ
(4) α 1 , α 2 , . . . , α s ∈ V λ ⇒ k 1 α 1 + k 2 α 2 + . . . + k s α s ∈ V λ , k i ∈ C \alpha_1,\alpha_2,...,\alpha_s∈ V_\lambda \Rightarrow k_1\alpha_1+k_2\alpha_2 + ... + k_s\alpha_s∈V_\lambda,k_i∈ C α1,α2,...,αs∈Vλ⇒k1α1+k2α2+...+ksαs∈Vλ,ki∈C
(5) V λ V_\lambda Vλ yes n n n Subspace of dimensional complex vector space
λ \lambda λ yes A A A The eigenvalues of the ⇔ \Leftrightarrow ⇔ V λ ≠ 0 V_\lambda ≠{0} Vλ=0
V λ V_\lambda Vλ be called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of
How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”
The reason for this thinking comes from a class of linear algebra problems :
Known matrix A = ? A=? A=?( Real symmetric matrix )
(1) Find invertible matrix P P P, bring P − 1 A P P^{-1}AP P−1AP It's a diagonal matrix
(2) Find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix
The students who have reviewed everything must be able to calculate it skillfully .
First question : All linearly independent eigenvectors corresponding to each eigenvalue are arranged together ;
Second questions : take The same eigenvalue After Schmidt orthogonalization of all corresponding linear independent eigenvectors , Normalize all eigenvectors , Then put all the linearly independent eigenvectors together .
\space
We all know that the eigenvectors corresponding to different eigenvalues of a real symmetric matrix are not only linearly independent , But also orthogonal ( This is easier to prove ); So we only need to orthogonalize all eigenvectors corresponding to the same eigenvalue .
But you may have never seen a class of problems ,“ Prescribed matrix A A A Not a real symmetric matrix , It is required to find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix ”, because In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized .
\space
Why is that ? The reason will be explained below .
“ In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized ” The fundamental reason is that Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector .
For example, two-dimensional vector , hypothesis A = ( 1 − 1 0 2 ) A=\left(\begin{matrix} 1 & -1 \\ 0 & 2 \end{matrix}\right) A=(10−12), Its eigenvalue λ 1 = 1 、 λ 2 = 2 \lambda_1=1、\lambda_2 = 2 λ1=1、λ2=2, The corresponding eigenvectors are α 1 = ( 1 0 ) 、 α 2 = ( 1 − 1 ) \alpha_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\alpha_2=\left(\begin{matrix} 1 \\ -1 \end{matrix}\right) α1=(10)、α2=(1−1), By orthogonalizing them, we get β 1 = ( 1 0 ) 、 β 2 = ( 0 1 ) \beta_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\beta_2=\left(\begin{matrix} 0 \\ 1 \end{matrix}\right) β1=(10)、β2=(01). But clearly , A β 2 ≠ λ 2 β 2 A\beta_2≠\lambda_2\beta_2 Aβ2=λ2β2, That is to say, after orthogonalization β 2 \beta_2 β2 The corresponding characteristic value is no longer λ 2 \lambda_2 λ2 了 .
Why? “ Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector ”
Visualize the above example :
After Schmidt orthogonalization, the two vectors are indeed orthogonal , But according to the following properties of eigenvalues and eigenvectors , because β 2 \beta_2 β2 Not by α 2 \alpha_2 α2 Linear representation , therefore β 2 \beta_2 β2 Not eigenvalues λ 2 \lambda_2 λ2 Eigenvector of , That is to say β 2 \beta_2 β2 be not in λ 2 \lambda_2 λ2 In the characteristic subspace of , therefore β 2 \beta_2 β2 The corresponding characteristic value is not λ 2 \lambda_2 λ2. In this way, orthogonalization is meaningless .
nature : if α 1 , α 2 , . . . , α t \alpha_1,\alpha_2,...,\alpha_t α1,α2,...,αt All are matrices. A A A Of are characteristic values λ \lambda λ Eigenvector of , Then when k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt When it's not zero , k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt Still matrix A A A Belong to characteristic value λ \lambda λ Eigenvector of .
To show more clearly , We expand into three-dimensional space .
Hypothetical matrix A = ( 3 2 0 1 2 0 2 0 1 2 0 3 2 ) A=\left(\begin{matrix} \frac{3}{2} & 0 & \frac{1}{2} \\ 0 & 2 & 0 \\ \frac{1}{2} & 0 & \frac{3}{2} \end{matrix}\right) A=⎝⎛2302102021023⎠⎞ Is a third-order real symmetric matrix , The three eigenvalues are λ 1 = 1 、 λ 2 = 2 、 λ 3 = 2 \lambda_1=1、\lambda_2=2、\lambda_3=2 λ1=1、λ2=2、λ3=2, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Because the number of linearly independent eigenvectors of a real symmetric matrix must be the same as the order , therefore λ 1 \lambda_1 λ1 The corresponding eigenvector is α 1 = ( − 1 , 0 , 1 ) T \alpha_1=(-1,0,1)^T α1=(−1,0,1)T, λ 2 \lambda_2 λ2 and λ 3 \lambda_3 λ3 The corresponding eigenvectors are α 2 = ( 0 , 1 , 0 ) T \alpha_2=(0,1,0)^T α2=(0,1,0)T and α 3 = ( 1 , 2 , 1 ) T \alpha_3=(1,2,1)^T α3=(1,2,1)T, And the eigenvectors corresponding to different eigenvalues are orthogonal , namely ( α 1 , α 2 ) = 0 (\alpha_1,\alpha_2)=0 (α1,α2)=0, ( α 1 , α 3 ) = 0 (\alpha_1,\alpha_3)=0 (α1,α3)=0, α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Not orthogonal .
hypothesis α 1 \alpha_1 α1、 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In 3D space, the following :

Yes α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization :

It can be found that the three vectors are orthogonal , And β 2 \beta_2 β2 and β 3 \beta_3 β3 The corresponding characteristic value is still 2 2 2.
If we deal with the matrix A A A Not a real symmetric matrix , Just a general matrix , That is to say, even the eigenvectors corresponding to different eigenvalues are not necessarily orthogonal , What happens if we orthogonalize two eigenvectors corresponding to different eigenvalues ?
First of all, we need to be clear about , If λ \lambda λ The characteristic subspace of is determined by two linearly independent vectors , Then the eigensubspace of the eigenvalue should be composed of all the vectors in the plane determined by the two linearly independent vectors , That is to say, the eigenvector satisfying the eigenvalue should be in the plane , The vectors that do not satisfy the eigenvalue are all out of plane .
Suppose for a non real symmetric third-order matrix A A A for , Its characteristic value is recorded as λ 1 、 λ 2 、 λ 3 \lambda_1、\lambda_2、\lambda_3 λ1、λ2、λ3, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Suppose the eigenvectors corresponding to the three eigenvalues α 1 \alpha_1 α1、 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 as follows :

Obviously, none of the three is orthogonal .
If the α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Schmidt orthogonalization , Because orthogonalization is only the addition and multiplication of two vectors , So it will not affect the plane formed by the two , in other words α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization β 2 \beta_2 β2 and β 3 \beta_3 β3 Still in α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane formed by , namely β 2 \beta_2 β2 and β 3 \beta_3 β3 In the eigensubspace of the same eigenvalue , That is to say, the eigenvalues of Schmidt orthogonality have not changed .
If the α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization , take α 2 \alpha_2 α2 As β 1 \beta_1 β1, take α 1 \alpha_1 α1 Convert to β 2 \beta_2 β2, You can get β 1 = α 2 = ( 1 , 1 , 0 ) T \beta_1 = \alpha_2 = (1, 1, 0)^T β1=α2=(1,1,0)T, β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T, The drawing is as follows :

The light blue dotted line indicates λ 1 \lambda_1 λ1 The characteristic subspace of , One dimensional space . After Schmidt orthogonalization λ 1 \lambda_1 λ1 The corresponding eigenvectors α 1 \alpha_1 α1 Turned into ( 1 2 , 1 , 1 2 ) T (\frac{1}{2}, 1, \frac{1}{2})^T (21,1,21)T, Obviously the vector is no longer λ 1 \lambda_1 λ1 In the characteristic subspace of , This explanation β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T The corresponding characteristic value is not λ 1 \lambda_1 λ1;
Similarly , Can also be α 1 \alpha_1 α1 As β 1 \beta_1 β1, take α 2 \alpha_2 α2 Convert to β 2 \beta_2 β2, After Schmidt orthogonalization, we will find that β 2 \beta_2 β2 No longer by α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane of , explain β 2 \beta_2 β2 The characteristic value of is changed .
Through the example above , It can be concluded that the Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will lead to the change of eigenvalues corresponding to new vectors , Because the essence of Schmidt orthogonalization is the addition and multiplication of vectors , After Schmidt orthogonalization, the vector in the same feature subspace is still in the original feature subspace , But it is hard to say that the vectors of different feature subspaces are orthogonalized .
Tang Jiafeng's teaching plan mentioned a property :
set up A A A by n n n Order matrix , λ 1 \lambda_1 λ1, λ 2 \lambda_2 λ2 by A A A Two different eigenvalues of , also A α = λ 1 α A\alpha=\lambda_1\alpha Aα=λ1α, A β = λ 2 β A\beta=\lambda_2\beta Aβ=λ2β( α \alpha α and β \beta β It's a non-zero vector ), For any a ≠ 0 a≠0 a=0, b ≠ 0 b ≠ 0 b=0, vector a α + b β a\alpha+b\beta aα+bβ It must not be an eigenvector .
\space
Schmidt orthogonalization is a process of linear combination , According to this property, it can also be explained that the eigenvectors corresponding to different eigenvalues are not eigenvectors obtained by orthogonalization .
Schmidt orthogonalization when doing problems
If you want to compute an orthogonal matrix Q Q Q, bring Q T A Q = Λ Q^TAQ=\Lambda QTAQ=Λ, be A A A It must be a real symmetric matrix , And it is necessary to Schmidt orthogonalize the eigenvector corresponding to the same eigenvalue , Then normalize all eigenvectors , The normalized eigenvectors are arranged to obtain Q Q Q;
If you want to calculate a reversible matrix P P P, bring P − 1 A P = Λ P^{-1}AP=\Lambda P−1AP=Λ, be A A A The requirements are not so strict . If A A A Is a non real symmetric matrix , that P P P It must not be transformed into an orthogonal matrix .
边栏推荐
- Why does EDR need defense in depth to combat ransomware?
- Cvpr2022 stereo matching of asymmetric resolution images
- The user adds a timer function in the handler () goroutine. If it times out, it will be kicked out
- Introduction de l'opérateur
- Is it reliable to open an account for stock trading on the mobile phone? Is it safe to open an account for stock trading on the Internet
- 手机上可以开户炒股吗 网上开户炒股安全吗
- ubuntu上安装mysql
- Smartbi gives you a piece to play with Boston matrix
- Leetcode 718. 最长重复子数组(暴力枚举,待解决)
- leetcode 1143. Longest common subsequence (medium)
猜你喜欢

简单测试轻量级表达式计算器Flee

PHP代码审计系列(一) 基础:方法、思路、流程

邮箱附件钓鱼常用技法
![[微服务]Nacos](/img/69/6641e943c4366d5591acdf9e12389c.png)
[微服务]Nacos

The processing of private chat function in go language
![Selenium电脑上怎么下载-Selenium下载和安装图文教程[超详细]](/img/ec/1c324dcf38d07742a139aac2bab02e.png)
Selenium电脑上怎么下载-Selenium下载和安装图文教程[超详细]

Solid and ambient colors

DAST black box vulnerability scanner part 5: vulnerability scanning engine and service capability

50 tips that unity beginners can definitely use

不会写免杀也能轻松过defender上线CS
随机推荐
Nacos安装指南
Unity4.6 Download
【LeetCode】1984. Minimum difference between highest and lowest of K scores
Tensorrt笔记(七)Tensorrt使用问题整理
手机上炒股开户可靠吗 网上开户炒股安全吗
Why don't I recommend going to sap training institution for training?
泰国安全又划算的支付方式
[微服务]Nacos
在手机开户买股票安全吗 网上开户炒股安全吗
简单测试轻量级表达式计算器Flee
买基金在哪里开户买比较安全
The client implements client Go client type definition connection
论文学习——降雨场次划分方法对降雨控制率的影响分析
利用burp精准定位攻击者
go语言的爬虫和中间件
买股票在手机上开户安全吗 网上开户炒股安全吗
新股民如何网上开户 网上开户炒股安全吗
【Try to Hack】正向shell和反向shell
炒股手机上开户可靠吗? 网上开户炒股安全吗
Is it safe to open an account on the mobile phone to buy stocks? Is it safe to open an account on the Internet to speculate in stocks