If S and T are matrices, and tilde (~) denotes the transpose of the matrix. \[\widetilde{\textbf{S}\textbf{T}}=\tilde{\textbf{T}}\tilde{\textbf{S}} \] How can I prove this statement.
similary \[\left({\textbf{S}\textbf{T}}\right)^{-1}={\textbf{T}}^{-1}{\textbf{S}^{-1}} \]
T(ST)' = T T'S' = S' ST(ST)' = SS' = 1 (ST)(ST)' = 1
So \[\textbf S \textbf T \widetilde{\textbf S \textbf T}=1\]\[\textbf T \widetilde{\textbf S \textbf T}=\textbf S^{-1}\]\[\widetilde{\textbf S \textbf T}=\tilde{\textbf T} \tilde{\textbf S}\] that is the whole proof
excusing the typographical error
hehe .. no probs man
How come we can assume the first line?
dosent it rely on the result?
can't seem to remember. I think it got something to do with the property of multiplication of Matrix with it's transpose. TT' = must equal to something Since if we consider matrix to be system of n vectors with m components. It gives you dot product with itself.
the inner product of a normalized vector with it self which must be unitary
well this is the special case of unitary matrix. It must be for general too.
hmm ok,
(a)ij 1<=i<=m, 1<=j<=n represent matrix A (b)kj 1<=j<=n, 1<=k<=0 represent matrix B AxB = summation j (a)ij(b)kj Must be something like that ... and since transpose changes rows into colums, A' = (a)ji
A = (a)ij, B = (b)jk, AB = summation j (a)ij(b)jk = (ab)ik (AB)T = (ab)ki = summation j (b)jk(a)ij = BTxAT excellent proof here https://www.google.com/search?rlz=1C1ASUT_enNP447NP447&sourceid=chrome&ie=UTF-8&q=how+to+prove+AB+transpose+is+B+transpose+A+transpose
thank you experimentX
summation j (b)jk(a)ij <--- it cannot be a matrix multiplication unless it is transpose (because rows are not equal to columns) You are welcome.
yeah because it works even if the matrices are not square .
That is a great article thank you again, the author even alludes to the outer product,
Join our real-time social learning platform and learn together with your friends!