Abstract
As I have mentioned already, we may also view matrices as the representations of operators acting on the elements of an LVS. In an n-dimensional Euclidean space (denoted by \(\mathbb {R}^n\)), the natural basisĀ is given by the ket vectors represented by the column matrices
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
I have yet to define the Hermitian conjugate of an operator, and what is meant by saying that it is Hermitian or self-adjoint (and the subtle distinction between the two). That will be done in Chap. 14, Sect. 14.2.3. For the present, take it to be that operator whose matrix representation is the Hermitian conjugate (or complex conjugate transpose) of the matrix representing the operator itself.
- 2.
Here I have used the symbol U to denote the matrix representing a general element of the group U(2), but this should not cause any confusion.
- 3.
I reiterate that we are considering finite-dimensional matrices here. In the case of infinite-dimensional matrices, it is possible for a matrix to have right eigenvectors but no left eigenvectors, or vice versa.
- 4.
Or columns. The statements that follow are valid if ārowā is replaced with ācolumnā. Recall also that the value of a determinant does not change if its rows and columns are interchanged.
- 5.
It is important to note that I am now referring to the linearly independent eigenvectors of the matrix, and not to its linearly independent rows or columns.
- 6.
As mentioned earlier, in the case of finite-dimensional matrices (which is what we are concerned with here), āself-adjointā is the same as āHermitianā. The distinction between Hermitian and self-adjoint operators in the general case will be explained in Chap. 14, Sect. 14.2.3.
- 7.
This binary operation is not to be confused with the inner product of two elements of the LVS.
- 8.
You might then ask why the generators \(J_i\) of rotations in three dimensions are Hermitian rather than antisymmetric. This has to do with real versus complex Lie algebras.
- 9.
I have used the notation \(|\psi _{i}\rangle \) and \(\langle \chi _{i}|\) for the right and left eigenvectors, so as to avoid any possible confusion with the natural basis, for which I have used the symbols \(|\phi _{i}\rangle \) and \(\langle \phi _{i}|\).
- 10.
Note that the term āspectral decompositionā is used sometimes for the representation \(S^{-1}MS = \Lambda \) of a diagonalizable matrix.
- 11.
That is, the level in each glass now becomes \(\tfrac{1}{2}(x_{0}+y_{0})\).
- 12.
Each column of M also adds upĀ to unity. Hence M is, in fact, a doubly stochastic matrix. This last property is no longer valid in the general case, i.e., when the process is conducted with \(n > 3\) glasses, as you will see subsequently.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
Ā© 2020 Authors
About this chapter
Cite this chapter
Balakrishnan, V. (2020). More About Matrices. In: Mathematical Physics. Springer, Cham. https://doi.org/10.1007/978-3-030-39680-0_12
Download citation
DOI: https://doi.org/10.1007/978-3-030-39680-0_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-39679-4
Online ISBN: 978-3-030-39680-0
eBook Packages: Physics and AstronomyPhysics and Astronomy (R0)