description/proof of that for tensors space or tensor product of vectors spaces, transition of standard bases or components is square matrix, and inverse is product of inverses
Topics
About: vectors space
The table of contents of this article
Starting Context
- The reader knows a definition of tensors space with respect to field and k vectors spaces and vectors space over field.
- The reader knows a definition of tensor product of \(k\) vectors spaces over field.
- The reader admits the proposition that for any field and any \(k\) finite-dimensional vectors spaces over the field, the tensors space with respect to the field and the vectors spaces and the field has the basis that consists of the tensor products of the elements of the dual bases of any bases of the vectors spaces.
- The reader admits the proposition that the tensor product of any \(k\) finite-dimensional vectors spaces has the basis that consists of the classes induced by any basis elements.
- The reader admits the proposition that for the tensors space with respect to any field and any \(k\) finite-dimensional vectors spaces over the field and the field, the transition of the standard bases with respect to any bases for the vector spaces is this.
- The reader admits the proposition that for the tensor product of any \(k\) finite-dimensional vectors spaces over any field, the transition of the standard bases with respect to any bases for the vector spaces is this.
- The reader admits the proposition that for any finite-dimensional vectors space, the transition of the components of any vector with respect to any change of bases is this.
Target Context
- The reader will have a description and a proof of the proposition that for the tensors space with respect to any field and any finite number of finite-dimensional the field vectors spaces and the field or the tensor product of any finite-dimensional vectors spaces over any field, the transition of any standard bases or the components is a square matrix, and the inverse matrix is the product of the inverses.
Orientation
There is a list of definitions discussed so far in this site.
There is a list of propositions discussed so far in this site.
Main Body
1: Structured Description
Here is the rules of Structured Description.
Entities:
\(F\): \(\in \{\text{ the fields }\}\)
\(\{V_1, ..., V_k\}\): \(\subseteq \{\text{ the finite-dimensional } F \text{ vectors spaces }\}\)
\(L (V_1, ..., V_k: F)\): \(= \text{ the tensors space }\)
\(V_1 \otimes ... \otimes V_k\): \(= \text{ the tensor product }\)
\(\{B_1, ..., B_k\}\): \(B_j \in \{\text{ the bases of } V_j\} = \{{b_j}_l \vert 1 \le l \le dim V_j\}\)
\(\{B'_1, ..., B'_k\}\): \(B'_j \in \{\text{ the bases of } V_j\} = \{{b'_j}_l \vert 1 \le l \le dim V_j\}\)
\(\{B^*_1, ..., B^*_k\}\): \(B^*_j = \text{ the dual basis of } B_j = \{{b_j}^l \vert 1 \le l \le dim V_j\}\)
\(\{B'^*_1, ..., B'^*_k\}\): \(B'^*_j = \text{ the dual basis of } B'_j = \{{b'_j}^l \vert 1 \le l \le dim V_j\}\)
\(B^*\): \(= \{{b_1}^{j_1} \otimes ... \otimes {b_k}^{j_k} \vert {b_l}^{j_l} \in B^*_l\}\), \(\in \{\text{ the bases for } L (V_1, ..., V_k: F)\}\)
\(B'^*\): \(= \{{b'_1}^{j_1} \otimes ... \otimes {b'_k}^{j_k} \vert {b'_l}^{j_l} \in B'^*_l\}\), \(\in \{\text{ the bases for } L (V_1, ..., V_k: F)\}\)
\(B\): \(= \{[(({b_1}_{j_1}, ..., {b_k}_{j_k}))] \vert {b_l}_{j_l} \in B_l\}\), \(\in \{\text{ the bases for } V_1 \otimes ... \otimes V_k\}\)
\(B'\): \(= \{[(({b'_1}_{j_1}, ..., {b'_k}_{j_k}))] \vert {b'_l}_{j_l} \in B'_l\}\), \(\in \{\text{ the bases for } V_1 \otimes ... \otimes V_k\}\)
//
Statements:
\({b'_j}_l = {b_j}_m {M_j}^m_l\)
\(\implies\)
(
(
\({b'_1}^{j_1} \otimes ... \otimes {b'_k}^{j_k} = {{M_1}^{-1}}^{j_1}_{l_1} ... {{M_k}^{-1}}^{j_k}_{l_k} {b_1}^{l_1} \otimes ... \otimes {b_k}^{l_k}\)
\(\land\)
\(M^{j_1, ..., j_k}_{l_1, ..., l_k} := {{M_1}^{-1}}^{j_1}_{l_1} ... {{M_k}^{-1}}^{j_k}_{l_k}\) is a square matrix
\(\land\)
\({M^{-1}}^{j_1, ..., j_k}_{l_1, ..., l_k} = {M_1}^{j_1}_{l_1} ... {M_k}^{j_k}_{l_k}\)
)
\(\land\)
(
\([(({b'_1}_{j_1}, ..., {b'_k}_{j_k}))] = [(({b_1}_{l_1}, ..., {b_k}_{l_k}))] {M_1}^{l_1}_{j_1} ... {M_k}^{l_k}_{j_k}\)
\(\land\)
\(M^{l_1, ..., l_k}_{j_1, ..., j_k} := {M_1}^{l_1}_{j_1} ... {M_k}^{l_k}_{j_k}\) is a square matrix
\(\land\)
\({M^{-1}}^{l_1, ..., l_k}_{j_1, ..., j_k} = {{M_1}^{-1}}^{l_1}_{j_1} ... {{M_k}^{-1}}^{l_k}_{j_k}\)
)
)
//
2: Proof
Whole Strategy: Step 1: see that the transition for the bases for \(L (V_1, ..., V_k: F)\) holds; Step 2: see that \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) is a square matrix; Step 3: see that the inverse of \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) is as is claimed; Step 4: see that the transition for the bases for \(V_1 \otimes ... \otimes V_k\) holds; Step 5: see that \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\) is a square matrix; Step 6: see that the inverse of \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\) is as is claimed; Step 7: see that also the components transitions are some square matrices.
Step 1:
\(B^*\) and \(B'^*\) are indeed some bases for \(L (V_1, ..., V_k: F)\), by the proposition that for any field and any \(k\) finite-dimensional vectors spaces over the field, the tensors space with respect to the field and the vectors spaces and the field has the basis that consists of the tensor products of the elements of the dual bases of any bases of the vectors spaces.
\({b'_1}^{j_1} \otimes ... \otimes {b'_k}^{j_k} = {{M_1}^{-1}}^{j_1}_{l_1} ... {{M_k}^{-1}}^{j_k}_{l_k} {b_1}^{l_1} \otimes ... \otimes {b_k}^{l_k}\) holds, by the proposition that for the tensors space with respect to any field and any \(k\) finite-dimensional vectors spaces over the field and the field, the transition of the standard bases with respect to any bases for the vector spaces is this.
Step 2:
\(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) may not look like any matrix unless \(k = 1\), because it is a multi-dimensional array.
But the set of the combinations, \(J := \{(j_1, ..., j_k) \vert 1 \le j_1 \le dim V_1, ..., 1 \le j_k \le dim V_k\}\), whose order is \(dim V_1 * ... * dim V_k\), can be regarded as a single index set. And \(J = \{(l_1, ..., l_k) \vert 1 \le l_1 \le dim V_1, ..., 1 \le l_k \le dim V_k\}\) can be regarded as a single index set.
So, \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) can be regarded to be a \((dim V_1 * ... * dim V_k) \times (dim V_1 * ... * dim V_k)\) square matrix: the order of the index, \(J\), can be chosen arbitrary, for example, \((1, ..., 1), (1, ..., 2), ..., (dim V_1, ..., dim V_k)\), which is the most natural one.
Also each of \({b'_1}^{j_1} \otimes ... \otimes {b'_k}^{j_k}\) and \({b_1}^{l_1} \otimes ... \otimes {b_k}^{l_k}\) can be regarded to be a column vector (a kind of matrix) with the chosen order of \(J\).
Then, \({b'_1}^{j_1} \otimes ... \otimes {b'_k}^{j_k} = M^{j_1, ..., j_k}_{l_1, ..., l_k} {b_1}^{l_1} \otimes ... \otimes {b_k}^{l_k}\) is the usual multiplication of the square matrix and the column vector.
In fact, that is natural, because it is a transition of bases for a vectors space, and although we denote the basis, \(B^*\), as \(\{{b_1}^{j_1} \otimes ... \otimes {b_k}^{j_k}\}\), just because that is somehow convenient for clarifying what each element is, in fact, the basis can be denoted also like \(\{e^1, ..., e^{dim V_1 * ... * dim V_k}\}\).
For any another matrix, \(N^{m_1, ..., m_k}_{j_1, ..., j_k}\), with the chosen order of \(J\), \(N^{m_1, ..., m_k}_{j_1, ..., j_k} M^{j_1, ..., j_k}_{l_1, ..., l_k}\) is the usual multiplication of the square matrices.
Step 3:
The reason why we want to regard \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) as a square matrix is that we want to take the inverse of it, while the reason why we want to take the inverse of it is that the inverse represents the transition of the components, by the proposition that for any finite-dimensional vectors space, the transition of the components of any vector with respect to any change of bases is this: it certainly has the inverse, because it is a transition of bases.
The inverse of \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) is the matrix, \(N^{m_1, ..., m_k}_{j_1, ..., j_k}\), such that \(N^{m_1, ..., m_k}_{j_1, ..., j_k} M^{j_1, ..., j_k}_{l_1, ..., l_k} = \delta^{m_1}_{l_1} ... \delta^{m_k}_{l_k}\): the product of the reverse order is automatically guaranteed to be \(I\), because we know that \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) is invertible: from \(N M = I\), \(M N = M N M M^{-1} = M I M^{-1} = I\).
There is \({M_1}^{m_1}_{j_1} ... {M_k}^{m_k}_{j_k}\), which is a \((dim V_1 * ... * dim V_k) \times (dim V_1 * ... * dim V_k)\) matrix.
\({M_1}^{m_1}_{j_1} ... {M_k}^{m_k}_{j_k} M^{j_1, ..., j_k}_{l_1, ..., l_k} = {M_1}^{m_1}_{j_1} ... {M_k}^{m_k}_{j_k} {{M_1}^{-1}}^{j_1}_{l_1} ... {{M_k}^{-1}}^{j_k}_{l_k} = {M_1}^{m_1}_{j_1} {{M_1}^{-1}}^{j_1}_{l_1} ... {M_k}^{m_k}_{j_k} ... {{M_k}^{-1}}^{j_k}_{l_k} = \delta^{m_1}_{l_1} ... \delta^{m_k}_{l_k}\), which means that \({M_1}^{m_1}_{j_1} ... {M_k}^{m_k}_{j_k}\) is the inverse of \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\).
So, \({M^{-1}}^{j_1, ..., j_k}_{l_1, ..., l_k} = {M_1}^{j_1}_{l_1} ... {M_k}^{j_k}_{l_k}\).
\({M^{-1}}^{j_1, ..., j_k}_{l_1, ..., l_k}\) represents the transition of tensor components, by the proposition that for any finite-dimensional vectors space, the transition of the components of any vector with respect to any change of bases is this.
Step 4:
\(B\) and \(B'\) are indeed some bases for \(V_1 \otimes ... \otimes V_k\), by the proposition that the tensor product of any \(k\) finite-dimensional vectors spaces has the basis that consists of the classes induced by any basis elements.
\([(({b'_1}_{j_1}, ..., {b'_k}_{j_k}))] = [(({b_1}_{l_1}, ..., {b_k}_{l_k}))] {M_1}^{l_1}_{j_1} ... {M_k}^{l_k}_{j_k}\) holds, by the proposition that for the tensor product of any \(k\) finite-dimensional vectors spaces over any field, the transition of the standard bases with respect to any bases for the vector spaces is this.
Step 5:
\(M^{l_1, ..., l_k}_{j_1, ..., j_k}\) may not look like any matrix unless \(k = 1\), because it is a multi-dimensional array.
But the set of the combinations, \(J := \{(j_1, ..., j_k) \vert 1 \le j_1 \le dim V_1, ..., 1 \le j_k \le dim V_k\}\), whose order is \(dim V_1 * ... * dim V_k\), can be regarded as a single index set. And \(J = \{(l_1, ..., l_k) \vert 1 \le l_1 \le dim V_1, ..., 1 \le l_k \le dim V_k\}\) can be regarded as a single index set.
So, \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\) can be regarded to be a \((dim V_1 * ... * dim V_k) \times (dim V_1 * ... * dim V_k)\) square matrix: the order of the index, \(J\), can be chosen arbitrary, for example, \((1, ..., 1), (1, ..., 2), ..., (dim V_1, ..., dim V_k)\), which is the most natural one.
Also each of \([(({b'_1}_{j_1}, ..., {b'_k}_{j_k}))]\) and \([(({b_1}_{l_1}, ..., {b_k}_{l_k}))]\) can be regarded to be a row vector (a kind of matrix) with the chosen order of \(J\).
Then, \([(({b'_1}_{j_1}, ..., {b'_k}_{j_k}))] = [(({b_1}_{l_1}, ..., {b_k}_{l_k}))] M^{l_1, ..., l_k}_{j_1, ..., j_k}\) is the usual multiplication of the row vector and the square matrix.
In fact, that is natural, because it is a transition of bases for a vectors space, and although we denote the basis, \(B\), as \(\{[(({b_1}_{j_1}, ..., {b_k}_{j_k}))]\}\), just because that is somehow convenient for clarifying what each element is, in fact, the basis can be denoted also like \(\{e_1, ..., e_{dim V_1 * ... * dim V_k}\}\).
For any another matrix, \(N^{m_1, ..., m_k}_{l_1, ..., l_k}\), with the chosen order of \(J\), \(N^{m_1, ..., m_k}_{l_1, ..., l_k} M^{l_1, ..., l_k}_{j_1, ..., j_k}\) is the usual multiplication of the square matrices.
Step 6:
The reason why we want to regard \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\) as a square matrix is that we want to take the inverse of it, while the reason why we want to take the inverse of it is that the inverse represents the transition of the components: it certainly has the inverse, because it is a transition of bases.
The inverse of \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\) is the matrix, \(N^{m_1, ..., m_k}_{l_1, ..., l_k}\), such that \(N^{m_1, ..., m_k}_{l_1, ..., l_k} M^{l_1, ..., l_k}_{j_1, ..., j_k} = \delta^{m_1}_{j_1} ... \delta^{m_k}_{j_k}\): the product of the reverse order is automatically guaranteed to be \(I\), because we know that \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\) is invertible.
There is \({{M_1}^{-1}}^{m_1}_{l_1} ... {{M_k}^{-1}}^{m_k}_{l_k}\), which is a \((dim V_1 * ... * dim V_k) \times (dim V_1 * ... * dim V_k)\) matrix.
\({{M_1}^{-1}}^{m_1}_{l_1} ... {{M_k}^{-1}}^{m_k}_{l_k} M^{l_1, ..., l_k}_{j_1, ..., j_k} = {{M_1}^{-1}}^{m_1}_{l_1} ... {{M_k}^{-1}}^{m_k}_{l_k} {M_1}^{l_1}_{j_1} ... {M_k}^{l_k}_{j_k} = {{M_1}^{-1}}^{m_1}_{l_1} {M_1}^{l_1}_{j_1} ... {{M_k}^{-1}}^{m_k}_{l_k} ... {M_k}^{l_k}_{j_k} = \delta^{m_1}_{j_1} ... \delta^{m_k}_{j_k}\), which means that \({{M_1}^{-1}}^{m_1}_{l_1} ... {{M_k}^{-1}}^{m_k}_{l_k}\) is the inverse of \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\).
So, \({M^{-1}}^{l_1, ..., l_k}_{j_1, ..., j_k} = {{M_1}^{-1}}^{l_1}_{j_1} ... {{M_k}^{-1}}^{l_k}_{j_k}\).
\({M^{-1}}^{l_1, ..., l_k}_{j_1, ..., j_k}\) represents the transition of tensor components, by the proposition that for any finite-dimensional vectors space, the transition of the components of any vector with respect to any change of bases is this.
Step 7:
So, we have gotten the components transitions, \({M^{-1}}^{j_1, ..., j_k}_{l_1, ..., l_k}\) and \({M^{-1}}^{l_1, ..., l_k}_{j_1, ..., j_k}\), also which are some matrices likewise, and the inverses are \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) and \(M^{l_1, ..., l_k}_{j_1, ..., j_k}\).
3: Note
We are talking about the transition of bases or the transition of components, not about the tensor components themselves: for example, for a tensor, \(t \in L ({V_1}^*, ..., {V_k}^*, V_1, ..., V_k: F)\), \(t\) can be expressed with the components with respect to a standard basis as \(t^{j_1, ..., j_k}_{l_1, ..., l_k}\), which resembles \(M^{j_1, ..., j_k}_{l_1, ..., l_k}\) in form, but it is not so natural to regard it as a matrix: thinking of \(t (v^1, ... v^k, v_1, ..., v_k) = t^{j_1, ..., j_k}_{l_1, ..., l_k} {v^1}_{j_1} ... {v^k}_{j_k} {v_1}^{l_1} ... {v_k}^{l_k}\), in order to regard it as the multiplication of a matrix and a column vector, the column vector would be like \(({v^1}_1 ... {v^k}_1 {v_1}^1 ... {v_k}^1, {v^1}_1 ... {v^k}_1 {v_1}^1 ... {v_k}^2, ..., {v^1}_{dim V_1} ... {v^k}_{dim V_k} {v_1}^{dim V_1} ... {v_k}^{dim V_k})^t\) instead of like \(({v_1}_1, ..., {v_1}_{dim V_1}, ..., {v_k}^1, ..., {v_k}^{dim V_k})\), which might not be particularly meaningful (if it is meaningful for your situation, of course, it is fine).
The reason why that might not be meaningful is that \(t\) is not any linear map in general (refer to the proposition that a multilinear map is not necessarily linear): \(t\) is \(: V_1^{*} \times ... \times V_k^{*} \times V_1 \times ... \times V_k \to F\), a non-linear map from a \((2 (dim V_1 + ... + dim V_k))\)-dimensional vectors space into \(F\), and thinking of the \((dim V_1 * ... * dim V_k)^2 \times (dim V_1 * ... * dim V_k)^2\) matrix is not meaningful in general.
Of course, as any matrix is just an arrangement of some ring elements, you can always regard \(t^{j_1, ..., j_k}_{l_1, ..., l_k}\) as a matrix, if you want to.