Although seemingly different, the various approaches to defining tensors describe the same geometric concept using different language and at different levels of abstraction.
As multidimensional arrays A tensor may be represented as a (potentially multidimensional) array. Just as a
vector in an -
dimensional space is represented by a
one-dimensional array with components with respect to a given
basis, any tensor with respect to a basis is represented by a multidimensional array. For example, a
linear operator is represented in a basis as a two-dimensional square array. The numbers in the multidimensional array are known as the
components or
elements of the tensor. They are denoted by indices giving their position in the array, as
subscripts and superscripts, following the symbolic name of the tensor. For example, the components of an order- tensor could be denoted , where and are indices running from to , or also by . Whether an index is displayed as a superscript or subscript depends on the transformation properties of the tensor, described below. Thus while and can both be expressed as
n-by-
n matrices, and are numerically related via
index juggling, the difference in their transformation laws indicates it would be improper to add them together. The total number of indices () required to identify each component uniquely is equal to the
dimension or the number of
ways of an array, which is why a tensor is sometimes referred to as an -dimensional array or an -way array. The total number of indices is also called the
order,
degree or
rank of a tensor, although the term "rank" generally has
another meaning in the context of matrices and tensors. Just as the components of a vector change when we change the
basis of the vector space, the components of a tensor also change under such a transformation. Each type of tensor comes equipped with a
transformation law that details how the components of the tensor respond to a
change of basis. The components of a vector can respond in two distinct ways to a
change of basis (see
Covariance and contravariance of vectors), where the new
basis vectors \mathbf{\hat{e}}_i are expressed in terms of the old basis vectors \mathbf{e}_j as, :\mathbf{\hat{e}}_i = \sum_{j=1}^n \mathbf{e}_j R^j_i = \mathbf{e}_j R^j_i . Here ''R'
j'i
are the entries of the change of basis matrix, and in the rightmost expression the summation sign was suppressed: this is the Einstein summation convention, which will be used throughout this article. The components v''
i of a column vector
v transform with the
inverse of the matrix
R, :\hat{v}^i = \left(R^{-1}\right)^i_j v^j, where the hat denotes the components in the new basis. This is called a
contravariant transformation law, because the vector components transform by the
inverse of the change of basis. In contrast, the components,
wi, of a covector (or row vector),
w, transform with the matrix
R itself, :\hat{w}_i = w_j R^j_i . This is called a
covariant transformation law, because the covector components transform by the
same matrix as the change of basis matrix. The components of a more general tensor are transformed by some combination of covariant and contravariant transformations, with one transformation law for each index. If the transformation matrix of an index is the inverse matrix of the basis transformation, then the index is called
contravariant and is conventionally denoted with an upper index (superscript). If the transformation matrix of an index is the basis transformation itself, then the index is called
covariant and is denoted with a lower index (subscript). As a simple example, the matrix of a linear operator with respect to a basis is a rectangular array T that transforms under a change of basis matrix R = \left(R^j_i\right) by \hat{T} = R^{-1}TR. For the individual matrix entries, this transformation law has the form \hat{T}^{i'}_{j'} = \left(R^{-1}\right)^{i'}_i T^i_j R^j_{j'} so the tensor corresponding to the matrix of a linear operator has one covariant and one contravariant index: it is of type (1,1). Combinations of covariant and contravariant components with the same index allow us to express geometric invariants. For example, the fact that a vector is the same object in different coordinate systems can be captured by the following equations, using the formulas defined above: :\mathbf{v} = \hat{v}^i \,\mathbf{\hat{e}}_i = \left( \left(R^{-1}\right)^i_j {v}^j \right) \left( \mathbf_k R^k_i \right) = \left( \left(R^{-1}\right)^i_j R^k_i \right) {v}^j \mathbf_k = \delta_j^k {v}^j \mathbf_k = {v}^k \,\mathbf_k = {v}^i \,\mathbf_i , where \delta^k_j is the
Kronecker delta, which functions similarly to the
identity matrix, and has the effect of renaming indices (
j into
k in this example). This shows several features of the component notation: the ability to re-arrange terms at will (
commutativity), the need to use different indices when working with multiple objects in the same expression, the ability to rename indices, and the manner in which contravariant and covariant tensors combine so that all instances of the transformation matrix and its inverse cancel, so that expressions like {v}^i \,\mathbf_i can immediately be seen to be geometrically identical in all coordinate systems. Similarly, a linear operator, viewed as a geometric object, does not actually depend on a basis: it is just a linear map that accepts a vector as an argument and produces another vector. The transformation law for how the matrix of components of a linear operator changes with the basis is consistent with the transformation law for a contravariant vector, so that the action of a linear operator on a contravariant vector is represented in coordinates as the matrix product of their respective coordinate representations. That is, the components (Tv)^i are given by (Tv)^i = T^i_j v^j. These components transform contravariantly, since :\left(\widehat{Tv}\right)^{i'} = \hat{T}^{i'}_{j'} \hat{v}^{j'} = \left[ \left(R^{-1}\right)^{i'}_i T^i_j R^j_{j'} \right] \left[ \left(R^{-1}\right)^{j'}_k v^k \right] = \left(R^{-1}\right)^{i'}_i (Tv)^i . The transformation law for an order tensor with
p contravariant indices and
q covariant indices is thus given as, : \hat{T}^{i'_1, \ldots, i'_p}_{j'_1, \ldots, j'_q} = \left(R^{-1}\right)^{i'_1}_{i_1} \cdots \left(R^{-1}\right)^{i'_p}_{i_p} T^{i_1, \ldots, i_p}_{j_1, \ldots, j_q} R^{j_1}_{j'_1}\cdots R^{j_q}_{j'_q}. Here the primed indices denote components in the new coordinates, and the unprimed indices denote the components in the old coordinates. Such a tensor is said to be of order or
type . The terms "order", "type", "rank", "valence", and "degree" are all sometimes used for the same concept. Here, the term "order" or "total order" will be used for the total dimension of the array (or its generalization in other definitions), in the preceding example, and the term "type" for the pair giving the number of contravariant and covariant indices. A tensor of type is also called a -tensor for short. This discussion motivates the following formal definition: {{blockquote|
Definition. A tensor of type (
p,
q) is an assignment of a multidimensional array :T^{i_1\dots i_p}_{j_{1}\dots j_{q}}[\mathbf{f}] to each basis of an
n-dimensional vector space such that, if we apply the change of basis :\mathbf{f}\mapsto \mathbf{f}\cdot R = \left( \mathbf{e}_i R^i_1, \dots, \mathbf{e}_i R^i_n \right) then the multidimensional array obeys the transformation law : T^{i'_1\dots i'_p}_{j'_1\dots j'_q}[\mathbf{f} \cdot R] = \left(R^{-1}\right)^{i'_1}_{i_1} \cdots \left(R^{-1}\right)^{i'_p}_{i_p} T^{i_1, \ldots, i_p}_{j_1, \ldots, j_q}[\mathbf{f}] R^{j_1}_{j'_1}\cdots R^{j_q}_{j'_q} . }} The definition of a tensor as a multidimensional array satisfying a transformation law traces back to the work of Ricci. and readily generalizes to other groups. In this approach, a type tensor
T is defined as a
multilinear map, : T: \underbrace{V^* \times\dots\times V^*}_{p \text{ copies}} \times \underbrace{ V \times\dots\times V}_{q \text{ copies}} \rightarrow \mathbb{R}, where
V∗ is the corresponding
dual space of covectors, which is linear in each of its arguments. The above assumes
V is a vector space over the
real numbers, . More generally,
V can be taken over any
field F (e.g. the
complex numbers), with
F replacing as the codomain of the multilinear maps. By applying a multilinear map
T of type to a basis {
ej} for
V and a canonical cobasis {
εi} for
V∗, :T^{i_1\dots i_p}_{j_1\dots j_q} \equiv T\left(\boldsymbol{\varepsilon}^{i_1}, \ldots,\boldsymbol{\varepsilon}^{i_p}, \mathbf{e}_{j_1}, \ldots, \mathbf{e}_{j_q}\right), a -dimensional array of components can be obtained. A different choice of basis will yield different components. But, because
T is linear in all of its arguments, the components satisfy the tensor transformation law used in the multilinear array definition. The multidimensional array of components of
T thus form a tensor according to that definition. Moreover, such an array can be realized as the components of some multilinear map
T. This motivates viewing multilinear maps as the intrinsic objects underlying tensors. In viewing a tensor as a multilinear map, it is conventional to identify the
double dual V∗∗ of the vector space
V, i.e., the space of linear functionals on the dual vector space
V∗, with the vector space
V. There is always a
natural linear map from
V to its double dual, given by evaluating a linear form in
V∗ against a vector in
V. This linear mapping is an isomorphism in finite dimensions, and it is often then expedient to identify
V with its double dual.
Using tensor products For some mathematical applications, a more abstract approach is sometimes useful. This can be achieved by defining tensors in terms of elements of
tensor products of vector spaces, which in turn are defined through a
universal property as explained
here and
here. A
type tensor is defined in this context as an element of the tensor product of vector spaces, :T \in \underbrace{V \otimes\dots\otimes V}_{p\text{ copies}} \otimes \underbrace{V^* \otimes\dots\otimes V^*}_{q \text{ copies}}. A basis of and basis of naturally induce a basis of the tensor product . The components of a tensor are the coefficients of the tensor with respect to the basis obtained from a basis {{math|{
ei}}} for and its dual basis {{math|{
ε}}}, i.e. :T = T^{i_1\dots i_p}_{j_1\dots j_q}\; \mathbf{e}_{i_1}\otimes\cdots\otimes \mathbf{e}_{i_p}\otimes \boldsymbol{\varepsilon}^{j_1}\otimes\cdots\otimes \boldsymbol{\varepsilon}^{j_q}. Using the properties of the tensor product, it can be shown that these components satisfy the transformation law for a type tensor. Moreover, the universal property of the tensor product gives a
one-to-one correspondence between tensors defined in this way and tensors defined as multilinear maps. This 1 to 1 correspondence can be achieved in the following way, because in the finite-dimensional case there exists a canonical isomorphism between a vector space and its double dual: :U \otimes V \cong\left(U^{* *}\right) \otimes\left(V^{* *}\right) \cong\left(U^{*} \otimes V^{*}\right)^{*} \cong \operatorname{Hom}^{2}\left(U^{*} \times V^{*} ; \mathbb{F}\right) The last line is using the universal property of the tensor product, that there is a 1 to 1 correspondence between maps from \operatorname{Hom}^{2}\left(U^{*} \times V^{*} ; \mathbb{F}\right) and \operatorname{Hom}\left(U^{*} \otimes V^{*} ; \mathbb{F}\right). Tensor products can be defined in great generality – for example,
involving arbitrary modules over a ring. In principle, one could define a "tensor" simply to be an element of any tensor product. However, the mathematics literature usually reserves the term
tensor for an element of a tensor product of any number of copies of a single vector space and its dual, as above.
Tensors in infinite dimensions This discussion of tensors so far assumes finite dimensionality of the spaces involved, where the spaces of tensors obtained by each of these constructions are
naturally isomorphic. Constructions of spaces of tensors based on the tensor product and multilinear mappings can be generalized, essentially without modification, to
vector bundles or
coherent sheaves. For infinite-dimensional vector spaces, inequivalent topologies lead to inequivalent notions of tensor, and these various isomorphisms may or may not hold depending on what exactly is meant by a tensor (see
topological tensor product). In some applications, it is the
tensor product of Hilbert spaces that is intended, whose properties are the most similar to the finite-dimensional case. A more modern view is that it is the tensors' structure as a
symmetric monoidal category that encodes their most important properties, rather than the specific models of those categories.
Tensor fields In many applications, especially in differential geometry and physics, it is natural to consider a tensor with components that are functions of the point in a space. This was the setting of Ricci's original work. In modern mathematical terminology such an object is called a
tensor field, often referred to simply as a tensor. In this context, a
coordinate basis is often chosen for the
tangent vector space. The transformation law may then be expressed in terms of
partial derivatives of the coordinate functions, :\bar{x}^i\left(x^1, \ldots, x^n\right), defining a coordinate transformation, : \hat{T}^{i'_1\dots i'_p}_{j'_1\dots j'_q}\left(\bar{x}^1, \ldots, \bar{x}^n\right) = \frac{\partial \bar{x}^{i'_1}}{\partial x^{i_1}} \cdots \frac{\partial \bar{x}^{i'_p}}{\partial x^{i_p}} \frac{\partial x^{j_1}}{\partial \bar{x}^{j'_1}} \cdots \frac{\partial x^{j_q}}{\partial \bar{x}^{j'_q}} T^{i_1\dots i_p}_{j_1\dots j_q}\left(x^1, \ldots, x^n\right). == History ==