In linear algebra, a coordinate vector is a representation of a vector as an ordered list of numbers that describes the vector in terms of a particular ordered basis. Coordinates are always specified relative to an ordered basis. Bases and their associated coordinate representations let one realize vector spaces and linear transformations concretely as column vectors, row vectors, and matrices, hence are useful in calculations.
The idea of a coordinate vector can also be used for infinite-dimensional vector spaces, as addressed below.
This is also called the representation of v with respect of B, or the B representation of v. The α-s are called the coordinates of v. The order of the basis becomes important here, since it determines the order in which the coefficients are listed in the coordinate vector.
The standard representation
We can mechanize the above transformation by defining a function , called the standard representation of V with respect to B, that takes every vector to its coordinate representation: . Then is a linear transformation from V to Fn. In fact, it is an isomorphism, and its inverse is simply
Alternatively, we could have defined to be the above function from the beginning, realized that is an isomorphism, and defined to be its inverse.
Let P4 be the space of all the algebraic polynomials in degree less than 4 (i.e. the highest exponent of x can be 3). This space is linear and spanned by the following polynomials:
then the corresponding coordinate vector to the polynomial
- is .
Basis transformation matrix
Let B and C be two different bases of a vector space V, and let us mark with the matrix which has columns consisting of the C representation of basis vectors b1, b2, ..., bn:
This matrix is referred to as the basis transformation matrix from B to C, and can be used for transforming any vector v from a B representation to a C representation, according to the following theorem:
If E is the standard basis, the transformation from B to E can be represented with the following simplified notation:
The matrix M is an invertible matrix and M−1 is the basis transformation matrix from C to B. In other words,
- The basis transformation matrix can be regarded as an automorphism over V.
- In order to easily remember the theorem
- notice that M 's superscript and v 's subscript indices are "canceling" each other and M 's subscript becomes v 's new subscript. This "canceling" of indices is not a real canceling but rather a convenient and intuitively appealing, although mathematically incorrect, manipulation of symbols, permitted by an appropriately chosen notation.
Infinite-dimensional vector spaces
Suppose V is an infinite-dimensional vector space over a field F. If the dimension is κ, then there is some basis of κ elements for V. After an order is chosen, the basis can be considered an ordered basis. The elements of V are finite linear combinations of elements in the basis, which give rise to unique coordinate representations exactly as described before. The only change is that the indexing set for the coordinates is not finite. Since a given vector v is a finite linear combination of basis elements, the only nonzero entries of the coordinate vector for v will be the nonzero coefficients of the linear combination representing v. Thus the coordinate vector for v is zero except in finitely many entries.
The linear transformations between (possibly) infinite-dimensional vector spaces can be modeled, analogously to the finite-dimensional case, with infinite matrices. The special case of the transformations from V into V is described in the full linear ring article.