Base (vector space)
In linear algebra , a basis is a subset of a vector space , with the help of which each vector of the space can be clearly represented as a finite linear combination . The coefficients of this linear combination are called the coordinates of the vector with respect to this base. One element of the basis is called a basis vector. If confusion with other basic terms (e.g. the shudder basis ) is to be feared, such a subset is also called the Hamel basis (after Georg Hamel ). A vector space generally has different bases, a change in the base forces a coordinate transformation . The Hamel base should not be confused with the base of a coordinate system , since these terms cannot be equated under certain conditions (e.g. curvilinear coordinates ).
Definition and basic terms
A basis of a vector space is a subset of with the following equivalent properties:
 Each element of can be represented as a linear combination of vectors and this representation is unique.
 is a minimal generating system of , so every vector from can be represented as a linear combination of ( is a linear envelope of ) and this property no longer applies if an element from is removed.
 is a maximal linearly independent subset of . So if another element is added from to , the new set is no longer linearly independent.
 is a linearly independent generating system of .
The elements of a basis are called basis vectors. If the vector space is a function space, the basis vectors are also called basis functions. A base can be described in the form using an index set , a finite base, for example, in the form . If such an index set is used, the family spelling is usually used to denote the base , i.e. H. instead .
Note that in family notation, an order relation on the index set creates an arrangement of the basis vectors; then means “orderly basis”. This is used to describe the orientation of vector spaces. An index set with an order relation makes it possible to introduce orientation classes (handedness) under the bases. Examples: countable infinite base , finite base .
The coefficients that appear in the representation of a vector as a linear combination of vectors from the base are called the coordinates of the vector with respect to . These are elements of the body on which the vector space is based (e.g. or ). Together these form a coordinate vector , which, however, lies in another vector space, the coordinate space . Caution: Since the assignment of the coordinates to their respective base vectors is crucial, the base vectors themselves must be used for indexing in the absence of a common index set.
Although bases are usually written down as sets, an “indexing” given by an index set is therefore more practical. The coordinate vectors then have the form which is coordinate space . If there is an order relation, a sequence of coordinates is also created for the coordinate vector. In the example , the coordinate vector has the form ("numbering" of the coordinates). The coordinate space is here , for real or complex vector spaces so or .
Important properties
 Every vector space has a basis. Proof of this statement is given in the Proof of Existence section .
 All bases of a vector space contain the same number of elements. This number, which can also be an infinite cardinal number , is called the dimension of the vector space.
 A subset of a Vektorraumes defining a linear mapping clearly , wherein the th standard unit vector designated.
 This mapping is exactly then
 injective if they are linearly independent;
 surjective , if they form a generating system;
 bijective , if they form a basis.
 This characterization is carried over to the more general case of modules over rings , see Basis (module) .
Examples
 In the Euclidean plane there is what is known as the standard basis . In addition, two vectors form a basis in this plane if and only if they do not point in the same (or the opposite) direction.
 The standard basis of vector space is the set of canonical unit vectors .
 The standard basis in the space of the matrices is formed by the standard matrices , in which there is exactly one entry and all other entries are.
 The zero vector space has dimension zero; its only basis is the empty set .
 The base is mostly used as the vector space . A set is a basis of over if and only if is not a real number. As vector space has a basis that cannot be specified explicitly.
 The vector space of the polynomials over a body has the basis . But there are also many other bases that are more complicated to write down, but are more practical in specific applications, for example the Legendre polynomials .
 In the vector space of the real number sequences the vectors form a linearly independent system, but no basis, because the sequence is not generated from it, for example , since a combination of infinitely many vectors is not a linear combination.
Proof of the equivalence of the definitions
The following considerations outline a proof that the four characterizing properties, which are used in this article as the definition of the term base, are equivalent. (Zorn's axiom of choice or lemma is not required for this proof.)
 If every vector can be clearly represented as a linear combination of vectors in , then in particular is a generating system (by definition). If there is not a minimal generating system, then there is a real subset that is also a generating system. Now be an element of which is not in . It can then be represented in at least two different ways as a linear combination of vectors in , namely once as a linear combination of vectors in and once as . There is a contradiction and therefore it is minimal. So (1) → (2) holds.
 Every minimal generating system must be linearly independent. Because if is not linearly independent, then there is a vector in which can be represented as a linear combination of vectors in . But then every linear combination of vectors in can also be described by a linear combination of vectors in and would not be minimal. So (2) → (4) holds.
 Every linearly independent generating system must be a maximally linearly independent set. If it were not maximally linearly independent, there would be one (which is not in ) which, together with linearly, would be independent. But can be represented as a linear combination of elements of what contradicts the linear independence. So (4) → (3) holds.
 A maximally linearly independent system is a generating system: Let be any vector. If in is included, then be of a linear combination of elements to write. But if is not contained in, then the set is a real superset of and therefore no longer linearly independent. The vectors that occur in a possible linear dependency cannot all be off , so one of them must (say ) be equal , with unequal 0. Therefore is . The uniqueness of this representation follows from the linear independence of . So (3) → (1) holds.
Proof of existence
With Zorn's lemma one can prove that every vector space must have a basis, even if one often cannot state it explicitly.
Let be a vector space. One would like to find a maximal linearly independent subset of the vector space. So it is obvious, the system of quantities
to consider that is semiordered by the relation . One can now show:
 is not empty (for example contains the empty set). If it not only consists of the zero vector , then every unit set with in and is also an element of .
 For every chain is also in .
From Zorn's lemma it follows that has a maximal element. The maximal elements of are now exactly the maximal linearly independent subsets of , i.e. the bases of . Therefore has a base, and it is also true that every linearly independent subset of is contained in a base of .
Basic supplementary set
Is a given set of linearly independent vectors and the above proof is based on
off, one obtains the statement that a maximal element contains. Since such a maximal element turns out to be a basis of , it is shown that one can add any set of linearly independent vectors to a basis of . This statement is called the basic supplementary sentence.
More statements about bases

Exchange lemma from Steinitz (after E. Steinitz ): If one basis of a vector space and another vector different from the zero vector , one of the basis vectors can be "exchanged" for, ie. that is, there is an index so that is also a base of . This statement is often used to show that all bases in a vector space consist of the same number of vectors.
 Each vector space is a free object above its base. This is a universal property of vector spaces in terms of category theory . Specifically, this means:
 A linear mapping of one vector space into another vector space is already completely determined by the images of the basis vectors.
 Any mapping of the base into the image space defines a linear mapping.
 In a dimensional vector space over a finite field with elements there are
 different bases.
Basic concepts in special vector spaces
Real and complex vector spaces usually have additional topological structures. This structure can result in a basic term that deviates from the one described here.
Basis and dual basis in threedimensional Euclidean vector space
In classical mechanics , the visual space is modeled with the threedimensional Euclidean vector space (V³, ·) , which makes it particularly relevant. Euclidean vector spaces are u. a. defined by the fact that there is a scalar product "·" in them , which gives these vector spaces special and noteworthy properties.
In the threedimensional Euclidean vector space there is exactly one dual basis for each basis , so that with the Kronecker delta δ the following applies: With an orthonormal basis , all basis vectors are normalized to length one and orthogonal in pairs . Then base and dual base match.
Each vector can now be represented as a linear combination of the basic vectors:
This is because the difference vectors from to the vectors to the right of the equal sign are zero vectors .
The threedimensional Euclidean vector space is a complete dot product space .
Hamel and Schauderbasis in scalar product spaces
When studying real or complex scalar product spaces , especially Hilbert spaces, there is another, more appropriate way of representing the elements of space. A basis consists of pairwise orthogonal unit vectors , and not only finite but also infinite sums (socalled series ) of basis vectors are permitted. Such a complete orthonormal system is never a basis in an infinitely dimensional space in the sense defined here; for a better distinction one also speaks of a shudder basis . The basic type described in this article is also called the Hamel base to distinguish it .
Auerbach bases
An Auerbach basis is a Hamel basis for a dense subspace in a normalized vector space , so that the distance of each basis vector from the product of the other vectors is equal to its norm .
Delimitation of the basic terms
 Both a Hamel base and a shudder base are linearly independent sets of vectors.
 A Hamel basis, or simply basis as described in this article, forms a generating system of vector space, i.e. That is, any vector of space can be represented as a linear combination of finitely many vectors of the Hamel base.
 In the case of a finitedimensional real or complex scalar product space, an orthonormal basis (i.e. a minimal generating system of normalized, mutually perpendicular vectors) is both Hamel and Schauder basis.
 In the case of an infinitedimensional, complete real or complex scalar product space (especially in an infinitedimensional Hilbert space) a shudder basis is never a Hamel basis and vice versa. In the infinitedimensional case, a Hamel basis often cannot even be orthonormalized.
 The Hamel basis of an infinitely dimensional, separable Hilbert space consists of an uncountable number of elements. A shudder base, however, in this case consists of a countable number of elements. There is therefore no Hilbert space of Hamel dimension .
 In Hilbert spaces, the basis (without addition) usually means a shudder basis , in vector spaces without a scalar product always a Hamel basis.
See also
literature
 Peter Knabner , Wolf Barth : Lineare Algebra. Basics and Applications. Springer Spectrum, Berlin / Heidelberg 2013, ISBN 9783642321856 .
 Uwe Storch, Hartmut Wiebe: Textbook of Mathematics. Volume II: Linear Algebra. BI science, Mannheim a. a. 1990, ISBN 9783411141012 .
Web links
Individual evidence
 ↑ In association theory , Steinitz's exchange lemma is placed within the framework of modular associations . It can be shown that it can be subsumed under KuroschOre's theorem . In this context, one speaks of the exchange rate in modular associations . (See: Helmuth Gericke : Theory of Associations. , 2nd edition. Bibliographisches Institut, Mannheim 1967, pp. 143–146)