A set of vectors is linearly independent if none of them can be built from the others. This is the formal way of saying that every vector in the set contributes something genuinely new — remove any one of them and the span shrinks. Independence is one half of what makes a basis: the half that eliminates waste.
Definition
A set of vectors {v1,v2,…,vk} in a vector space V is linearly independent if the equation
c1v1+c2v2+⋯+ckvk=0
has only the trivial solution c1=c2=⋯=ck=0. If a nontrivial solution exists — any assignment of scalars, not all zero, that produces 0 — the set is linearly dependent.
Dependence means at least one vector in the set can be expressed as a linear combination of the others. If cj=0 in a nontrivial relation, then dividing by cj isolates vj:
A single nonzero vector is always independent — the equation cv=0 with v=0 forces c=0. The zero vector, by contrast, is always dependent: 1⋅0=0 is a nontrivial relation. Any set containing the zero vector is therefore dependent.
Geometric Interpretation in Rⁿ
In R2 and R3, independence has clean geometric meaning.
Two vectors in R2 are dependent if and only if they are parallel — one is a scalar multiple of the other. They point along the same line through the origin, so neither adds a direction that the other does not already cover. Two non-parallel vectors are independent and span all of R2.
Three vectors in R3 are dependent if and only if they are coplanar — all three lie in a single plane through the origin. The third vector can be written as a combination of the first two, so it provides no new reach. Three non-coplanar vectors are independent and span all of R3.
In Rn, at most n vectors can be independent. A set of n+1 or more vectors in Rn is automatically dependent, regardless of what the vectors are. This is because Rn has dimensionn, and no independent set can exceed the dimension.
Testing Independence: The Homogeneous System
For vectors in Rm, independence can be tested by row reduction. Arrange v1,…,vk as columns of an m×kmatrixA. The independence equation c1v1+⋯+ckvk=0 is equivalent to the homogeneous system Ac=0.
Row reduce A. The vectors are independent if and only if every column contains a pivot — that is, there are no free variables. If any column lacks a pivot, the corresponding variable is free, a nontrivial solution exists, and the set is dependent.
Worked Example: Independent Set
Test whether v1=(1,2,−1), v2=(0,1,3), v3=(2,0,1) are independent. Form the matrix and reduce:
Three pivots in three columns — no free variables. The set is independent.
Worked Example: Dependent Set
Test v1=(1,0,2), v2=(3,1,4), v3=(5,2,6), v4=(0,1,−2) in R3. The matrix is 3×4 — four columns but only three rows, so at most three pivots. At least one column must be free. The set is dependent without any computation, simply because four vectors in R3 cannot be independent.
Testing Independence: The Determinant
When the number of vectors equals the dimension of the space, the independence question reduces to a single number. Arrange n vectors in Rn as columns of an n×n matrix A. The set is independent if and only if
det(A)=0
This follows from the invertibility equivalence: A is invertible if and only if its columns are independent, and A is invertible if and only if det(A)=0.
Worked Example
Test whether (1,3) and (2,5) are independent in R2:
det(1325)=5−6=−1=0
The vectors are independent.
The determinant test applies only to the square case — exactly n vectors in Rn. For fewer than n vectors, or for vectors in an abstract vector space, the row-reduction approach or the definition itself must be used.
Properties of Independent Sets
Several structural facts constrain how independence behaves under set operations.
Every subset of an independent set is independent. If no nontrivial combination of {v1,…,vk} gives 0, then no nontrivial combination of a subset can either — fewer vectors means fewer coefficients, all of which must still be zero.
Adding a vector w to an independent set {v1,…,vk} preserves independence if and only if w∈/Span{v1,…,vk}. If w is already in the span, it can be written as a combination of the existing vectors, creating a dependence relation. If it is outside the span, no such relation exists.
In an n-dimensional space, any independent set has at most n elements. An independent set with exactly n elements is automatically a basis — the spanning condition comes for free once the count reaches the dimension. An independent set with fewer than n elements can always be extended to a basis by adding more vectors.
Removing a vector from a dependent set may or may not restore independence. It depends on which vector is removed and which vectors participate in the dependence relation.
Dependence Relations
When a set is dependent, the nontrivial solution c1v1+⋯+ckvk=0 is called a dependence relation. It identifies which vectors participate in the redundancy: those with nonzero coefficients.
A common misconception is that dependence means every vector in the set is a combination of the others. This is false. Consider {(1,0),(0,1),(2,3)} in R2. The set is dependent because (2,3)=2(1,0)+3(0,1), but neither (1,0) nor (0,1) is a combination of the other two alone — removing either of the first two would make the remaining pair independent.
The dependence relation 2(1,0)+3(0,1)+(−1)(2,3)=(0,0) has all three coefficients nonzero, so all three vectors participate. But the vector that can be "removed" without shrinking the span is (2,3), because it is the one that lies in the span of the other two — not because it has the largest coefficient.
When the null space of the column matrix has dimension greater than 1, there are multiple linearly independent dependence relations, and different vectors can be expressed in terms of different subsets. The pivot/free column structure from row reduction clarifies which vectors are redundant.
The Wronskian Test for Functions
In function spaces, the column-matrix test does not apply directly because the vectors are functions, not finite tuples. The Wronskian provides an alternative.
Given n functions f1,…,fn, each differentiable at least n−1 times, the Wronskian is the determinant of the matrix whose rows are successive derivatives:
If W(f1,…,fn)(x0)=0 at any single point x0, the functions are linearly independent.
The converse requires caution. A Wronskian that vanishes identically does not always imply dependence unless the functions are known to be solutions of a single linear ordinary differential equation with continuous coefficients. Without that structural guarantee, counterexamples exist where the Wronskian is zero everywhere yet the functions are independent.
Independence in Abstract Vector Spaces
The definition of independence — the only combination giving 0 is the trivial one — applies in any vector space, not just Rn.
In the polynomial space Pn, the set {1,x,x2,…,xn} is independent. The equation c0+c1x+c2x2+⋯+cnxn=0 for all x forces every coefficient to be zero, because a nonzero polynomial of degree at most n can have at most n roots, but this equation must hold for every real number. Since there are n+1 vectors and dim(Pn)=n+1, this independent set is a basis.
are independent. The equation c1E11+c2E12+c3E21+c4E22=O gives (c1c3c2c4)=(0000), which forces all four scalars to zero.
In abstract spaces where vectors are not columns of numbers, the row-reduction shortcut is unavailable. Verification returns to the definition: write down the combination equal to 0 and show that all coefficients must vanish.