00.03.01 · precalc / equations-lines

Linear equations and the line

shipped3 tiersLean: partial

Anchor (Master): Descartes 1637 La Géométrie; Cramer 1750 Introduction à l'analyse des lignes courbes algébriques; Klein 1872 Erlangen Programm; Lang Linear Algebra Ch. I; Audin Geometry Ch. I

Intuition [Beginner]

A linear equation in two variables and is an equation that can be written in the form , where , , are numbers and , are not both zero. The set of points in the plane that satisfy the equation is a straight line. Every line in the plane is the solution set of some linear equation, and the correspondence runs both ways.

The most familiar form is slope-intercept form, . The number is the slope — how steeply the line rises as you move to the right — measured as rise over run. The number is the y-intercept, the height at which the line crosses the vertical axis. Once you know and , you can draw the line: mark on the vertical axis, then step right by and up by to reach the next point.

Two lines in the plane behave in one of three ways. They cross at exactly one point if their slopes differ. They never meet if they have the same slope but different intercepts (parallel lines). They lie on top of each other if they have the same slope and the same intercept (the same line). The slope is the single number that decides which of the three holds.

Visual [Beginner]

Picture two lines drawn on a coordinate grid. The first is , rising gently from the left, crossing the vertical axis at height , and continuing upward to the right. The second is , descending from the upper left, crossing the first line at the point , and continuing down past the horizontal axis. The intersection point is marked with a dot, and both axes are labelled.

Two straight lines on a coordinate grid, one rising and one falling, meeting at a single marked intersection point with both axes labelled.

The picture records the central fact. Each line is the solution set of one linear equation, and the intersection of the two lines is the simultaneous solution of the two equations. When two lines have different slopes, they meet at exactly one point. When the slopes match, they either never meet (parallel) or coincide everywhere (the same line). Reading the slope off the picture is the first thing to do when reasoning about a system of two linear equations.

Worked example [Beginner]

Take the equation and put it into slope-intercept form. Subtract from both sides to get , then divide by to get . The slope is and the y-intercept is . The line crosses the vertical axis at . To find where it crosses the horizontal axis, set : , so . The line passes through and .

Now solve the two-equation system and . Add the two equations side by side: the terms cancel, leaving , so . Substitute back into the second equation: , so . The unique intersection point is . Check by plugging into the first equation: , correct.

What this tells us: a linear equation in two variables is the same algebraic object as a line in the plane, and a system of two such equations asks for the meeting point of two lines. The arithmetic of solving the system matches the geometry of finding where the lines cross.

Check your understanding [Beginner]

Formal definition [Intermediate+]

A linear equation in variables over a field is an equation of the form

with coefficients and the not all zero [Lang — Basic Mathematics Ch. 3–4]. A solution is a tuple satisfying the equation. The solution set is a hyperplane in : a codimension- affine subspace.

For over , the solution set of with is a line in . Two equivalent presentations describe the same line. The implicit form is the equation itself. The parametric form is , where is any solution and is a direction vector satisfying . The vector is the normal to the line, since the implicit equation expresses orthogonality to after translation by .

When the implicit form rearranges to slope-intercept form with the slope and the y-intercept. The vertical line (with ) has no slope-intercept form; its slope is undefined. Two non-vertical lines are parallel iff they share the same slope, and perpendicular iff the product of their slopes is . The product-equals- criterion follows from the fact that the direction vectors of perpendicular lines satisfy , and the slope of a line with direction vector is .

Counterexamples to common slips

  • A linear equation in two variables is not a function in general. The vertical line has no slope-intercept form because solving for in either has every as a solution or none. The implicit form covers every line, including the vertical ones.
  • "Parallel" requires distinct lines. Two identical lines and are not parallel in the geometric sense — they coincide. The classification by the determinant (below) handles both parallel and coincident lines as the case , with the relative proportionality of distinguishing them.
  • The slope-perpendicularity rule assumes both lines are non-vertical. A vertical line is perpendicular to every horizontal line (), but the product is not in any algebraic sense. The product criterion is a corollary of the vector dot-product criterion, which works uniformly for all lines.

Key theorem with proof [Intermediate+]

Theorem (classification of two-line intersections). Let and be the lines in defined by the linear equations

with for . Let

Then exactly one of the following holds.

(a) If , the system has a unique solution, given by Cramer's rule: $$ x = \frac{c_1 b_2 - c_2 b_1}{D}, \qquad y = \frac{a_1 c_2 - a_2 c_1}{D}. $$ The two lines meet in exactly one point.

(b) If and , are proportional (there exists with ), the two equations define the same line. The two lines coincide and have infinitely many common points.

(c) If and the triples are not proportional, the two lines are parallel and have no common point.

Proof. Case . The coefficient matrix has determinant , hence is invertible with inverse . Solving gives

Substituting back verifies that this pair satisfies both equations. Uniqueness is automatic from invertibility of .

Case , proportional triples. Write with . The second equation is then , equivalent to the first after dividing by . Every solution of the first is a solution of the second and vice versa, so the two lines coincide.

Case , non-proportional triples. Suppose for contradiction that the two lines share a common point . Then and . The condition means , so the vectors and are proportional in . Pick with . (This is unique because ; if then and the proportionality is read off the first coordinate, and similarly otherwise.) Evaluating: . Then , contradicting the non-proportionality hypothesis. So the two lines share no common point.

Bridge. The classification of two-line intersections builds toward the general theory of linear systems and the algebraic-geometric correspondence that organises every later use of "line" in the curriculum. First, the determinant is the simplest nonzero example of the determinant introduced in 01.01.07 as the obstruction to invertibility of a matrix; the case is the first place this obstruction appears, and Cramer's rule for the system is the prototype of Cramer's rule for systems with the -by- determinant in the denominator. Second, the matrix equation is the simplest instance of the abstract setup for a linear map between vector spaces, the central object of linear algebra in [01.01.*]: the unique-solution case corresponds to an isomorphism, the no-solution case to , and the infinitely-many-solutions case to . Third, the parametric form is the prototype of an affine subspace — a translate of a linear subspace — and the distinction between affine and linear (lines through the origin are linear, general lines are affine) is the first appearance of a distinction that runs through every later geometry unit.

Putting these together, the linear equation in two variables is the foundational object whose algebra generalises in three coordinated directions. The determinant generalises from the scalar to the alternating multilinear form, reappearing as the Jacobian determinant in the multivariable change-of-variables formula and as the obstruction to invertibility in every linear map between equal-dimension spaces. The line generalises from a one-dimensional affine subspace of to the affine flats of (codimension- flats are solution sets of independent linear equations) and then to the hyperplanes and lines of projective space . The system generalises from two equations in two unknowns to Gaussian elimination on an system, the entire theory of rank, kernel, image, and rank-nullity. The line is the foundational geometric object whose algebra is the seed of all of linear algebra, and the classification proved above is the seed of the entire structure theory of linear systems.

Exercises [Intermediate+]

Lean formalization [Intermediate+]

[object Promise]

The four named statements compile against Mathlib's Matrix infrastructure. The proofs are deferred — Mathlib supplies Matrix.det_fin_two for the two-by-two determinant, Matrix.cramer for the unique-solution formula when the determinant is a unit, and AffineSubspace.mk' for the affine-subspace structure on the solution set of a single linear equation. The human reviewer named in the frontmatter signs off on the coverage claim.

Advanced results [Master]

The classification of two-line intersections is the prototype of the structure theory of linear systems, and the generalisations follow in three coordinated directions: dimension, base field, and the affine-versus-linear distinction.

Linear versus affine. A linear subspace of is a subset closed under addition and scalar multiplication, equivalently the kernel of a linear map or the solution set of a homogeneous linear system . An affine subspace is a translate of a linear subspace, equivalently the solution set of a possibly inhomogeneous system for some [Audin — Geometry Ch. I]. A line through the origin is a one-dimensional linear subspace; a general line is a one-dimensional affine subspace. The set of affine subspaces of of a given dimension forms a homogeneous space under the affine group , the semidirect product of translations with the general linear group. The distinction between the two becomes structural in the differential-geometric setting: a tangent space at a point is a linear subspace, while a tangent affine subspace (the actual tangent line, plane, etc., physically located at the point) is its affine translate.

Hyperplanes and flats. A hyperplane in is a codimension- affine subspace, the solution set of a single linear equation with . A general flat of codimension is the intersection of hyperplanes whose normal vectors are linearly independent, equivalently the solution set of a system where is of rank . The classification of two-line intersections generalises to the Frobenius / Kronecker-Capelli theorem: a linear system is consistent iff , and the solution set, when nonempty, is an affine subspace of dimension . The unique-solution case ( invertible) recovers Cramer's rule, the no-solution case corresponds to , and the infinitely-many-solutions case corresponds to .

Cramer's rule in general. For a square system with and , the unique solution has -th component

where is the matrix obtained from by replacing the -th column with [Cramer — Introduction à l'analyse des lignes courbes algébriques]. The case of the theorem above is the instance . Cramer's rule is rarely used computationally for large (Gaussian elimination scales as while expansion of an determinant scales as ), but it remains structurally important: it expresses each component of the solution as a ratio of determinants, making the algebraic dependence on the matrix entries explicit and supplying the determinant identity behind the adjugate-inverse formula .

Projective lines. The projective line over is , where for . As a set, : the equivalence class of with is identified with , and the class of is the point at infinity. The projective line is the natural ambient space for projective geometry — Klein's 1872 Erlangen Programm placed the geometry of as the symmetry under the projective linear group , acting by Möbius transformations with [Klein — Vergleichende Betrachtungen über neuere geometrische Forschungen]. Three distinct points of form a frame: any three points map to any other three by a unique Möbius transformation, and the canonical reference frame is . The projective line over is the Riemann sphere , central to complex analysis.

Linear codes. In coding theory, a linear code over of block length and dimension is a -dimensional linear subspace . The code is the kernel of a parity-check matrix , equivalently the image of a generator matrix . The minimum-distance and rate-versus-distance trade-off of a linear code (the Singleton bound , achieved by MDS codes such as Reed-Solomon) is the central problem of algebraic coding theory. The geometric content: a linear code is a finite-field analogue of a linear subspace of , and the codewords are the lattice points of that subspace.

Convex polytopes and half-spaces. A half-space in is the solution set of a linear inequality . A convex polytope is a bounded intersection of finitely many half-spaces. The structure theory of polytopes — vertices, edges, faces, the face lattice, the -vector — sits on top of the linear-equation foundation: each face of a polytope is the intersection of the polytope with a hyperplane that touches it at the face and bounds it on one side. The duality between vertex description (-polytope) and half-space description (-polytope), proved by Minkowski-Weyl, is the polytopal generalisation of the implicit-versus-parametric duality of a single linear equation.

Synthesis. The bridge between the elementary classification of two-line intersections and the broader theory of linear systems is the recognition that the determinant is the single load-bearing invariant — once the determinant has been identified as the obstruction to invertibility, the unique-solution case, the parallel case, and the coincident case all read off as the three possibilities for the rank of the coefficient matrix. The determinant generalises to the determinant, the classification to the Frobenius / Kronecker-Capelli theorem, the line to the hyperplane and the affine flat, and Cramer's rule to the general -by- solution formula. This is precisely the structural content the algebra strand reads off in 01.01.07: the determinant is the obstruction to invertibility, the rank measures the dimension of the image, and the solution geometry of a linear system is governed by exactly two integer invariants, and .

The same machinery extends to projective geometry, where the line is replaced by the projective subspace and the affine group by the projective linear group. Klein's Erlangen Programm organised the resulting hierarchy: affine geometry studies invariants under , projective geometry under , Euclidean under , and each lower geometry's invariants are a special case of the higher one's. The line is the foundational geometric object whose algebra is the seed of all of linear algebra, and whose generalisations are the entry points to affine geometry, projective geometry, convex geometry, and algebraic coding theory. Putting these together, the foundational insight is that a linear equation in variables is the abstract presentation of a hyperplane, the solution set of a system is the intersection of hyperplanes, and the structure theory of linear systems is the algebraic shadow of the geometry of flats in affine and projective space. This unit identifies the line as the prototype hyperplane, identifies the determinant as the obstruction to unique intersection, identifies the parametric form as the affine version of a linear subspace, and identifies the projective line as the natural completion of the affine line.

Full proof set [Master]

Proposition (Cramer's rule, case). Let with and . The unique solution of has components , where is with its -th column replaced by .

Proof. Since , is invertible with , where the entry of is the cofactor of . Then has -th component , where is the minor obtained by deleting row and column of . The numerator is the cofactor expansion of along the -th column (which contains the entries ). Hence .

Proposition (Frobenius / Kronecker-Capelli). A linear system with , has a solution iff , where is the augmented matrix. When nonempty, the solution set is an affine subspace of of dimension .

Proof. The system has a solution iff , the column space of . The column space of is spanned by the columns of together with , so iff . When the system is consistent, the solution set is the preimage . Fix any particular solution . Then iff , that is, . So the solution set is the affine subspace . The rank-nullity theorem gives .

Proposition (transitive action of Möbius transformations). The group acts on by , and the action is sharply -transitive: for any two ordered triples and of distinct points of , there is a unique element of sending to for .

Proof sketch. Existence: it suffices to show every triple of distinct points can be sent to the canonical frame by an element of , then compose. The Möbius transformation sends , , (with appropriate limit interpretation if any ). Uniqueness: a Möbius transformation fixing three distinct points is the identity. If fixes , then fixing gives , fixing gives , fixing gives , so the matrix is a nonzero scalar multiple of the identity and represents the identity in .

Proposition (Minkowski-Weyl, statement). A subset is a bounded intersection of finitely many half-spaces (an -polytope) iff it is the convex hull of finitely many points (a -polytope).

Proof. Stated without proof; see Ziegler Lectures on Polytopes Ch. 1 for the standard development via Fourier-Motzkin elimination, the linear-programming duality theorem, and Carathéodory's theorem. The bridge from the linear-equation foundation here is that each facet (codimension- face) of a polytope corresponds to one of the defining linear inequalities turning into a tight linear equation, so the face lattice of a polytope is the combinatorial shadow of a finite collection of linear equations and inequalities.

Connections [Master]

  • The two-by-two determinant proved here as the obstruction to unique intersection is the simplest nonzero instance of the general determinant introduced in 01.01.07. Cramer's rule for the system is the prototype of Cramer's rule for the system, and the classification of two-line intersections is the prototype of the Frobenius / Kronecker-Capelli theorem governing arbitrary linear systems. The determinant reappears throughout the curriculum: as the Jacobian determinant in multivariable change of variables in 02.05.04, as the obstruction to local invertibility of a smooth map at a point, and as the top exterior power of a linear endomorphism in 01.01.10 pending.

  • The matrix equation studied here is the simplest instance of the abstract setup for a linear map between vector spaces, the central object of linear algebra in 01.01.03 (vector space) and 01.01.06 pending (linear transformation, kernel, image, rank-nullity). The three cases of the classification — unique solution, no solution, infinitely many solutions — correspond exactly to the three structural possibilities for a linear map between equal-dimension spaces: isomorphism, inconsistency (), and nontrivial kernel. The line itself is the simplest example of an affine subspace, the geometric setting where vectors are points rather than displacements.

  • The projective line and the action of by Möbius transformations introduced in the Advanced results section reappear in 06.01.07 (Riemann sphere) and 06.01.08 (Möbius transformations) as the central organising structures of complex analysis on . The same group-theoretic content — three distinct points form a sharp frame — is the basis of the cross-ratio invariant of projective geometry, and the same algebra of matrices acting on the projective line drives the spectral theory of Schwarzian derivatives and one-dimensional projective structures.

  • The affine-versus-linear distinction sharpened here recurs throughout differential geometry. A tangent space at a point of a manifold is a linear subspace of the ambient vector space, while the tangent affine subspace (the actual tangent line, plane, etc.) is its affine translate located at the point. This distinction is the source of the affine-vs-linear gap in the formulation of geodesics, in the affine connections of 03.02.01, and in the principal-bundle reformulation of gauge theory where the affine action of a structure group on a fibre is the load-bearing operation.

Historical & philosophical context [Master]

The geometric line predates the algebraic line by more than two millennia. Euclid's Elements (~300 BCE) Book I opens with the postulates of plane geometry, taking the line as a primitive object characterised by axioms: through any two distinct points there is exactly one line, any line can be extended indefinitely, and through any point not on a given line there is exactly one line parallel to it (the parallel postulate). The Euclidean line is a synthetic object with no algebraic description. René Descartes, in his 1637 La Géométrie (an appendix to the Discours de la méthode), introduced the correspondence between algebraic equations and geometric loci that organises every modern treatment of the line: a curve in the plane is the solution set of an equation , and the simplest non-degenerate case is the linear equation , whose solution set is a straight line [Descartes — La Géométrie]. Pierre de Fermat, in unpublished manuscripts circulating from 1636, independently introduced the same algebraic-geometric correspondence; his Ad locos planos et solidos isagoge (composed ~1636, published posthumously in 1679) gave the equation of a straight line in essentially modern form.

The determinant rule for solving systems of linear equations is due to Gabriel Cramer, who stated and applied it in the appendix to his 1750 Introduction à l'analyse des lignes courbes algébriques [Cramer — Introduction à l'analyse des lignes courbes algébriques]. Cramer derived the rule for systems by extrapolating from the small- cases without explicit recursive structure; the modern cofactor-expansion development of the determinant came later, with Pierre-Simon Laplace (1772), Augustin-Louis Cauchy (1812), and Carl Gustav Jacob Jacobi (1841) supplying the systematic theory. Felix Klein, in his 1872 inaugural lecture at Erlangen, organised geometry by symmetry groups: the projective line with its symmetry sits at the top of a hierarchy whose lower floors are affine geometry, similarity geometry, and Euclidean geometry, each obtained by restricting the symmetry group [Klein — Vergleichende Betrachtungen über neuere geometrische Forschungen]. The line, in each of these geometries, is the canonical one-dimensional flat: a line through the origin in linear algebra, an affine line in affine geometry, a projective line in projective geometry.

Bibliography [Master]

[object Promise]