HOME

TheInfoList



OR:

In
mathematics Mathematics is a field of study that discovers and organizes methods, Mathematical theory, theories and theorems that are developed and Mathematical proof, proved for the needs of empirical sciences and mathematics itself. There are many ar ...
, a matrix (: matrices) is a rectangular array or table of
number A number is a mathematical object used to count, measure, and label. The most basic examples are the natural numbers 1, 2, 3, 4, and so forth. Numbers can be represented in language with number words. More universally, individual numbers can ...
s,
symbol A symbol is a mark, Sign (semiotics), sign, or word that indicates, signifies, or is understood as representing an idea, physical object, object, or wikt:relationship, relationship. Symbols allow people to go beyond what is known or seen by cr ...
s, or expressions, with elements or entries arranged in rows and columns, which is used to represent a
mathematical object A mathematical object is an abstract concept arising in mathematics. Typically, a mathematical object can be a value that can be assigned to a Glossary of mathematical symbols, symbol, and therefore can be involved in formulas. Commonly encounter ...
or property of such an object. For example, \begin1 & 9 & -13 \\20 & 5 & -6 \end is a matrix with two rows and three columns. This is often referred to as a "two-by-three matrix", a " matrix", or a matrix of dimension . Matrices are commonly used in
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, where they represent
linear map In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that p ...
s. In
geometry Geometry (; ) is a branch of mathematics concerned with properties of space such as the distance, shape, size, and relative position of figures. Geometry is, along with arithmetic, one of the oldest branches of mathematics. A mathematician w ...
, matrices are widely used for specifying and representing geometric transformations (for example
rotation Rotation or rotational/rotary motion is the circular movement of an object around a central line, known as an ''axis of rotation''. A plane figure can rotate in either a clockwise or counterclockwise sense around a perpendicular axis intersect ...
s) and coordinate changes. In
numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic computation, symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of ...
, many computational problems are solved by reducing them to a matrix computation, and this often involves computing with matrices of huge dimensions. Matrices are used in most areas of mathematics and scientific fields, either directly, or through their use in geometry and numerical analysis. '' Square matrices'', matrices with the same number of rows and columns, play a major role in matrix theory. The
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of a square matrix is a number associated with the matrix, which is fundamental for the study of a square matrix; for example, a square matrix is
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
if and only if it has a nonzero determinant and the
eigenvalue In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
s of a square matrix are the roots of a
polynomial In mathematics, a polynomial is a Expression (mathematics), mathematical expression consisting of indeterminate (variable), indeterminates (also called variable (mathematics), variables) and coefficients, that involves only the operations of addit ...
determinant. Matrix theory is the branch of mathematics that focuses on the study of matrices. It was initially a sub-branch of
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, but soon grew to include subjects related to
graph theory In mathematics and computer science, graph theory is the study of ''graph (discrete mathematics), graphs'', which are mathematical structures used to model pairwise relations between objects. A graph in this context is made up of ''Vertex (graph ...
,
algebra Algebra is a branch of mathematics that deals with abstract systems, known as algebraic structures, and the manipulation of expressions within those systems. It is a generalization of arithmetic that introduces variables and algebraic ope ...
,
combinatorics Combinatorics is an area of mathematics primarily concerned with counting, both as a means and as an end to obtaining results, and certain properties of finite structures. It is closely related to many other areas of mathematics and has many ...
and
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
.


Definition

A matrix is a rectangular array of
number A number is a mathematical object used to count, measure, and label. The most basic examples are the natural numbers 1, 2, 3, 4, and so forth. Numbers can be represented in language with number words. More universally, individual numbers can ...
s (or other mathematical objects), called the "entries" of the matrix. Matrices are subject to standard operations such as
addition Addition (usually signified by the Plus and minus signs#Plus sign, plus symbol, +) is one of the four basic Operation (mathematics), operations of arithmetic, the other three being subtraction, multiplication, and Division (mathematics), divis ...
and
multiplication Multiplication is one of the four elementary mathematical operations of arithmetic, with the other ones being addition, subtraction, and division (mathematics), division. The result of a multiplication operation is called a ''Product (mathem ...
. Most commonly, a matrix over a field F is a rectangular array of elements of . A real matrix and a complex matrix are matrices whose entries are respectively
real number In mathematics, a real number is a number that can be used to measure a continuous one- dimensional quantity such as a duration or temperature. Here, ''continuous'' means that pairs of values can have arbitrarily small differences. Every re ...
s or
complex number In mathematics, a complex number is an element of a number system that extends the real numbers with a specific element denoted , called the imaginary unit and satisfying the equation i^= -1; every complex number can be expressed in the for ...
s. More general types of entries are discussed below. For instance, this is a real matrix: \mathbf = \begin -1.3 & 0.6 \\ 20.4 & 5.5 \\ 9.7 & -6.2 \end. The numbers, symbols, or expressions in the matrix are called its ''entries'' or its ''elements''. The horizontal and vertical lines of entries in a matrix are respectively called ''rows'' and ''columns''.


Size

The size of a matrix is defined by the number of rows and columns it contains. There is no limit to the number of rows and columns that a matrix (in the usual sense) can have as long as they are positive integers. A matrix with m rows and n columns is called an m \times n matrix, or -by- matrix, where and are called its ''dimensions''. For example, the matrix above is a matrix. Matrices with a single row are called '' row matrices'' or ''row vectors'', and those with a single column are called '' column matrices'' or ''column vectors''. A matrix with the same number of rows and columns is called a ''
square matrix In mathematics, a square matrix is a Matrix (mathematics), matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied. Squ ...
''. A matrix with an infinite number of rows or columns (or both) is called an '' infinite matrix''. In some contexts, such as computer algebra programs, it is useful to consider a matrix with no rows or no columns, called an '' empty matrix''.


Notation

The specifics of symbolic matrix notation vary widely, with some prevailing trends. Matrices are commonly written in square brackets or parentheses, so that an m \times n matrix \mathbf is represented as \mathbf = \begin a_ & a_ & \cdots & a_ \\ a_ & a_ & \cdots & a_ \\ \vdots & \vdots & \ddots & \vdots \\ a_ & a_ & \cdots & a_ \end = \begin a_ & a_ & \cdots & a_ \\ a_ & a_ & \cdots & a_ \\ \vdots & \vdots & \ddots & \vdots \\ a_ & a_ & \cdots & a_ \end. This may be abbreviated by writing only a single generic term, possibly along with indices, as in \mathbf = \left(a_\right), \quad \left a_\right \quad \text \quad \left(a_\right)_ or \mathbf=(a_)_ in the case that . Matrices are usually symbolized using upper-case letters (such as in the examples above), while the corresponding lower-case letters, with two subscript indices (e.g., , or ), represent the entries. In addition to using upper-case letters to symbolize matrices, many authors use a special typographical style, commonly boldface Roman (non-italic), to further distinguish matrices from other mathematical objects. An alternative notation involves the use of a double-underline with the variable name, with or without boldface style, as in . The entry in the th row and th column of a matrix is sometimes referred to as the or (i,j) entry of the matrix, and commonly denoted by a_ or . Alternative notations for that entry are and . For example, the (1,3) entry of the following matrix \mathbf is (also denoted , , \mathbf ,3/math> or ): \mathbf=\begin 4 & -7 & \color & 0 \\ -2 & 0 & 11 & 8 \\ 19 & 1 & -3 & 12 \end Sometimes, the entries of a matrix can be defined by a formula such as . For example, each of the entries of the following matrix \mathbf is determined by the formula . \mathbf A = \begin 0 & -1 & -2 & -3\\ 1 & 0 & -1 & -2\\ 2 & 1 & 0 & -1 \end In this case, the matrix itself is sometimes defined by that formula, within square brackets or double parentheses. For example, the matrix above is defined as = -j/math> or . If matrix size is , the above-mentioned formula f(i,j) is valid for any i=1,\dots,m and any . This can be specified separately or indicated using m\times n as a subscript. For instance, the matrix \mathbf above is , and can be defined as = -j(i=1,2,3; j=1,\dots,4) or . Some programming languages utilize doubly subscripted arrays (or arrays of arrays) to represent an -by- matrix. Some programming languages start the numbering of array indexes at zero, in which case the entries of an -by- matrix are indexed by 0\leq i \leq m-1 and . This article follows the more common convention in mathematical writing where enumeration starts from . The
set Set, The Set, SET or SETS may refer to: Science, technology, and mathematics Mathematics *Set (mathematics), a collection of elements *Category of sets, the category whose objects and morphisms are sets and total functions, respectively Electro ...
of all -by- real matrices is often denoted , or . The set of all -by- matrices over another field, or over a ring , is similarly denoted , or . If , such as in the case of square matrices, one does not repeat the dimension: , or . Often, , or , is used in place of .


Basic operations

Several basic operations can be applied to matrices. Some, such as ''transposition'' and ''submatrix'' do not depend on the nature of the entries. Others, such as ''matrix addition'', ''scalar multiplication'', ''matrix multiplication'', and ''row operations'' involve operations on matrix entries and therefore require that matrix entries are numbers or belong to a field or a ring. In this section, it is supposed that matrix entries belong to a fixed ring, which is typically a field of numbers.


Addition

Matrix addition and subtraction require matrices of a consistent size, and are calculated entrywise. The ''sum'' and the difference of two matrices are: \begin (\bold A + \bold B)_ = \bold A_ + \bold B_, \quad 1 \leq i \leq m, \quad 1 \leq j \leq n. \\ (\bold A - \bold B)_ = \bold A_ - \bold B_, \quad 1 \leq i \leq m, \quad 1 \leq j \leq n. \end For example, \begin 1 & 3 & 1 \\ 1 & 0 & 0 \end + \begin 0 & 0 & 5 \\ 7 & 5 & 0 \end = \begin 1+0 & 3+0 & 1+5 \\ 1+7 & 0+5 & 0+0 \end = \begin 1 & 3 & 6 \\ 8 & 5 & 0 \end Familiar properties of numbers extend to these operations on matrices: for example, addition is
commutative In mathematics, a binary operation is commutative if changing the order of the operands does not change the result. It is a fundamental property of many binary operations, and many mathematical proofs depend on it. Perhaps most familiar as a pr ...
, that is, the matrix sum does not depend on the order of the summands: .


Scalar multiplication

The product of a number (also called a scalar in this context) and a matrix is computed by multiplying each entry of by : (c\bold A)_ = c \cdot \bold A_ This operation is called ''scalar multiplication'', but its result is not named "scalar product" to avoid confusion, since "scalar product" is often used as a synonym for "
inner product In mathematics, an inner product space (or, rarely, a Hausdorff pre-Hilbert space) is a real vector space or a complex vector space with an operation called an inner product. The inner product of two vectors in the space is a scalar, ofte ...
". For example: 2 \cdot \begin 1 & 8 & -3 \\ 4 & -2 & 5 \end = \begin 2 \cdot 1 & 2\cdot 8 & 2\cdot -3 \\ 2\cdot 4 & 2\cdot -2 & 2\cdot 5 \end = \begin 2 & 16 & -6 \\ 8 & -4 & 10 \end Matrix subtraction is consistent with composition of matrix addition with scalar multiplication by : \mathbf A - \mathbf B = \mathbf A +(-1)\cdot \mathbf B


Transpose

The ''transpose'' of an matrix is the matrix (also denoted or ) formed by turning rows into columns and vice versa: \left( \bold A^ \right)_ = \bold A_. For example: \begin 1 & 2 & 3 \\ 0 & -6 & 7 \end^\mathrm = \begin 1 & 0 \\ 2 & -6 \\ 3 & 7 \end The transpose is compatible with addition and scalar multiplication, as expressed by and . Finally, .


Matrix multiplication

''Multiplication'' of two matrices corresponds to the composition of linear transformations represented by each matrix. It is defined if and only if the number of columns of the left matrix is the same as the number of rows of the right matrix. If is an matrix and is an matrix, then their ''matrix product'' is the matrix whose entries are given by the
dot product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
of the corresponding row of and the corresponding column of : mathbf = a_b_ + a_b_ + \cdots + a_b_ = \sum_^n a_b_, where and . For example, the underlined entry 2340 in the product is calculated as \begin \begin \underline & \underline 3 & \underline 4 \\ 1 & 0 & 0 \\ \end \begin 0 & \underline \\ 1 & \underline \\ 0 & \underline \\ \end &= \begin 3 & \underline \\ 0 & 1000 \\ \end. \end Matrix multiplication satisfies the rules (
associativity In mathematics, the associative property is a property of some binary operations that rearranging the parentheses in an expression will not change the result. In propositional logic, associativity is a Validity (logic), valid rule of replaceme ...
), and as well as (left and right
distributivity In mathematics, the distributive property of binary operations is a generalization of the distributive law, which asserts that the equality x \cdot (y + z) = x \cdot y + x \cdot z is always true in elementary algebra. For example, in elementary ...
), whenever the size of the matrices is such that the various products are defined. The product may be defined without being defined, namely if and are and matrices, respectively, and Even if both products are defined, they generally need not be equal, that is: \bold \ne \bold. In other words, matrix multiplication is not
commutative In mathematics, a binary operation is commutative if changing the order of the operands does not change the result. It is a fundamental property of many binary operations, and many mathematical proofs depend on it. Perhaps most familiar as a pr ...
,
in marked contrast to (rational, real, or complex) numbers, whose product is independent of the order of the factors. An example of two matrices not commuting with each other is: \begin 1 & 2\\ 3 & 4\\ \end \begin 0 & 1\\ 0 & 0\\ \end= \begin 0 & 1\\ 0 & 3\\ \end, whereas \begin 0 & 1\\ 0 & 0\\ \end \begin 1 & 2\\ 3 & 4\\ \end= \begin 3 & 4\\ 0 & 0\\ \end. Besides the ordinary matrix multiplication just described, other less frequently used operations on matrices that can be considered forms of multiplication also exist, such as the Hadamard product and the
Kronecker product In mathematics, the Kronecker product, sometimes denoted by ⊗, is an operation on two matrices of arbitrary size resulting in a block matrix. It is a specialization of the tensor product (which is denoted by the same symbol) from vector ...
. They arise in solving matrix equations such as the Sylvester equation.


Row operations

There are three types of row operations: # row addition, that is, adding a row to another. # row multiplication, that is, multiplying all entries of a row by a non-zero constant; # row switching, that is, interchanging two rows of a matrix; These operations are used in several ways, including solving
linear equation In mathematics, a linear equation is an equation that may be put in the form a_1x_1+\ldots+a_nx_n+b=0, where x_1,\ldots,x_n are the variables (or unknowns), and b,a_1,\ldots,a_n are the coefficients, which are often real numbers. The coeffici ...
s and finding
matrix inverse In linear algebra, an invertible matrix (''non-singular'', ''non-degenarate'' or ''regular'') is a square matrix that has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by an ...
s with Gauss elimination and Gauss–Jordan elimination, respectively.


Submatrix

A submatrix of a matrix is a matrix obtained by deleting any collection of rows and/or columns. For example, from the following 3-by-4 matrix, we can construct a 2-by-3 submatrix by removing row 3 and column 2: \mathbf=\begin 1 & \color & 3 & 4 \\ 5 & \color & 7 & 8 \\ \color & \color & \color & \color \end \rightarrow \begin 1 & 3 & 4 \\ 5 & 7 & 8 \end. The minors and cofactors of a matrix are found by computing the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of certain submatrices. A principal submatrix is a square submatrix obtained by removing certain rows and columns. The definition varies from author to author. According to some authors, a principal submatrix is a submatrix in which the set of row indices that remain is the same as the set of column indices that remain. Other authors define a principal submatrix as one in which the first rows and columns, for some number , are the ones that remain; this type of submatrix has also been called a leading principal submatrix.


Linear equations

Matrices can be used to compactly write and work with multiple linear equations, that is, systems of linear equations. For example, if is an matrix, designates a column vector (that is, -matrix) of variables and is an -column vector, then the matrix equation \mathbf = \mathbf is equivalent to the system of linear equations \begin a_x_1 + a_x_2 + &\cdots + a_x_n = b_1 \\ &\ \ \vdots \\ a_x_1 + a_x_2 + &\cdots + a_x_n = b_m \end Using matrices, this can be solved more compactly than would be possible by writing out all the equations separately. If and the equations are independent, then this can be done by writing \mathbf = \mathbf^ \mathbf where is the inverse matrix of . If has no inverse, solutions—if any—can be found using its
generalized inverse In mathematics, and in particular, algebra, a generalized inverse (or, g-inverse) of an element ''x'' is an element ''y'' that has some properties of an inverse element but not necessarily all of them. The purpose of constructing a generalized inv ...
.


Linear transformations

Matrices and matrix multiplication reveal their essential features when related to ''linear transformations'', also known as ''linear maps''. A real -by- matrix gives rise to a linear transformation \R^n \to \R^m mapping each vector in to the (matrix) product , which is a vector in Conversely, each linear transformation f: \R^n \to \R^m arises from a unique -by- matrix : explicitly, the -entry of is the th coordinate of , where is the
unit vector In mathematics, a unit vector in a normed vector space is a Vector (mathematics and physics), vector (often a vector (geometry), spatial vector) of Norm (mathematics), length 1. A unit vector is often denoted by a lowercase letter with a circumfle ...
with in the th position and elsewhere.
The matrix is said to represent the linear map , and is called the ''transformation matrix'' of . For example, the 2×2 matrix \mathbf = \begin a & c\\b & d \end can be viewed as the transform of the
unit square In mathematics, a unit square is a square whose sides have length . Often, ''the'' unit square refers specifically to the square in the Cartesian plane with corners at the four points ), , , and . Cartesian coordinates In a Cartesian coordinat ...
into a
parallelogram In Euclidean geometry, a parallelogram is a simple polygon, simple (non-list of self-intersecting polygons, self-intersecting) quadrilateral with two pairs of Parallel (geometry), parallel sides. The opposite or facing sides of a parallelogram a ...
with vertices at , , , and . The parallelogram pictured at the right is obtained by multiplying with each of the column vectors , , , and in turn. These vectors define the vertices of the unit square. The following table shows several 2×2 real matrices with the associated linear maps of . The original is mapped to the grid and shapes. The origin is marked with a black point. Under the 1-to-1 correspondence between matrices and linear maps, matrix multiplication corresponds to
composition Composition or Compositions may refer to: Arts and literature *Composition (dance), practice and teaching of choreography * Composition (language), in literature and rhetoric, producing a work in spoken tradition and written discourse, to include ...
of maps: if a -by- matrix represents another linear map , then the composition is represented by since (g \circ f)(\bold x) = g(f(\bold x)) = g(\bold) = \bold B(\bold ) = (\bold) \bold x. The last equality follows from the above-mentioned associativity of matrix multiplication. The
rank of a matrix In linear algebra, the rank of a matrix is the dimension of the vector space generated (or spanned) by its columns. p. 48, § 1.16 This corresponds to the maximal number of linearly independent columns of . This, in turn, is identical to the dime ...
is the maximum number of
linearly independent In the theory of vector spaces, a set of vectors is said to be if there exists no nontrivial linear combination of the vectors that equals the zero vector. If such a linear combination exists, then the vectors are said to be . These concep ...
row vectors of the matrix, which is the same as the maximum number of linearly independent column vectors. Equivalently it is the
dimension In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coo ...
of the
image An image or picture is a visual representation. An image can be Two-dimensional space, two-dimensional, such as a drawing, painting, or photograph, or Three-dimensional space, three-dimensional, such as a carving or sculpture. Images may be di ...
of the linear map represented by . The rank–nullity theorem states that the dimension of the kernel of a matrix plus the rank equals the number of columns of the matrix.


Square matrix

A
square matrix In mathematics, a square matrix is a Matrix (mathematics), matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied. Squ ...
is a matrix with the same number of rows and columns. An -by- matrix is known as a square matrix of order . Any two square matrices of the same order can be added and multiplied. The entries form the main diagonal of a square matrix. They lie on the imaginary line running from the top left corner to the bottom right corner of the matrix. Square matrices of a given dimension form a
noncommutative ring In mathematics, a noncommutative ring is a ring whose multiplication is not commutative; that is, there exist ''a'' and ''b'' in the ring such that ''ab'' and ''ba'' are different. Equivalently, a ''noncommutative ring'' is a ring that is not ...
, which is one of the most common examples of a noncommutative ring.


Main types

:


Diagonal and triangular matrix

If all entries of below the main diagonal are zero, is called an ''upper
triangular matrix In mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called if all the entries ''above'' the main diagonal are zero. Similarly, a square matrix is called if all the entries ''below'' the main diagonal are z ...
''. Similarly, if all entries of above the main diagonal are zero, is called a ''lower triangular matrix''. If all entries outside the main diagonal are zero, is called a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
.


Identity matrix

The ''identity matrix'' of size is the -by- matrix in which all the elements on the main diagonal are equal to and all other elements are equal to , for example, \begin \mathbf_1 &= \begin 1 \end, \\ pt \mathbf_2 &= \begin 1 & 0 \\ 0 & 1 \end, \\ pt \vdots& \\ pt \mathbf_n &= \begin 1 & 0 & \cdots & 0 \\ 0 & 1 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & 1 \end \end It is a square matrix of order , and also a special kind of
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
. It is called an identity matrix because multiplication with it leaves a matrix unchanged: \bold_n = \bold I_m \bold A = \bold A for any -by- matrix . A scalar multiple of an identity matrix is called a ''scalar'' matrix.


Symmetric or skew-symmetric matrix

A square matrix that is equal to its transpose, that is, , is a
symmetric matrix In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, Because equal matrices have equal dimensions, only square matrices can be symmetric. The entries of a symmetric matrix are symmetric with ...
. If instead, is equal to the negative of its transpose, that is, , then is a
skew-symmetric matrix In mathematics, particularly in linear algebra, a skew-symmetric (or antisymmetric or antimetric) matrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition In terms of the entries of the matrix, if a ...
. In complex matrices, symmetry is often replaced by the concept of Hermitian matrices, which satisfies , where the star or
asterisk The asterisk ( ), from Late Latin , from Ancient Greek , , "little star", is a Typography, typographical symbol. It is so called because it resembles a conventional image of a star (heraldry), heraldic star. Computer scientists and Mathematici ...
denotes the conjugate transpose of the matrix, that is, the transpose of the
complex conjugate In mathematics, the complex conjugate of a complex number is the number with an equal real part and an imaginary part equal in magnitude but opposite in sign. That is, if a and b are real numbers, then the complex conjugate of a + bi is a - ...
of . By the
spectral theorem In linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful because computations involvin ...
, real symmetric matrices and complex Hermitian matrices have an eigenbasis; that is, every vector is expressible as a
linear combination In mathematics, a linear combination or superposition is an Expression (mathematics), expression constructed from a Set (mathematics), set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' a ...
of eigenvectors. In both cases, all eigenvalues are real. This theorem can be generalized to infinite-dimensional situations related to matrices with infinitely many rows and columns.


Invertible matrix and its inverse

A square matrix is called ''
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
'' or ''non-singular'' if there exists a matrix such that \bold = \bold = \bold I_n, where is the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
with for each entry on the main diagonal and elsewhere. If exists, it is unique and is called the '' inverse matrix'' of , denoted . There are many
algorithm In mathematics and computer science, an algorithm () is a finite sequence of Rigour#Mathematics, mathematically rigorous instructions, typically used to solve a class of specific Computational problem, problems or to perform a computation. Algo ...
s for testing whether a square matrix is invertible, and, if it is, computing its inverse. One of the oldest, which is still in common use is
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
.


Definite matrix

A symmetric real matrix is called ''positive-definite'' if the associated
quadratic form In mathematics, a quadratic form is a polynomial with terms all of degree two (" form" is another name for a homogeneous polynomial). For example, 4x^2 + 2xy - 3y^2 is a quadratic form in the variables and . The coefficients usually belong t ...
f(\bold x) = \bold x^ \bold has a positive value for every nonzero vector in . If yields only negative values then is ''negative-definite''; if does produce both negative and positive values then is '' indefinite''. If the quadratic form yields only non-negative values (positive or zero), the symmetric matrix is called ''positive-semidefinite'' (or if only non-positive values, then negative-semidefinite); hence the matrix is indefinite precisely when it is neither positive-semidefinite nor negative-semidefinite. A symmetric matrix is positive-definite if and only if all its eigenvalues are positive, that is, the matrix is positive-semidefinite and it is invertible. The table at the right shows two possibilities for 2-by-2 matrices. The eigenvalues of a diagonal matrix are simply the entries along the diagonal, and so in these examples, the eigenvalues can be read directly from the matrices themselves. The first matrix has two eigenvalues that are both positive, while the second has one that is positive and another that is negative. Allowing as input two different vectors instead yields the
bilinear form In mathematics, a bilinear form is a bilinear map on a vector space (the elements of which are called '' vectors'') over a field ''K'' (the elements of which are called '' scalars''). In other words, a bilinear form is a function that is linea ...
associated to : B_ (\bold x, \bold y) = \bold x^ \bold. In the case of complex matrices, the same terminology and results apply, with ''symmetric matrix'', ''quadratic form'', ''bilinear form'', and ''transpose'' replaced respectively by
Hermitian matrix In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the ...
,
Hermitian form In mathematics, a sesquilinear form is a generalization of a bilinear form that, in turn, is a generalization of the concept of the dot product of Euclidean space. A bilinear form is linear map, linear in each of its arguments, but a sesquilinear f ...
, sesquilinear form, and conjugate transpose .


Orthogonal matrix

An ''orthogonal matrix'' is a square matrix with real entries whose columns and rows are
orthogonal In mathematics, orthogonality (mathematics), orthogonality is the generalization of the geometric notion of ''perpendicularity''. Although many authors use the two terms ''perpendicular'' and ''orthogonal'' interchangeably, the term ''perpendic ...
unit vector In mathematics, a unit vector in a normed vector space is a Vector (mathematics and physics), vector (often a vector (geometry), spatial vector) of Norm (mathematics), length 1. A unit vector is often denoted by a lowercase letter with a circumfle ...
s (that is, orthonormal vectors). Equivalently, a matrix is orthogonal if its
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
is equal to its inverse: \mathbf^\mathrm=\mathbf^, \, which entails \mathbf^\mathrm \mathbf = \mathbf \mathbf^\mathrm = \mathbf_n, where is the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
of size . An orthogonal matrix is necessarily
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
(with inverse ), unitary (), and normal (). The
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of any orthogonal matrix is either or . A ''special orthogonal matrix'' is an orthogonal matrix with
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
. As a
linear transformation In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pr ...
, every orthogonal matrix with determinant is a pure
rotation Rotation or rotational/rotary motion is the circular movement of an object around a central line, known as an ''axis of rotation''. A plane figure can rotate in either a clockwise or counterclockwise sense around a perpendicular axis intersect ...
without reflection, i.e., the transformation preserves the orientation of the transformed structure, while every orthogonal matrix with determinant reverses the orientation, i.e., is a composition of a pure reflection and a (possibly null) rotation. The identity matrices have determinant and are pure rotations by an angle zero. The
complex Complex commonly refers to: * Complexity, the behaviour of a system whose components interact in multiple ways so possible interactions are difficult to describe ** Complex system, a system composed of many components which may interact with each ...
analog of an orthogonal matrix is a unitary matrix.


Main operations


Trace

The trace, of a square matrix is the sum of its diagonal entries. While matrix multiplication is not commutative as mentioned above, the trace of the product of two matrices is independent of the order of the factors: \operatorname(\mathbf) =\operatorname(\mathbf). This is immediate from the definition of matrix multiplication: \operatorname(\mathbf) = \sum_^m \sum_^n a_ b_ = \operatorname(\mathbf). It follows that the trace of the product of more than two matrices is independent of
cyclic permutation In mathematics, and in particular in group theory, a cyclic permutation is a permutation consisting of a single cycle. In some cases, cyclic permutations are referred to as cycles; if a cyclic permutation has ''k'' elements, it may be called a ''k ...
s of the matrices; however, this does not in general apply for arbitrary permutations. For example, , in general. Also, the trace of a matrix is equal to that of its transpose, that is, \operatorname(\bold A) = \operatorname(\bold A^).


Determinant

The ''determinant'' of a square matrix (denoted or ) is a number encoding certain properties of the matrix. A matrix is invertible
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
its determinant is nonzero. Its
absolute value In mathematics, the absolute value or modulus of a real number x, is the non-negative value without regard to its sign. Namely, , x, =x if x is a positive number, and , x, =-x if x is negative (in which case negating x makes -x positive), ...
equals the area (in ) or volume (in ) of the image of the unit square (or cube), while its sign corresponds to the orientation of the corresponding linear map: the determinant is positive if and only if the orientation is preserved. The determinant of 2-by-2 matrices is given by \det \begina&b\\c&d\end = ad-bc. The determinant of 3-by-3 matrices involves 6 terms ( rule of Sarrus). The more lengthy Leibniz formula generalizes these two formulae to all dimensions. The determinant of a product of square matrices equals the product of their determinants: \det(\bold) = \det(\bold A) \cdot \det(\bold B), or using alternate notation: , \bold, = , \bold A, \cdot , \bold B, . Adding a multiple of any row to another row, or a multiple of any column to another column, does not change the determinant. Interchanging two rows or two columns affects the determinant by multiplying it by . Using these operations, any matrix can be transformed to a lower (or upper) triangular matrix, and for such matrices, the determinant equals the product of the entries on the main diagonal; this provides a method to calculate the determinant of any matrix. Finally, the Laplace expansion expresses the determinant in terms of minors, that is, determinants of smaller matrices. This expansion can be used for a recursive definition of determinants (taking as starting case the determinant of a 1-by-1 matrix, which is its unique entry, or even the determinant of a 0-by-0 matrix, which is ), that can be seen to be equivalent to the Leibniz formula. Determinants can be used to solve linear systems using
Cramer's rule In linear algebra, Cramer's rule is an explicit formula for the solution of a system of linear equations with as many equations as unknowns, valid whenever the system has a unique solution. It expresses the solution in terms of the determinants of ...
, where the division of the determinants of two related square matrices equates to the value of each of the system's variables.


Eigenvalues and eigenvectors

A number \lambda and a nonzero vector satisfying \mathbf \mathbf = \lambda \mathbf are called an ''eigenvalue'' and an ''eigenvector'' of , respectively. The number is an eigenvalue of an -matrix if and only if is not invertible, which is equivalent to \det(\mathbf-\lambda \mathbf) = 0. The polynomial in an indeterminate given by evaluation of the determinant is called the
characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
of . It is a
monic polynomial In algebra, a monic polynomial is a non-zero univariate polynomial (that is, a polynomial in a single variable) in which the leading coefficient (the nonzero coefficient of highest degree) is equal to 1. That is to say, a monic polynomial is one ...
of degree . Therefore the polynomial equation has at most different solutions, that is, eigenvalues of the matrix. They may be complex even if the entries of are real. According to the Cayley–Hamilton theorem, , that is, the result of substituting the matrix itself into its characteristic polynomial yields the
zero matrix In mathematics, particularly linear algebra, a zero matrix or null matrix is a matrix all of whose entries are zero. It also serves as the additive identity of the additive group of m \times n matrices, and is denoted by the symbol O or 0 followe ...
.


Computational aspects

Matrix calculations can be often performed with different techniques. Many problems can be solved by both direct algorithms and iterative approaches. For example, the eigenvectors of a square matrix can be obtained by finding a
sequence In mathematics, a sequence is an enumerated collection of objects in which repetitions are allowed and order matters. Like a set, it contains members (also called ''elements'', or ''terms''). The number of elements (possibly infinite) is cal ...
of vectors converging to an eigenvector when tends to
infinity Infinity is something which is boundless, endless, or larger than any natural number. It is denoted by \infty, called the infinity symbol. From the time of the Ancient Greek mathematics, ancient Greeks, the Infinity (philosophy), philosophic ...
. To choose the most appropriate algorithm for each specific problem, it is important to determine both the effectiveness and precision of all the available algorithms. The domain studying these matters is called numerical linear algebra. As with other numerical situations, two main aspects are the
complexity Complexity characterizes the behavior of a system or model whose components interact in multiple ways and follow local rules, leading to non-linearity, randomness, collective dynamics, hierarchy, and emergence. The term is generally used to c ...
of algorithms and their
numerical stability In the mathematical subfield of numerical analysis, numerical stability is a generally desirable property of numerical algorithms. The precise definition of stability depends on the context: one important context is numerical linear algebra, and ...
. Determining the complexity of an algorithm means finding
upper bound In mathematics, particularly in order theory, an upper bound or majorant of a subset of some preordered set is an element of that is every element of . Dually, a lower bound or minorant of is defined to be an element of that is less ...
s or estimates of how many elementary operations such as additions and multiplications of scalars are necessary to perform some algorithm, for example, multiplication of matrices. Calculating the matrix product of two -by- matrices using the definition given above needs multiplications, since for any of the entries of the product, multiplications are necessary. The Strassen algorithm outperforms this "naive" algorithm; it needs only multiplications. Theoretically faster but impractical
matrix multiplication algorithm Because matrix multiplication is such a central operation in many numerical algorithms, much work has been invested in making matrix multiplication algorithms efficient. Applications of matrix multiplication in computational problems are found in m ...
s have been developed, as have speedups to this problem using parallel algorithms or distributed computation systems such as
MapReduce MapReduce is a programming model and an associated implementation for processing and generating big data sets with a parallel and distributed algorithm on a cluster. A MapReduce program is composed of a ''map'' procedure, which performs filte ...
. In many practical situations, additional information about the matrices involved is known. An important case concerns sparse matrices, that is, matrices whose entries are mostly zero. There are specifically adapted algorithms for, say, solving linear systems for sparse matrices , such as the conjugate gradient method. An algorithm is, roughly speaking, numerically stable if little deviations in the input values do not lead to big deviations in the result. For example, one can calculate the inverse of a matrix by computing its adjugate matrix: \bold A^ = \operatorname(\bold A) / \det(\bold A). However, this may lead to significant rounding errors if the determinant of the matrix is very small. The norm of a matrix can be used to capture the conditioning of linear algebraic problems, such as computing a matrix's inverse.


Decomposition

There are several methods to render matrices into a more easily accessible form. They are generally referred to as ''matrix decomposition'' or ''matrix factorization'' techniques. These techniques are of interest because they can make computations easier. The LU decomposition factors matrices as a product of lower () and an upper triangular matrices (). Once this decomposition is calculated, linear systems can be solved more efficiently by a simple technique called forward and back substitution. Likewise, inverses of triangular matrices are algorithmically easier to calculate. The ''Gaussian elimination'' is a similar algorithm; it transforms any matrix to
row echelon form In linear algebra, a matrix is in row echelon form if it can be obtained as the result of Gaussian elimination. Every matrix can be put in row echelon form by applying a sequence of elementary row operations. The term ''echelon'' comes from the F ...
. Both methods proceed by multiplying the matrix by suitable elementary matrices, which correspond to permuting rows or columns and adding multiples of one row to another row.
Singular value decomposition In linear algebra, the singular value decomposition (SVD) is a Matrix decomposition, factorization of a real number, real or complex number, complex matrix (mathematics), matrix into a rotation, followed by a rescaling followed by another rota ...
expresses any matrix as a product , where and are unitary matrices and is a diagonal matrix. The eigendecomposition or ''diagonalization'' expresses as a product , where is a diagonal matrix and is a suitable invertible matrix. If can be written in this form, it is called diagonalizable. More generally, and applicable to all matrices, the Jordan decomposition transforms a matrix into
Jordan normal form \begin \lambda_1 1\hphantom\hphantom\\ \hphantom\lambda_1 1\hphantom\\ \hphantom\lambda_1\hphantom\\ \hphantom\lambda_2 1\hphantom\hphantom\\ \hphantom\hphantom\lambda_2\hphantom\\ \hphantom\lambda_3\hphantom\\ \hphantom\ddots\hphantom\\ ...
, that is to say matrices whose only nonzero entries are the eigenvalues to of , placed on the main diagonal and possibly entries equal to one directly above the main diagonal, as shown at the right. Given the eigendecomposition, the th power of (that is, -fold iterated matrix multiplication) can be calculated via \bold A^n = (\bold^)^ = \bold^ \bold^ \ldots \bold^ = \bold^n \bold V^ and the power of a diagonal matrix can be calculated by taking the corresponding powers of the diagonal entries, which is much easier than doing the exponentiation for instead. This can be used to compute the
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
, a need frequently arising in solving
linear differential equation In mathematics, a linear differential equation is a differential equation that is linear equation, linear in the unknown function and its derivatives, so it can be written in the form a_0(x)y + a_1(x)y' + a_2(x)y'' \cdots + a_n(x)y^ = b(x) wher ...
s, matrix logarithms and square roots of matrices. To avoid numerically ill-conditioned situations, further algorithms such as the
Schur decomposition In the mathematical discipline of linear algebra, the Schur decomposition or Schur triangulation, named after Issai Schur, is a matrix decomposition. It allows one to write an arbitrary complex square matrix as unitarily similar to an upper tria ...
can be employed.


Abstract algebraic aspects and generalizations

Matrices can be generalized in different ways. Abstract algebra uses matrices with entries in more general fields or even rings, while linear algebra codifies properties of matrices in the notion of linear maps. It is possible to consider matrices with infinitely many columns and rows. Another extension is
tensor In mathematics, a tensor is an algebraic object that describes a multilinear relationship between sets of algebraic objects associated with a vector space. Tensors may map between different objects such as vectors, scalars, and even other ...
s, which can be seen as higher-dimensional arrays of numbers, as opposed to vectors, which can often be realized as sequences of numbers, while matrices are rectangular or two-dimensional arrays of numbers. Matrices, subject to certain requirements tend to form groups known as matrix groups. Similarly under certain conditions matrices form rings known as matrix rings. Though the product of matrices is not in general commutative certain matrices form fields known as matrix fields. In general, matrices and their
multiplication Multiplication is one of the four elementary mathematical operations of arithmetic, with the other ones being addition, subtraction, and division (mathematics), division. The result of a multiplication operation is called a ''Product (mathem ...
also form a
category Category, plural categories, may refer to: General uses *Classification, the general act of allocating things to classes/categories Philosophy * Category of being * ''Categories'' (Aristotle) * Category (Kant) * Categories (Peirce) * Category ( ...
, the
category of matrices In mathematics, the category of matrices, often denoted \mathbf, is the category (category theory), category whose object (category theory), objects are natural numbers and whose morphisms are matrix (mathematics), matrices, with composition give ...
.


Matrices with more general entries

This article focuses on matrices whose entries are real or complex numbers. However, matrices can be considered with much more general types of entries than real or complex numbers. As a first step of generalization, any field, that is, a
set Set, The Set, SET or SETS may refer to: Science, technology, and mathematics Mathematics *Set (mathematics), a collection of elements *Category of sets, the category whose objects and morphisms are sets and total functions, respectively Electro ...
where
addition Addition (usually signified by the Plus and minus signs#Plus sign, plus symbol, +) is one of the four basic Operation (mathematics), operations of arithmetic, the other three being subtraction, multiplication, and Division (mathematics), divis ...
,
subtraction Subtraction (which is signified by the minus sign, –) is one of the four Arithmetic#Arithmetic operations, arithmetic operations along with addition, multiplication and Division (mathematics), division. Subtraction is an operation that repre ...
,
multiplication Multiplication is one of the four elementary mathematical operations of arithmetic, with the other ones being addition, subtraction, and division (mathematics), division. The result of a multiplication operation is called a ''Product (mathem ...
, and division operations are defined and well-behaved, may be used instead of or , for example
rational number In mathematics, a rational number is a number that can be expressed as the quotient or fraction of two integers, a numerator and a non-zero denominator . For example, is a rational number, as is every integer (for example, The set of all ...
s or
finite field In mathematics, a finite field or Galois field (so-named in honor of Évariste Galois) is a field (mathematics), field that contains a finite number of Element (mathematics), elements. As with any field, a finite field is a Set (mathematics), s ...
s. For example,
coding theory Coding theory is the study of the properties of codes and their respective fitness for specific applications. Codes are used for data compression, cryptography, error detection and correction, data transmission and computer data storage, data sto ...
makes use of matrices over finite fields. Wherever
eigenvalue In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
s are considered, as these are roots of a polynomial they may exist only in a larger field than that of the entries of the matrix; for instance, they may be complex in the case of a matrix with real entries. The possibility to reinterpret the entries of a matrix as elements of a larger field (for example, to view a real matrix as a complex matrix whose entries happen to be all real) then allows considering each square matrix to possess a full set of eigenvalues. Alternatively one can consider only matrices with entries in an
algebraically closed field In mathematics, a field is algebraically closed if every non-constant polynomial in (the univariate polynomial ring with coefficients in ) has a root in . In other words, a field is algebraically closed if the fundamental theorem of algebra ...
, such as from the outset. More generally, matrices with entries in a ring are widely used in mathematics. Rings are a more general notion than fields in that a division operation need not exist. The very same addition and multiplication operations of matrices extend to this setting, too. The set (also denoted ) of all square -by- matrices over is a ring called matrix ring, isomorphic to the
endomorphism ring In mathematics, the endomorphisms of an abelian group ''X'' form a ring. This ring is called the endomorphism ring of ''X'', denoted by End(''X''); the set of all homomorphisms of ''X'' into itself. Addition of endomorphisms arises naturally in ...
of the left - module . If the ring is
commutative In mathematics, a binary operation is commutative if changing the order of the operands does not change the result. It is a fundamental property of many binary operations, and many mathematical proofs depend on it. Perhaps most familiar as a pr ...
, that is, its multiplication is commutative, then the ring is also an
associative algebra In mathematics, an associative algebra ''A'' over a commutative ring (often a field) ''K'' is a ring ''A'' together with a ring homomorphism from ''K'' into the center of ''A''. This is thus an algebraic structure with an addition, a mult ...
over . The
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of square matrices over a commutative ring can still be defined using the Leibniz formula; such a matrix is invertible if and only if its determinant is
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
in , generalizing the situation over a field , where every nonzero element is invertible. Matrices over superrings are called supermatrices. Matrices do not always have all their entries in the same ring– or even in any ring at all. One special but common case is block matrices, which may be considered as matrices whose entries themselves are matrices. The entries need not be square matrices, and thus need not be members of any ring; but their sizes must fulfill certain compatibility conditions.


Relationship to linear maps

Linear maps \R^n \to \R^m are equivalent to -by- matrices, as described above. More generally, any linear map between finite- dimensional
vector space In mathematics and physics, a vector space (also called a linear space) is a set (mathematics), set whose elements, often called vector (mathematics and physics), ''vectors'', can be added together and multiplied ("scaled") by numbers called sc ...
s can be described by a matrix , after choosing bases of , and of (so is the dimension of and is the dimension of ), which is such that f(\mathbf_j) = \sum_^m a_ \mathbf_i\qquad\mbox\ j=1,\ldots,n. In other words, column of expresses the image of in terms of the basis vectors of ; thus this relation uniquely determines the entries of the matrix . The matrix depends on the choice of the bases: different choices of bases give rise to different, but equivalent matrices. Many of the above concrete notions can be reinterpreted in this light, for example, the transpose matrix describes the transpose of the linear map given by , concerning the dual bases. These properties can be restated more naturally: the
category of matrices In mathematics, the category of matrices, often denoted \mathbf, is the category (category theory), category whose object (category theory), objects are natural numbers and whose morphisms are matrix (mathematics), matrices, with composition give ...
with entries in a field k with multiplication as composition is equivalent to the category of finite-dimensional
vector space In mathematics and physics, a vector space (also called a linear space) is a set (mathematics), set whose elements, often called vector (mathematics and physics), ''vectors'', can be added together and multiplied ("scaled") by numbers called sc ...
s and linear maps over this field. More generally, the set of matrices can be used to represent the -linear maps between the free modules and for an arbitrary ring with unity. When composition of these maps is possible, and this gives rise to the matrix ring of matrices representing the
endomorphism ring In mathematics, the endomorphisms of an abelian group ''X'' form a ring. This ring is called the endomorphism ring of ''X'', denoted by End(''X''); the set of all homomorphisms of ''X'' into itself. Addition of endomorphisms arises naturally in ...
of .


Matrix groups

A group is a mathematical structure consisting of a set of objects together with a
binary operation In mathematics, a binary operation or dyadic operation is a rule for combining two elements (called operands) to produce another element. More formally, a binary operation is an operation of arity two. More specifically, a binary operation ...
, that is, an operation combining any two objects to a third, subject to certain requirements. A group in which the objects are matrices and the group operation is matrix multiplication is called a ''matrix group''. Since a group of every element must be invertible, the most general matrix groups are the groups of all invertible matrices of a given size, called the
general linear group In mathematics, the general linear group of degree n is the set of n\times n invertible matrices, together with the operation of ordinary matrix multiplication. This forms a group, because the product of two invertible matrices is again inve ...
s. Any property of matrices that is preserved under matrix products and inverses can be used to define further matrix groups. For example, matrices with a given size and with a determinant of form a
subgroup In group theory, a branch of mathematics, a subset of a group G is a subgroup of G if the members of that subset form a group with respect to the group operation in G. Formally, given a group (mathematics), group under a binary operation  ...
of (that is, a smaller group contained in) their general linear group, called a special linear group. Orthogonal matrices, determined by the condition \bold M^ \bold M = \bold I, form the
orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the Group (mathematics), group of isometry, distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by ...
. Every orthogonal matrix has
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
or . Orthogonal matrices with determinant form a subgroup called the ''special orthogonal group''. Every
finite group In abstract algebra, a finite group is a group whose underlying set is finite. Finite groups often arise when considering symmetry of mathematical or physical objects, when those objects admit just a finite number of structure-preserving tra ...
is
isomorphic In mathematics, an isomorphism is a structure-preserving mapping or morphism between two structures of the same type that can be reversed by an inverse mapping. Two mathematical structures are isomorphic if an isomorphism exists between the ...
to a matrix group, as one can see by considering the regular representation of the
symmetric group In abstract algebra, the symmetric group defined over any set is the group whose elements are all the bijections from the set to itself, and whose group operation is the composition of functions. In particular, the finite symmetric grou ...
. General groups can be studied using matrix groups, which are comparatively well understood, using
representation theory Representation theory is a branch of mathematics that studies abstract algebra, abstract algebraic structures by ''representing'' their element (set theory), elements as linear transformations of vector spaces, and studies Module (mathematics), ...
.


Infinite matrices

It is also possible to consider matrices with infinitely many rows and/or columns. The basic operations introduced above are defined the same way in this case. Matrix multiplication, however, and all operations stemming therefrom are only meaningful when restricted to certain matrices, since the sum featuring in the above definition of the matrix product will contain an infinity of summands. An easy way to circumvent this issue is to restrict to matrices all of whose rows (or columns) contain only finitely many nonzero terms. As in the finite case (see above), where matrices describe linear maps, infinite matrices can be used to describe operators on Hilbert spaces, where convergence and continuity questions arise. However, the explicit point of view of matrices tends to obfuscate the matter, and the abstract and more powerful tools of
functional analysis Functional analysis is a branch of mathematical analysis, the core of which is formed by the study of vector spaces endowed with some kind of limit-related structure (for example, Inner product space#Definition, inner product, Norm (mathematics ...
are used instead, by relating matrices to linear maps (as in the finite case above), but imposing additional convergence and continuity constraints.


Empty matrix

An ''empty matrix'' is a matrix in which the number of rows or columns (or both) is zero."A matrix having at least one dimension equal to zero is called an empty matrix"
MATLAB Data Structures
Empty matrices help to deal with maps involving the zero vector space. For example, if is a 3-by-0 matrix and is a 0-by-3 matrix, then is the 3-by-3 zero matrix corresponding to the null map from a 3-dimensional space to itself, while is a 0-by-0 matrix. There is no common notation for empty matrices, but most
computer algebra system A computer algebra system (CAS) or symbolic algebra system (SAS) is any mathematical software with the ability to manipulate mathematical expressions in a way similar to the traditional manual computations of mathematicians and scientists. The de ...
s allow creating and computing with them. The determinant of the 0-by-0 matrix is 1 as follows regarding the
empty product In mathematics, an empty product, or nullary product or vacuous product, is the result of multiplication, multiplying no factors. It is by convention equal to the multiplicative identity (assuming there is an identity for the multiplication operat ...
occurring in the Leibniz formula for the determinant as 1. This value is also consistent with the fact that the identity map from any finite-dimensional space to itself has determinant1, a fact that is often used as a part of the characterization of determinants.


Applications

There are numerous applications of matrices, both in mathematics and other sciences. Some of them merely take advantage of the compact representation of a set of numbers in a matrix. For example, Text mining and automated
thesaurus A thesaurus (: thesauri or thesauruses), sometimes called a synonym dictionary or dictionary of synonyms, is a reference work which arranges words by their meanings (or in simpler terms, a book where one can find different words with similar me ...
compilation makes use of document-term matrices such as tf-idf to track frequencies of certain words in several documents. Complex numbers can be represented by particular real 2-by-2 matrices via a + ib \leftrightarrow \begin a & -b \\ b & a \end, under which addition and multiplication of complex numbers and matrices correspond to each other. For example, 2-by-2 rotation matrices represent the multiplication with some complex number of
absolute value In mathematics, the absolute value or modulus of a real number x, is the non-negative value without regard to its sign. Namely, , x, =x if x is a positive number, and , x, =-x if x is negative (in which case negating x makes -x positive), ...
1, as above. A similar interpretation is possible for
quaternion In mathematics, the quaternion number system extends the complex numbers. Quaternions were first described by the Irish mathematician William Rowan Hamilton in 1843 and applied to mechanics in three-dimensional space. The algebra of quater ...
s and
Clifford algebra In mathematics, a Clifford algebra is an algebra generated by a vector space with a quadratic form, and is a unital associative algebra with the additional structure of a distinguished subspace. As -algebras, they generalize the real number ...
s in general. In
game theory Game theory is the study of mathematical models of strategic interactions. It has applications in many fields of social science, and is used extensively in economics, logic, systems science and computer science. Initially, game theory addressed ...
and
economics Economics () is a behavioral science that studies the Production (economics), production, distribution (economics), distribution, and Consumption (economics), consumption of goods and services. Economics focuses on the behaviour and interac ...
, the payoff matrix encodes the payoff for two players, depending on which out of a given (finite) set of strategies the players choose. The expected outcome of the game, when both players play mixed strategies, is obtained by multiplying this matrix on both sides by vectors representing the strategies. The
minimax theorem In the mathematical area of game theory and of convex optimization, a minimax theorem is a theorem that claims that : \max_ \min_ f(x,y) = \min_ \max_f(x,y) under certain conditions on the sets X and Y and on the function f. It is always true that ...
central to game theory is closely related to the duality theory of linear programs, which are often formulated in terms of matrix-vector products. Early
encryption In Cryptography law, cryptography, encryption (more specifically, Code, encoding) is the process of transforming information in a way that, ideally, only authorized parties can decode. This process converts the original representation of the inf ...
techniques such as the Hill cipher also used matrices. However, due to the linear nature of matrices, these codes are comparatively easy to break.
Computer graphics Computer graphics deals with generating images and art with the aid of computers. Computer graphics is a core technology in digital photography, film, video games, digital art, cell phone and computer displays, and many specialized applications. ...
uses matrices to represent objects; to calculate transformations of objects using affine rotation matrices to accomplish tasks such as projecting a three-dimensional object onto a two-dimensional screen, corresponding to a theoretical camera observation; and to apply image convolutions such as sharpening, blurring, edge detection, and more. Matrices over a
polynomial ring In mathematics, especially in the field of algebra, a polynomial ring or polynomial algebra is a ring formed from the set of polynomials in one or more indeterminates (traditionally also called variables) with coefficients in another ring, ...
are important in the study of
control theory Control theory is a field of control engineering and applied mathematics that deals with the control system, control of dynamical systems in engineered processes and machines. The objective is to develop a model or algorithm governing the applic ...
.
Chemistry Chemistry is the scientific study of the properties and behavior of matter. It is a physical science within the natural sciences that studies the chemical elements that make up matter and chemical compound, compounds made of atoms, molecules a ...
makes use of matrices in various ways, particularly since the use of quantum theory to discuss molecular bonding and
spectroscopy Spectroscopy is the field of study that measures and interprets electromagnetic spectra. In narrower contexts, spectroscopy is the precise study of color as generalized from visible light to all bands of the electromagnetic spectrum. Spectro ...
. Examples are the overlap matrix and the
Fock matrix The Fock matrix is defined by the Fock operator. In its general form the Fock operator writes: :\hat F(i) = \hat h(i)+\sum_^ hat J_j(i)-\hat K_j(i)/math> Where ''i'' runs over the total ''N'' spin orbitals. In the closed-shell case, it can be si ...
used in solving the Roothaan equations to obtain the molecular orbitals of the Hartree–Fock method.


Graph theory

The
adjacency matrix In graph theory and computer science, an adjacency matrix is a square matrix used to represent a finite graph (discrete mathematics), graph. The elements of the matrix (mathematics), matrix indicate whether pairs of Vertex (graph theory), vertices ...
of a finite graph is a basic notion of
graph theory In mathematics and computer science, graph theory is the study of ''graph (discrete mathematics), graphs'', which are mathematical structures used to model pairwise relations between objects. A graph in this context is made up of ''Vertex (graph ...
. It records which vertices of the graph are connected by an edge. Matrices containing just two different values ( and meaning for example "yes" and "no", respectively) are called logical matrices. The distance (or cost) matrix contains information about the distances of the edges. These concepts can be applied to
website A website (also written as a web site) is any web page whose content is identified by a common domain name and is published on at least one web server. Websites are typically dedicated to a particular topic or purpose, such as news, educatio ...
s connected by
hyperlink In computing, a hyperlink, or simply a link, is a digital reference providing direct access to Data (computing), data by a user (computing), user's point and click, clicking or touchscreen, tapping. A hyperlink points to a whole document or to ...
s, or cities connected by roads etc., in which case (unless the connection network is extremely dense) the matrices tend to be sparse, that is, contain few nonzero entries. Therefore, specifically tailored matrix algorithms can be used in
network theory In mathematics, computer science, and network science, network theory is a part of graph theory. It defines networks as Graph (discrete mathematics), graphs where the vertices or edges possess attributes. Network theory analyses these networks ...
.


Analysis and geometry

The
Hessian matrix In mathematics, the Hessian matrix, Hessian or (less commonly) Hesse matrix is a square matrix of second-order partial derivatives of a scalar-valued Function (mathematics), function, or scalar field. It describes the local curvature of a functio ...
of a
differentiable function In mathematics, a differentiable function of one real variable is a function whose derivative exists at each point in its domain. In other words, the graph of a differentiable function has a non- vertical tangent line at each interior point in ...
f: \R^n \to \R consists of the
second derivative In calculus, the second derivative, or the second-order derivative, of a function is the derivative of the derivative of . Informally, the second derivative can be phrased as "the rate of change of the rate of change"; for example, the secon ...
s of concerning the several coordinate directions, that is, H(f) = \left frac \right It encodes information about the local growth behavior of the function: given a critical point , that is, a point where the first
partial derivative In mathematics, a partial derivative of a function of several variables is its derivative with respect to one of those variables, with the others held constant (as opposed to the total derivative, in which all variables are allowed to vary). P ...
s \partial f / \partial x_i of vanish, the function has a
local minimum In mathematical analysis, the maximum and minimum of a function are, respectively, the greatest and least value taken by the function. Known generically as extremum, they may be defined either within a given range (the ''local'' or ''relative ...
if the Hessian matrix is positive definite. Quadratic programming can be used to find global minima or maxima of quadratic functions closely related to the ones attached to matrices (see above). Another matrix frequently used in geometrical situations is the Jacobi matrix of a differentiable map . If denote the components of , then the Jacobi matrix is defined as J(f) = \left frac \right . If , and if the rank of the Jacobi matrix attains its maximal value , is locally invertible at that point, by the implicit function theorem.
Partial differential equation In mathematics, a partial differential equation (PDE) is an equation which involves a multivariable function and one or more of its partial derivatives. The function is often thought of as an "unknown" that solves the equation, similar to ho ...
s can be classified by considering the matrix of coefficients of the highest-order differential operators of the equation. For
elliptic partial differential equation In mathematics, an elliptic partial differential equation is a type of partial differential equation (PDE). In mathematical modeling, elliptic PDEs are frequently used to model steady states, unlike parabolic PDE and hyperbolic PDE which gene ...
s this matrix is positive definite, which has a decisive influence on the set of possible solutions of the equation in question. The
finite element method Finite element method (FEM) is a popular method for numerically solving differential equations arising in engineering and mathematical modeling. Typical problem areas of interest include the traditional fields of structural analysis, heat tran ...
is an important numerical method to solve partial differential equations, widely applied in simulating complex physical systems. It attempts to approximate the solution to some equation by piecewise linear functions, where the pieces are chosen concerning a sufficiently fine grid, which in turn can be recast as a matrix equation.


Probability theory and statistics

Stochastic matrices are square matrices whose rows are probability vectors, that is, whose entries are non-negative and sum up to one. Stochastic matrices are used to define
Markov chain In probability theory and statistics, a Markov chain or Markov process is a stochastic process describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally ...
s with finitely many states. A row of the stochastic matrix gives the probability distribution for the next position of some particle currently in the state that corresponds to the row. Properties of the Markov chain—like absorbing states, that is, states that any particle attains eventually—can be read off the eigenvectors of the transition matrices. Statistics also makes use of matrices in many different forms.
Descriptive statistics A descriptive statistic (in the count noun sense) is a summary statistic that quantitatively describes or summarizes features from a collection of information, while descriptive statistics (in the mass noun sense) is the process of using and an ...
is concerned with describing data sets, which can often be represented as data matrices, which may then be subjected to dimensionality reduction techniques. The covariance matrix encodes the mutual
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
of several
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s. Another technique using matrices are linear least squares, a method that approximates a finite set of pairs , by a linear function y_i \approx ax_i + b, \quad i = 1, \ldots, N which can be formulated in terms of matrices, related to the
singular value decomposition In linear algebra, the singular value decomposition (SVD) is a Matrix decomposition, factorization of a real number, real or complex number, complex matrix (mathematics), matrix into a rotation, followed by a rescaling followed by another rota ...
of matrices. Random matrices are matrices whose entries are random numbers, subject to suitable
probability distribution In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
s, such as matrix normal distribution. Beyond probability theory, they are applied in domains ranging from
number theory Number theory is a branch of pure mathematics devoted primarily to the study of the integers and arithmetic functions. Number theorists study prime numbers as well as the properties of mathematical objects constructed from integers (for example ...
to
physics Physics is the scientific study of matter, its Elementary particle, fundamental constituents, its motion and behavior through space and time, and the related entities of energy and force. "Physical science is that department of knowledge whi ...
.


Quantum mechanics and particle physics

The first model of
quantum mechanics Quantum mechanics is the fundamental physical Scientific theory, theory that describes the behavior of matter and of light; its unusual characteristics typically occur at and below the scale of atoms. Reprinted, Addison-Wesley, 1989, It is ...
( Heisenberg, 1925) used infinite-dimensional matrices to define the operators that took over the role of variables like position, momentum and energy from classical physics. (This is sometimes referred to as matrix mechanics.) Matrices, both finite and infinite-dimensional, have since been employed for many purposes in quantum mechanics. One particular example is the
density matrix In quantum mechanics, a density matrix (or density operator) is a matrix used in calculating the probabilities of the outcomes of measurements performed on physical systems. It is a generalization of the state vectors or wavefunctions: while th ...
, a tool used in calculating the
probabilities Probability is a branch of mathematics and statistics concerning Event (probability theory), events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probab ...
of the outcomes of measurements performed on
physical system A physical system is a collection of physical objects under study. The collection differs from a set: all the objects must coexist and have some physical relationship. In other words, it is a portion of the physical universe chosen for analys ...
s. Linear transformations and the associated symmetries play a key role in modern physics. For example,
elementary particle In particle physics, an elementary particle or fundamental particle is a subatomic particle that is not composed of other particles. The Standard Model presently recognizes seventeen distinct particles—twelve fermions and five bosons. As a c ...
s in
quantum field theory In theoretical physics, quantum field theory (QFT) is a theoretical framework that combines Field theory (physics), field theory and the principle of relativity with ideas behind quantum mechanics. QFT is used in particle physics to construct phy ...
are classified as representations of the
Lorentz group In physics and mathematics, the Lorentz group is the group of all Lorentz transformations of Minkowski spacetime, the classical and quantum setting for all (non-gravitational) physical phenomena. The Lorentz group is named for the Dutch physi ...
of special relativity and, more specifically, by their behavior under the spin group. Concrete representations involving the
Pauli matrices In mathematical physics and mathematics, the Pauli matrices are a set of three complex matrices that are traceless, Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma (), they are occasionally denoted by tau () ...
and more general
gamma matrices In mathematical physics, the gamma matrices, \ \left\\ , also called the Dirac matrices, are a set of conventional matrices with specific anticommutation relations that ensure they generate a matrix representation of the Clifford algebra \ \mathr ...
are an integral part of the physical description of
fermion In particle physics, a fermion is a subatomic particle that follows Fermi–Dirac statistics. Fermions have a half-integer spin (spin 1/2, spin , Spin (physics)#Higher spins, spin , etc.) and obey the Pauli exclusion principle. These particles i ...
s, which behave as
spinor In geometry and physics, spinors (pronounced "spinner" IPA ) are elements of a complex numbers, complex vector space that can be associated with Euclidean space. A spinor transforms linearly when the Euclidean space is subjected to a slight (infi ...
s. For the three lightest
quark A quark () is a type of elementary particle and a fundamental constituent of matter. Quarks combine to form composite particles called hadrons, the most stable of which are protons and neutrons, the components of atomic nucleus, atomic nuclei ...
s, there is a group-theoretical representation involving the
special unitary group In mathematics, the special unitary group of degree , denoted , is the Lie group of unitary matrices with determinant 1. The matrices of the more general unitary group may have complex determinants with absolute value 1, rather than real 1 ...
SU(3); for their calculations, physicists use a convenient matrix representation known as the Gell-Mann matrices, which are also used for the SU(3) gauge group that forms the basis of the modern description of strong nuclear interactions,
quantum chromodynamics In theoretical physics, quantum chromodynamics (QCD) is the study of the strong interaction between quarks mediated by gluons. Quarks are fundamental particles that make up composite hadrons such as the proton, neutron and pion. QCD is a type of ...
. The Cabibbo–Kobayashi–Maskawa matrix, in turn, expresses the fact that the basic quark states that are important for
weak interaction In nuclear physics and particle physics, the weak interaction, weak force or the weak nuclear force, is one of the four known fundamental interactions, with the others being electromagnetism, the strong interaction, and gravitation. It is th ...
s are not the same as, but linearly related to the basic quark states that define particles with specific and distinct
mass Mass is an Intrinsic and extrinsic properties, intrinsic property of a physical body, body. It was traditionally believed to be related to the physical quantity, quantity of matter in a body, until the discovery of the atom and particle physi ...
es. Another matrix serves as a key tool for describing the scattering experiments that form the cornerstone of experimental particle physics: Collision reactions such as occur in
particle accelerator A particle accelerator is a machine that uses electromagnetic fields to propel electric charge, charged particles to very high speeds and energies to contain them in well-defined particle beam, beams. Small accelerators are used for fundamental ...
s, where non-interacting particles head towards each other and collide in a small interaction zone, with a new set of non-interacting particles as the result, can be described as the scalar product of outgoing particle states and a linear combination of ingoing particle states. The linear combination is given by a matrix known as the
S-matrix In physics, the ''S''-matrix or scattering matrix is a Matrix (mathematics), matrix that relates the initial state and the final state of a physical system undergoing a scattering, scattering process. It is used in quantum mechanics, scattering ...
, which encodes all information about the possible interactions between particles.


Normal modes

A general application of matrices in physics is the description of linearly coupled harmonic systems. The
equations of motion In physics, equations of motion are equations that describe the behavior of a physical system in terms of its motion as a function of time. More specifically, the equations of motion describe the behavior of a physical system as a set of mathem ...
of such systems can be described in matrix form, with a mass matrix multiplying a generalized velocity to give the kinetic term, and a
force In physics, a force is an influence that can cause an Physical object, object to change its velocity unless counterbalanced by other forces. In mechanics, force makes ideas like 'pushing' or 'pulling' mathematically precise. Because the Magnitu ...
matrix multiplying a displacement vector to characterize the interactions. The best way to obtain solutions is to determine the system's
eigenvector In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by ...
s, its
normal mode A normal mode of a dynamical system is a pattern of motion in which all parts of the system move sinusoidally with the same frequency and with a fixed phase relation. The free motion described by the normal modes takes place at fixed frequencies ...
s, by diagonalizing the matrix equation. Techniques like this are crucial when it comes to the internal dynamics of
molecules A molecule is a group of two or more atoms that are held together by attractive forces known as chemical bonds; depending on context, the term may or may not include ions that satisfy this criterion. In quantum physics, organic chemistry ...
: the internal vibrations of systems consisting of mutually bound component atoms. They are also needed for describing mechanical vibrations, and oscillations in electrical circuits.


Geometrical optics

Geometrical optics Geometrical optics, or ray optics, is a model of optics that describes light Wave propagation, propagation in terms of ''ray (optics), rays''. The ray in geometrical optics is an abstract object, abstraction useful for approximating the paths along ...
provides further matrix applications. In this approximative theory, the wave nature of light is neglected. The result is a model in which light rays are indeed geometrical rays. If the deflection of light rays by optical elements is small, the action of a
lens A lens is a transmissive optical device that focuses or disperses a light beam by means of refraction. A simple lens consists of a single piece of transparent material, while a compound lens consists of several simple lenses (''elements'') ...
or reflective element on a given light ray can be expressed as multiplication of a two-component vector with a two-by-two matrix called ray transfer matrix analysis: the vector's components are the light ray's slope and its distance from the optical axis, while the matrix encodes the properties of the optical element. There are two kinds of matrices, viz. a ''refraction matrix'' describing the refraction at a lens surface, and a ''translation matrix'', describing the translation of the plane of reference to the next refracting surface, where another refraction matrix applies. The optical system, consisting of a combination of lenses and/or reflective elements, is simply described by the matrix resulting from the product of the components' matrices. The
Jones calculus In optics, polarized light can be described using the Jones calculus, invented by R. C. Jones in 1941. Polarized light is represented by a Jones vector, and linear optical elements are represented by ''Jones matrices''. When light crosses an opt ...
models the polarization of a light source as a 2\times 2 vector, and the effects of optical filters on this polarization vector as a matrix.


Electronics

Electronic circuits that are composed of linear components (such as resistors, inductors and capacitors) obey
Kirchhoff's circuit laws Kirchhoff's circuit laws are two equalities that deal with the current and potential difference (commonly known as voltage) in the lumped element model of electrical circuits. They were first described in 1845 by German physicist Gustav Kirc ...
, which leads to a system of linear equations, which can be described with a matrix equation that relates the source currents and voltages to the resultant currents and voltages at each point in the circuit, and where the matrix entries are determined by the circuit.


History

Matrices have a long history of application in solving
linear equation In mathematics, a linear equation is an equation that may be put in the form a_1x_1+\ldots+a_nx_n+b=0, where x_1,\ldots,x_n are the variables (or unknowns), and b,a_1,\ldots,a_n are the coefficients, which are often real numbers. The coeffici ...
s but they were known as arrays until the 1800s. The Chinese text ''
The Nine Chapters on the Mathematical Art ''The Nine Chapters on the Mathematical Art'' is a Chinese mathematics book, composed by several generations of scholars from the 10th–2nd century BCE, its latest stage being from the 1st century CE. This book is one of the earliest surviving ...
'' written in the 10th–2nd century BCE is the first example of the use of array methods to solve
simultaneous equations In mathematics, a set of simultaneous equations, also known as a system of equations or an equation system, is a finite set of equations for which common solutions are sought. An equation system is usually classified in the same manner as single e ...
, including the concept of
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
s. In 1545 Italian mathematician
Gerolamo Cardano Gerolamo Cardano (; also Girolamo or Geronimo; ; ; 24 September 1501– 21 September 1576) was an Italian polymath whose interests and proficiencies ranged through those of mathematician, physician, biologist, physicist, chemist, astrologer, as ...
introduced the method to Europe when he published ''Ars Magna''. The Japanese mathematician Seki used the same array methods to solve simultaneous equations in 1683. The Dutch mathematician'' '' Jan de Witt represented transformations using arrays in his 1659 book ''Elements of Curves'' (1659). Between 1700 and 1710
Gottfried Wilhelm Leibniz Gottfried Wilhelm Leibniz (or Leibnitz; – 14 November 1716) was a German polymath active as a mathematician, philosopher, scientist and diplomat who is credited, alongside Sir Isaac Newton, with the creation of calculus in addition to ...
publicized the use of arrays for recording information or solutions and experimented with over 50 different systems of arrays. Cramer presented his rule in 1750. The term "matrix" (Latin for "womb", "dam" (non-human female animal kept for breeding), "source", "origin", "list", and "register", are derived from '' mater''—mother) was coined by
James Joseph Sylvester James Joseph Sylvester (3 September 1814 – 15 March 1897) was an English mathematician. He made fundamental contributions to matrix theory, invariant theory, number theory, partition theory, and combinatorics. He played a leadership ...
in 1850, who understood a matrix as an object giving rise to several determinants today called minors, that is to say, determinants of smaller matrices that derive from the original one by removing columns and rows. In an 1851 paper, Sylvester explains: Arthur Cayley published a treatise on geometric transformations using matrices that were not rotated versions of the coefficients being investigated as had previously been done. Instead, he defined operations such as addition, subtraction, multiplication, and division as transformations of those matrices and showed the associative and distributive properties held. Cayley investigated and demonstrated the non-commutative property of matrix multiplication as well as the commutative property of matrix addition. Early matrix theory had limited the use of arrays almost exclusively to determinants and Cayley's abstract matrix operations were revolutionary. He was instrumental in proposing a matrix concept independent of equation systems. In 1858, Cayley published his ''A memoir on the theory of matrices'' in which he proposed and demonstrated the Cayley–Hamilton theorem. The English mathematician Cuthbert Edmund Cullis was the first to use modern bracket notation for matrices in 1913 and he simultaneously demonstrated the first significant use of the notation to represent a matrix where refers to the th row and the th column. The modern study of determinants sprang from several sources. Number-theoretical problems led
Gauss Johann Carl Friedrich Gauss (; ; ; 30 April 177723 February 1855) was a German mathematician, astronomer, Geodesy, geodesist, and physicist, who contributed to many fields in mathematics and science. He was director of the Göttingen Observat ...
to relate coefficients of
quadratic form In mathematics, a quadratic form is a polynomial with terms all of degree two (" form" is another name for a homogeneous polynomial). For example, 4x^2 + 2xy - 3y^2 is a quadratic form in the variables and . The coefficients usually belong t ...
s, that is, expressions such as , and
linear map In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that p ...
s in three dimensions to matrices. Eisenstein further developed these notions, including the remark that, in modern parlance, matrix products are non-commutative. Cauchy was the first to prove general statements about determinants, using as the definition of the determinant of a matrix the following: replace the powers by in the
polynomial In mathematics, a polynomial is a Expression (mathematics), mathematical expression consisting of indeterminate (variable), indeterminates (also called variable (mathematics), variables) and coefficients, that involves only the operations of addit ...
a_1 a_2 \cdots a_n \prod_ (a_j - a_i), where \textstyle\prod denotes the product of the indicated terms. He also showed, in 1829, that the
eigenvalue In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
s of symmetric matrices are real. Jacobi studied "functional determinants"—later called Jacobi determinants by Sylvester—which can be used to describe geometric transformations at a local (or
infinitesimal In mathematics, an infinitesimal number is a non-zero quantity that is closer to 0 than any non-zero real number is. The word ''infinitesimal'' comes from a 17th-century Modern Latin coinage ''infinitesimus'', which originally referred to the " ...
) level, see above. Kronecker's ''Vorlesungen über die Theorie der Determinanten'' and
Weierstrass Karl Theodor Wilhelm Weierstrass (; ; 31 October 1815 – 19 February 1897) was a German mathematician often cited as the " father of modern analysis". Despite leaving university without a degree, he studied mathematics and trained as a school t ...
's ''Zur Determinantentheorie'', both published in 1903, first treated determinants
axiom An axiom, postulate, or assumption is a statement that is taken to be true, to serve as a premise or starting point for further reasoning and arguments. The word comes from the Ancient Greek word (), meaning 'that which is thought worthy or ...
atically, as opposed to previous more concrete approaches such as the mentioned formula of Cauchy. At that point, determinants were firmly established. Many theorems were first established for small matrices only, for example, the Cayley–Hamilton theorem was proved for 2×2 matrices by Cayley in the aforementioned memoir, and by
Hamilton Hamilton may refer to: * Alexander Hamilton (1755/1757–1804), first U.S. Secretary of the Treasury and one of the Founding Fathers of the United States * ''Hamilton'' (musical), a 2015 Broadway musical by Lin-Manuel Miranda ** ''Hamilton'' (al ...
for 4×4 matrices.
Frobenius Frobenius is a surname. Notable people with the surname include: * Ferdinand Georg Frobenius (1849–1917), mathematician ** Frobenius algebra ** Frobenius endomorphism ** Frobenius inner product ** Frobenius norm ** Frobenius method ** Frobenius g ...
, working on
bilinear form In mathematics, a bilinear form is a bilinear map on a vector space (the elements of which are called '' vectors'') over a field ''K'' (the elements of which are called '' scalars''). In other words, a bilinear form is a function that is linea ...
s, generalized the theorem to all dimensions (1898). Also at the end of the 19th century, the Gauss–Jordan elimination (generalizing a special case now known as Gauss elimination) was established by Wilhelm Jordan. In the early 20th century, matrices attained a central role in linear algebra, partially due to their use in the classification of the
hypercomplex number In mathematics, hypercomplex number is a traditional term for an element (mathematics), element of a finite-dimensional Algebra over a field#Unital algebra, unital algebra over a field, algebra over the field (mathematics), field of real numbers. ...
systems of the previous century. The inception of matrix mechanics by Heisenberg, Born and
Jordan Jordan, officially the Hashemite Kingdom of Jordan, is a country in the Southern Levant region of West Asia. Jordan is bordered by Syria to the north, Iraq to the east, Saudi Arabia to the south, and Israel and the occupied Palestinian ter ...
led to studying matrices with infinitely many rows and columns. Later, von Neumann carried out the
mathematical formulation of quantum mechanics The mathematical formulations of quantum mechanics are those mathematical formalisms that permit a rigorous description of quantum mechanics. This mathematical formalism uses mainly a part of functional analysis, especially Hilbert spaces, whic ...
, by further developing functional analytic notions such as
linear operator In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pr ...
s on
Hilbert space In mathematics, a Hilbert space is a real number, real or complex number, complex inner product space that is also a complete metric space with respect to the metric induced by the inner product. It generalizes the notion of Euclidean space. The ...
s, which, very roughly speaking, correspond to
Euclidean space Euclidean space is the fundamental space of geometry, intended to represent physical space. Originally, in Euclid's ''Elements'', it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are ''Euclidean spaces ...
, but with an infinity of independent directions.


Other historical usages of the word "matrix" in mathematics

The word has been used in unusual ways by at least two authors of historical importance.
Bertrand Russell Bertrand Arthur William Russell, 3rd Earl Russell, (18 May 1872 – 2 February 1970) was a British philosopher, logician, mathematician, and public intellectual. He had influence on mathematics, logic, set theory, and various areas of analytic ...
and
Alfred North Whitehead Alfred North Whitehead (15 February 1861 – 30 December 1947) was an English mathematician and philosopher. He created the philosophical school known as process philosophy, which has been applied in a wide variety of disciplines, inclu ...
in their ''
Principia Mathematica The ''Principia Mathematica'' (often abbreviated ''PM'') is a three-volume work on the foundations of mathematics written by the mathematician–philosophers Alfred North Whitehead and Bertrand Russell and published in 1910, 1912, and 1 ...
'' (1910–1913) use the word "matrix" in the context of their
axiom of reducibility The axiom of reducibility was introduced by Bertrand Russell in the early 20th century as part of his ramified theory of types. Russell devised and introduced the axiom in an attempt to manage the contradictions he had discovered in his analysis ...
. They proposed this axiom as a means to reduce any function to one of lower type, successively, so that at the "bottom" (0 order) the function is identical to its extension:Whitehead, Alfred North; and Russell, Bertrand (1913) ''Principia Mathematica to *56'', Cambridge at the University Press, Cambridge UK (republished 1962) cf page 162ff. For example, a function of two variables and can be reduced to a ''collection'' of functions of a single variable, such as , by "considering" the function for all possible values of "individuals" substituted in place of a variable . And then the resulting collection of functions of the single variable , that is, , can be reduced to a "matrix" of values by "considering" the function for all possible values of "individuals" substituted in place of variable : \forall b_j \forall a_i \colon \phi(a_i, b_j).
Alfred Tarski Alfred Tarski (; ; born Alfred Teitelbaum;School of Mathematics and Statistics, University of St Andrews ''School of Mathematics and Statistics, University of St Andrews''. January 14, 1901 – October 26, 1983) was a Polish-American logician ...
in his 1941 ''Introduction to Logic'' used the word "matrix" synonymously with the notion of
truth table A truth table is a mathematical table used in logic—specifically in connection with Boolean algebra, Boolean functions, and propositional calculus—which sets out the functional values of logical expressions on each of their functional arg ...
as used in mathematical logic.


See also

* List of named matrices * * Irregular matrix * *


Notes


References


Mathematical references

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * . * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *


Physics references

* * * * * * * * * * * * * * * *


Historical references

* , reprint of the 1907 original edition * ; reprinted in ''The collected mathematical papers of Arthur Cayley'', vol. II, Cambridge University Press, 1889
pp. 475–496
* * * * * * * * * * * * * ; reprint of 1946 corrected printing, Dover Publications, 1995, *


Further reading

* * *


External links

*



{{authority control