Matrix Exponential
In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives the exponential map (Lie theory), exponential map between a matrix Lie algebra and the corresponding Lie group. Let be an real number, real or complex number, complex matrix (mathematics), matrix. The exponential of , denoted by or , is the matrix given by the power series e^X = \sum_^\infty \frac X^k where X^0 is defined to be the identity matrix I with the same dimensions as X, and . The series always converges, so the exponential of is well-defined. Equivalently, e^X = \lim_ \left(I + \frac \right)^k for integer-valued , where is the identity matrix. Equivalently, given by the solution to the differential equation \frac d e^ = X e^, \quad e^ = I When is an diagonal matrix then will be an diagonal matr ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Mathematics
Mathematics is a field of study that discovers and organizes methods, Mathematical theory, theories and theorems that are developed and Mathematical proof, proved for the needs of empirical sciences and mathematics itself. There are many areas of mathematics, which include number theory (the study of numbers), algebra (the study of formulas and related structures), geometry (the study of shapes and spaces that contain them), Mathematical analysis, analysis (the study of continuous changes), and set theory (presently used as a foundation for all mathematics). Mathematics involves the description and manipulation of mathematical object, abstract objects that consist of either abstraction (mathematics), abstractions from nature orin modern mathematicspurely abstract entities that are stipulated to have certain properties, called axioms. Mathematics uses pure reason to proof (mathematics), prove properties of objects, a ''proof'' consisting of a succession of applications of in ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Invertible Matrix
In linear algebra, an invertible matrix (''non-singular'', ''non-degenarate'' or ''regular'') is a square matrix that has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by an inverse to undo the operation. An invertible matrix multiplied by its inverse yields the identity matrix. Invertible matrices are the same size as their inverse. Definition An -by- square matrix is called invertible if there exists an -by- square matrix such that\mathbf = \mathbf = \mathbf_n ,where denotes the -by- identity matrix and the multiplication used is ordinary matrix multiplication. If this is the case, then the matrix is uniquely determined by , and is called the (multiplicative) ''inverse'' of , denoted by . Matrix inversion is the process of finding the matrix which when multiplied by the original matrix gives the identity matrix. Over a field, a square matrix that is ''not'' invertible is called singular or deg ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Jacobi's Formula
In matrix calculus, Jacobi's formula expresses the derivative of the determinant of a matrix ''A'' in terms of the adjugate of ''A'' and the derivative of ''A''., Part Three, Section 8.3 If is a differentiable map from the real numbers to matrices, then : \frac \det A(t) = \operatorname \left (\operatorname(A(t)) \, \frac\right ) = \left(\det A(t) \right) \cdot \operatorname \left (A(t)^ \cdot \, \frac\right ) where is the trace of the matrix and \operatorname(X) is its adjugate matrix. (The latter equality only holds if ''A''(''t'') is invertible.) As a special case, : = \operatorname(A)_. Equivalently, if stands for the differential of , the general formula is : d \det (A) = \operatorname (\operatorname(A) \, dA) = \det (A) \operatorname \left (A^ d A\right ) The formula is named after the mathematician Carl Jacobi. Derivation Via matrix computation Theorem. (Jacobi's formula) For any differentiable map ''A'' from the real numbers to ''n'' × ''n'' matr ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Magnus Series
In mathematics and physics, the Magnus expansion, named after Wilhelm Magnus (1907–1990), provides an exponential representation of the product integral solution of a first-order homogeneous linear differential equation for a linear operator. In particular, it furnishes the fundamental matrix of a system of linear ordinary differential equations of order with varying coefficients. The exponent is aggregated as an infinite series, whose terms involve multiple integrals and nested commutators. The deterministic case Magnus approach and its interpretation Given the coefficient matrix , one wishes to solve the initial-value problem associated with the linear ordinary differential equation : Y'(t) = A(t) Y(t), \quad Y(t_0) = Y_0 for the unknown -dimensional vector function . When ''n'' = 1, the solution is given as a product integral : Y(t) = \exp \left( \int_^t A(s)\,ds \right) Y_0. This is still valid for ''n'' > 1 if the matrix satisfies for a ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Applications
Application may refer to: Mathematics and computing * Application software, computer software designed to help the user to perform specific tasks ** Application layer, an abstraction layer that specifies protocols and interface methods used in a communications network * Function application, in mathematics and computer science Processes and documents * Application for employment, a form or forms that an individual seeking employment must fill out * College application College application is the process by which individuals apply to gain entry into a college or university. Although specific details vary by country and institution, applications generally require basic background information of the applicant, such ..., the process by which prospective students apply for entry into a college or university * Patent application, a document filed at a patent office to support the grant of a patent Other uses * Application (virtue), a characteristic encapsulated in diligence * Topic ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Ordinary Differential Equations
In mathematics, an ordinary differential equation (ODE) is a differential equation (DE) dependent on only a single independent variable. As with any other DE, its unknown(s) consists of one (or more) function(s) and involves the derivatives of those functions. The term "ordinary" is used in contrast with ''partial'' differential equations (PDEs) which may be with respect to one independent variable, and, less commonly, in contrast with ''stochastic'' differential equations (SDEs) where the progression is random. Differential equations A linear differential equation is a differential equation that is defined by a linear polynomial in the unknown function and its derivatives, that is an equation of the form :a_0(x)y +a_1(x)y' + a_2(x)y'' +\cdots +a_n(x)y^+b(x)=0, where a_0(x),\ldots,a_n(x) and b(x) are arbitrary differentiable functions that do not need to be linear, and y',\ldots, y^ are the successive derivatives of the unknown function y of the variable x. Among ord ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Resolvent Formalism
In mathematics, the resolvent formalism is a technique for applying concepts from complex analysis to the study of the spectrum of operators on Banach spaces and more general spaces. Formal justification for the manipulations can be found in the framework of holomorphic functional calculus. The resolvent captures the spectral properties of an operator in the analytic structure of the functional. Given an operator , the resolvent may be defined as : R(z;A)= (A-zI)^~. Among other uses, the resolvent may be used to solve the inhomogeneous Fredholm integral equations; a commonly used approach is a series solution, the Liouville–Neumann series. The resolvent of can be used to directly obtain information about the spectral decomposition of . For example, suppose is an isolated eigenvalue in the spectrum of . That is, suppose there exists a simple closed curve C_\lambda in the complex plane that separates from the rest of the spectrum of . Then the residue : -\frac ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Laplace Transform
In mathematics, the Laplace transform, named after Pierre-Simon Laplace (), is an integral transform that converts a Function (mathematics), function of a Real number, real Variable (mathematics), variable (usually t, in the ''time domain'') to a function of a Complex number, complex variable s (in the complex-valued frequency domain, also known as ''s''-domain, or ''s''-plane). The transform is useful for converting derivative, differentiation and integral, integration in the time domain into much easier multiplication and Division (mathematics), division in the Laplace domain (analogous to how logarithms are useful for simplifying multiplication and division into addition and subtraction). This gives the transform many applications in science and engineering, mostly as a tool for solving linear differential equations and dynamical systems by simplifying ordinary differential equations and integral equations into algebraic equation, algebraic polynomial equations, and by simplifyin ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Unitary Matrix
In linear algebra, an invertible complex square matrix is unitary if its matrix inverse equals its conjugate transpose , that is, if U^* U = UU^* = I, where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose is referred to as the Hermitian adjoint of a matrix and is denoted by a dagger (), so the equation above is written U^\dagger U = UU^\dagger = I. A complex matrix is special unitary if it is unitary and its matrix determinant equals . For real numbers, the analogue of a unitary matrix is an orthogonal matrix. Unitary matrices have significant importance in quantum mechanics because they preserve norms, and thus, probability amplitudes. Properties For any unitary matrix of finite size, the following hold: * Given two complex vectors and , multiplication by preserves their inner product; that is, . * is normal (U^* U = UU^*). * is diagonalizable; that is, is unitarily similar to a diagonal matrix, as a consequence of ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Skew-Hermitian Matrix
__NOTOC__ In linear algebra, a square matrix with complex entries is said to be skew-Hermitian or anti-Hermitian if its conjugate transpose is the negative of the original matrix. That is, the matrix A is skew-Hermitian if it satisfies the relation where A^\textsf denotes the conjugate transpose of the matrix A. In component form, this means that for all indices i and j, where a_ is the element in the i-th row and j-th column of A, and the overline denotes complex conjugation. Skew-Hermitian matrices can be understood as the complex versions of real skew-symmetric matrices, or as the matrix analogue of the purely imaginary numbers., §4.1.2 The set of all skew-Hermitian n \times n matrices forms the u(n) Lie algebra, which corresponds to the Lie group U(n). The concept can be generalized to include linear transformations of any complex vector space with a sesquilinear norm. Note that the adjoint of an operator depends on the scalar product considered on the n dimension ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Hermitian Matrix
In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the -th row and -th column, for all indices and : A \text \quad \iff \quad a_ = \overline or in matrix form: A \text \quad \iff \quad A = \overline . Hermitian matrices can be understood as the complex extension of real symmetric matrices. If the conjugate transpose of a matrix A is denoted by A^\mathsf, then the Hermitian property can be written concisely as A \text \quad \iff \quad A = A^\mathsf Hermitian matrices are named after Charles Hermite, who demonstrated in 1855 that matrices of this form share a property with real symmetric matrices of always having real eigenvalues. Other, equivalent notations in common use are A^\mathsf = A^\dagger = A^\ast, although in quantum mechanics, A^\ast typically means the complex conjugate onl ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Orthogonal Matrix
In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identity matrix. This leads to the equivalent characterization: a matrix is orthogonal if its transpose is equal to its inverse: Q^\mathrm=Q^, where is the inverse of . An orthogonal matrix is necessarily invertible (with inverse ), unitary (), where is the Hermitian adjoint ( conjugate transpose) of , and therefore normal () over the real numbers. The determinant of any orthogonal matrix is either +1 or −1. As a linear transformation, an orthogonal matrix preserves the inner product of vectors, and therefore acts as an isometry of Euclidean space, such as a rotation, reflection or rotoreflection. In other words, it is a unitary transformation. The set of orthogonal matrices, under multiplication, forms the group , known as th ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |