Dual Space
In mathematics, any vector space ''V'' has a corresponding dual vector space (or just dual space for short) consisting of all linear forms on ''V'', together with the vector space structure of pointwise addition and scalar multiplication by constants. The dual space as defined above is defined for all vector spaces, and to avoid ambiguity may also be called the . When defined for a topological vector space, there is a subspace of the dual space, corresponding to continuous linear functionals, called the ''continuous dual space''. Dual vector spaces find application in many branches of mathematics that use vector spaces, such as in tensor analysis with finitedimensional vector spaces. When applied to vector spaces of functions (which are typically infinitedimensional), dual spaces are used to describe measures, distributions, and Hilbert spaces. Consequently, the dual space is an important concept in functional analysis. Early terms for ''dual'' include ''polarer Raum'' ah ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Mathematics
Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics with the major subdisciplines of number theory, algebra, geometry, and mathematical analysis, analysis, respectively. There is no general consensus among mathematicians about a common definition for their academic discipline. Most mathematical activity involves the discovery of properties of mathematical object, abstract objects and the use of pure reason to proof (mathematics), prove them. These objects consist of either abstraction (mathematics), abstractions from nature orin modern mathematicsentities that are stipulated to have certain properties, called axioms. A ''proof'' consists of a succession of applications of inference rule, deductive rules to already established results. These results include previously proved theorems, axioms ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Quantum Mechanics
Quantum mechanics is a fundamental theory in physics that provides a description of the physical properties of nature at the scale of atoms and subatomic particles. It is the foundation of all quantum physics including quantum chemistry, quantum field theory, quantum technology, and quantum information science. Classical physics, the collection of theories that existed before the advent of quantum mechanics, describes many aspects of nature at an ordinary ( macroscopic) scale, but is not sufficient for describing them at small (atomic and subatomic) scales. Most theories in classical physics can be derived from quantum mechanics as an approximation valid at large (macroscopic) scale. Quantum mechanics differs from classical physics in that energy, momentum, angular momentum, and other quantities of a bound system are restricted to discrete values ( quantization); objects have characteristics of both particles and waves ( wave–particle duality); and there are lim ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Dimension (vector Space)
In mathematics, the dimension of a vector space ''V'' is the cardinality (i.e., the number of vectors) of a basis of ''V'' over its base field. p. 44, §2.36 It is sometimes called Hamel dimension (after Georg Hamel) or algebraic dimension to distinguish it from other types of dimension. For every vector space there exists a basis, and all bases of a vector space have equal cardinality; as a result, the dimension of a vector space is uniquely defined. We say V is if the dimension of V is finite, and if its dimension is infinite. The dimension of the vector space V over the field F can be written as \dim_F(V) or as : F read "dimension of V over F". When F can be inferred from context, \dim(V) is typically written. Examples The vector space \R^3 has \left\ as a standard basis, and therefore \dim_(\R^3) = 3. More generally, \dim_(\R^n) = n, and even more generally, \dim_(F^n) = n for any field F. The complex numbers \Complex are both a real and complex vector spac ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Sequence
In mathematics, a sequence is an enumerated collection of objects in which repetitions are allowed and order matters. Like a set, it contains members (also called ''elements'', or ''terms''). The number of elements (possibly infinite) is called the ''length'' of the sequence. Unlike a set, the same elements can appear multiple times at different positions in a sequence, and unlike a set, the order does matter. Formally, a sequence can be defined as a function from natural numbers (the positions of elements in the sequence) to the elements at each position. The notion of a sequence can be generalized to an indexed family, defined as a function from an ''arbitrary'' index set. For example, (M, A, R, Y) is a sequence of letters with the letter 'M' first and 'Y' last. This sequence differs from (A, R, M, Y). Also, the sequence (1, 1, 2, 3, 5, 8), which contains the number 1 at two different positions, is a valid sequence. Sequences can be '' finite'', as in these examples, or '' in ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Linearly Independent
In the theory of vector spaces, a set of vectors is said to be if there is a nontrivial linear combination of the vectors that equals the zero vector. If no such linear combination exists, then the vectors are said to be . These concepts are central to the definition of dimension. A vector space can be of finite dimension or infinite dimension depending on the maximum number of linearly independent vectors. The definition of linear dependence and the ability to determine whether a subset of vectors in a vector space is linearly dependent are central to determining the dimension of a vector space. Definition A sequence of vectors \mathbf_1, \mathbf_2, \dots, \mathbf_k from a vector space is said to be ''linearly dependent'', if there exist scalars a_1, a_2, \dots, a_k, not all zero, such that :a_1\mathbf_1 + a_2\mathbf_2 + \cdots + a_k\mathbf_k = \mathbf, where \mathbf denotes the zero vector. This implies that at least one of the scalars is nonzero, say a_1\ne 0, and ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Axiom Of Choice
In mathematics, the axiom of choice, or AC, is an axiom of set theory equivalent to the statement that ''a Cartesian product of a collection of nonempty sets is nonempty''. Informally put, the axiom of choice says that given any collection of sets, each containing at least one element, it is possible to construct a new set by arbitrarily choosing one element from each set, even if the collection is infinite. Formally, it states that for every indexed family (S_i)_ of nonempty sets, there exists an indexed set (x_i)_ such that x_i \in S_i for every i \in I. The axiom of choice was formulated in 1904 by Ernst Zermelo in order to formalize his proof of the wellordering theorem. In many cases, a set arising from choosing elements arbitrarily can be made without invoking the axiom of choice; this is, in particular, the case if the number of sets from which to choose the elements is finite, or if a canonical rule on how to choose the elements is available – some distinguis ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Basis (linear Algebra)
In mathematics, a set of vectors in a vector space is called a basis if every element of may be written in a unique way as a finite linear combination of elements of . The coefficients of this linear combination are referred to as components or coordinates of the vector with respect to . The elements of a basis are called . Equivalently, a set is a basis if its elements are linearly independent and every element of is a linear combination of elements of . In other words, a basis is a linearly independent spanning set. A vector space can have several bases; however all the bases have the same number of elements, called the ''dimension'' of the vector space. This article deals mainly with finitedimensional vector spaces. However, many of the principles are also valid for infinitedimensional vector spaces. Definition A basis of a vector space over a field (such as the real numbers or the complex numbers ) is a linearly independent subset of that spans . Th ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Vector (geometric)
In mathematics, physics, and engineering, a Euclidean vector or simply a vector (sometimes called a geometric vector or spatial vector) is a geometric object that has magnitude (or length) and direction. Vectors can be added to other vectors according to vector algebra. A Euclidean vector is frequently represented by a '' directed line segment'', or graphically as an arrow connecting an ''initial point'' ''A'' with a ''terminal point'' ''B'', and denoted by \overrightarrow . A vector is what is needed to "carry" the point ''A'' to the point ''B''; the Latin word ''vector'' means "carrier". It was first used by 18th century astronomers investigating planetary revolution around the Sun. The magnitude of the vector is the distance between the two points, and the direction refers to the direction of displacement from ''A'' to ''B''. Many algebraic operations on real numbers such as addition, subtraction, multiplication, and negation have close analogues for vectors, operat ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Matrix Multiplication
In mathematics, particularly in linear algebra, matrix multiplication is a binary operation that produces a matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the number of rows in the second matrix. The resulting matrix, known as the matrix product, has the number of rows of the first and the number of columns of the second matrix. The product of matrices and is denoted as . Matrix multiplication was first described by the French mathematician Jacques Philippe Marie Binet in 1812, to represent the composition of linear maps that are represented by matrices. Matrix multiplication is thus a basic tool of linear algebra, and as such has numerous applications in many areas of mathematics, as well as in applied mathematics, statistics, physics, economics, and engineering. Computing matrix products is a central operation in all computational applications of linear algebra. Notation This article will use the following notat ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Real Number
In mathematics, a real number is a number that can be used to measure a ''continuous'' onedimensional quantity such as a distance, duration or temperature. Here, ''continuous'' means that values can have arbitrarily small variations. Every real number can be almost uniquely represented by an infinite decimal expansion. The real numbers are fundamental in calculus (and more generally in all mathematics), in particular by their role in the classical definitions of limits, continuity and derivatives. The set of real numbers is denoted or \mathbb and is sometimes called "the reals". The adjective ''real'' in this context was introduced in the 17th century by René Descartes to distinguish real numbers, associated with physical reality, from imaginary numbers (such as the square roots of ), which seemed like a theoretical contrivance unrelated to physical reality. The real numbers include the rational numbers, such as the integer and the fraction . The rest of the rea ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Kronecker Delta
In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just nonnegative integers. The function is 1 if the variables are equal, and 0 otherwise: \delta_ = \begin 0 &\text i \neq j, \\ 1 &\text i=j. \end or with use of Iverson brackets: \delta_ = =j, where the Kronecker delta is a piecewise function of variables and . For example, , whereas . The Kronecker delta appears naturally in many areas of mathematics, physics and engineering, as a means of compactly expressing its definition above. In linear algebra, the identity matrix has entries equal to the Kronecker delta: I_ = \delta_ where and take the values , and the inner product of vectors can be written as \mathbf\cdot\mathbf = \sum_^n a_\delta_b_ = \sum_^n a_ b_. Here the Euclidean vectors are defined as tuples: \mathbf = (a_1, a_2, \dots, a_n) and \mathbf= (b_1, b_2, ..., b_n) and the last step is obtained by using the values of the Kronecker del ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 

Dual Basis
In linear algebra, given a vector space ''V'' with a basis ''B'' of vectors indexed by an index set ''I'' (the cardinality of ''I'' is the dimension of ''V''), the dual set of ''B'' is a set ''B''∗ of vectors in the dual space ''V''∗ with the same index set ''I'' such that ''B'' and ''B''∗ form a biorthogonal system. The dual set is always linearly independent but does not necessarily span ''V''∗. If it does span ''V''∗, then ''B''∗ is called the dual basis or reciprocal basis for the basis ''B''. Denoting the indexed vector sets as B = \_ and B^ = \_, being biorthogonal means that the elements pair to have an inner product equal to 1 if the indexes are equal, and equal to 0 otherwise. Symbolically, evaluating a dual vector in ''V''∗ on a vector in the original space ''V'': : v^i\cdot v_j = \delta^i_j = \begin 1 & \text i = j\\ 0 & \text i \ne j\text \end where \delta^i_j is the Kronecker delta symbol. Introduction To perform operations with a vector, we must ... [...More Info...] [...Related Items...] OR: [Wikipedia] [Google] [Baidu] 