HOME
*





Immanant
In mathematics, the immanant of a matrix (mathematics), matrix was defined by Dudley E. Littlewood and Archibald Read Richardson as a generalisation of the concepts of determinant and Permanent (mathematics), permanent. Let \lambda=(\lambda_1,\lambda_2,\ldots) be a Partition (number theory), partition of an integer n and let \chi_\lambda be the corresponding irreducible Group representation, representation-theoretic Character (group theory), character of the symmetric group S_n. The ''immanant'' of an n\times n Matrix (mathematics), matrix A=(a_) associated with the character \chi_\lambda is defined as the expression :\operatorname_\lambda(A)=\sum_ \chi_\lambda(\sigma) a_ a_ \cdots a_. Examples The determinant is a special case of the immanant, where \chi_\lambda is the alternating character \sgn, of ''S''''n'', defined by the parity of a permutation. The permanent is the case where \chi_\lambda is the trivial character, which is identically equal to 1. For example, for 3 ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Determinant
In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if and only if the matrix is invertible and the linear map represented by the matrix is an isomorphism. The determinant of a product of matrices is the product of their determinants (the preceding property is a corollary of this one). The determinant of a matrix is denoted , , or . The determinant of a matrix is :\begin a & b\\c & d \end=ad-bc, and the determinant of a matrix is : \begin a & b & c \\ d & e & f \\ g & h & i \end= aei + bfg + cdh - ceg - bdi - afh. The determinant of a matrix can be defined in several equivalent ways. Leibniz formula expresses the determinant as a sum of signed products of matrix entries such that each summand is the product of different entries, and the number of these summands is n!, the factorial of ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Archibald Read Richardson
Archibald Read Richardson FRS (21 August 1881 – 4 November 1954) was a British mathematician known for his work in algebra. Career Richardson collaborated with Dudley E. Littlewood on invariants and group representation theory. They introduced the immanant of a matrix, studied Schur functions and developed the Littlewood–Richardson rule for their multiplication. Awards and honours Richardson was elected a Fellow of the Royal Society Fellowship of the Royal Society (FRS, ForMemRS and HonFRS) is an award granted by the judges of the Royal Society of London to individuals who have made a "substantial contribution to the improvement of natural knowledge, including mathemat ... on 21 March 1946. See also * Quasideterminant References 1881 births 1954 deaths 20th-century British mathematicians Algebraists Fellows of the Royal Society {{UK-mathematician-stub ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Permanent (mathematics)
In linear algebra, the permanent of a square matrix is a function of the matrix similar to the determinant. The permanent, as well as the determinant, is a polynomial in the entries of the matrix. Both are special cases of a more general function of a matrix called the immanant. Definition The permanent of an matrix is defined as \operatorname(A)=\sum_\prod_^n a_. The sum here extends over all elements σ of the symmetric group ''S''''n''; i.e. over all permutations of the numbers 1, 2, ..., ''n''. For example, \operatorname\begina&b \\ c&d\end=ad+bc, and \operatorname\begina&b&c \\ d&e&f \\ g&h&i \end=aei + bfg + cdh + ceg + bdi + afh. The definition of the permanent of ''A'' differs from that of the determinant of ''A'' in that the signatures of the permutations are not taken into account. The permanent of a matrix A is denoted per ''A'', perm ''A'', or Per ''A'', sometimes with parentheses around the argument. Minc uses Per(''A'') for the permanent of rectangular m ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Matrix (mathematics)
In mathematics, a matrix (plural matrices) is a rectangular array or table of numbers, symbols, or expressions, arranged in rows and columns, which is used to represent a mathematical object or a property of such an object. For example, \begin1 & 9 & -13 \\20 & 5 & -6 \end is a matrix with two rows and three columns. This is often referred to as a "two by three matrix", a "-matrix", or a matrix of dimension . Without further specifications, matrices represent linear maps, and allow explicit computations in linear algebra. Therefore, the study of matrices is a large part of linear algebra, and most properties and operations of abstract linear algebra can be expressed in terms of matrices. For example, matrix multiplication represents composition of linear maps. Not all matrices are related to linear algebra. This is, in particular, the case in graph theory, of incidence matrices, and adjacency matrices. ''This article focuses on matrices related to linear algebra, and, un ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Multilinear Form
In abstract algebra and multilinear algebra, a multilinear form on a vector space V over a field K is a map :f\colon V^k \to K that is separately ''K''-linear in each of its ''k'' arguments. More generally, one can define multilinear forms on a module over a commutative ring. The rest of this article, however, will only consider multilinear forms on finite-dimensional vector spaces. A multilinear ''k''-form on V over \mathbf is called a (covariant) ''k''-tensor, and the vector space of such forms is usually denoted \mathcal^k(V) or \mathcal^k(V). Tensor product Given a ''k''-tensor f\in\mathcal^k(V) and an ''ℓ''-tensor g\in\mathcal^\ell(V), a product f\otimes g\in\mathcal^(V), known as the tensor product, can be defined by the property : (f\otimes g)(v_1,\ldots,v_k,v_,\ldots, v_)=f(v_1,\ldots,v_k)g(v_,\ldots, v_), for all v_1,\ldots,v_\in V. The tensor product of multilinear forms is not commutative; however it is bilinear and associative: : f\otimes(ag_1+bg_2)=a(f\ot ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Linear Algebra
Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices. Linear algebra is central to almost all areas of mathematics. For instance, linear algebra is fundamental in modern presentations of geometry, including for defining basic objects such as lines, planes and rotations. Also, functional analysis, a branch of mathematical analysis, may be viewed as the application of linear algebra to spaces of functions. Linear algebra is also used in most sciences and fields of engineering, because it allows modeling many natural phenomena, and computing efficiently with such models. For nonlinear systems, which cannot be modeled with linear algebra, it is often used for dealing with first-order approximations, using the fact that the differential of a multivariate function at a point is the linea ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Algebra
Algebra () is one of the broad areas of mathematics. Roughly speaking, algebra is the study of mathematical symbols and the rules for manipulating these symbols in formulas; it is a unifying thread of almost all of mathematics. Elementary algebra deals with the manipulation of variables (commonly represented by Roman letters) as if they were numbers and is therefore essential in all applications of mathematics. Abstract algebra is the name given, mostly in education, to the study of algebraic structures such as groups, rings, and fields (the term is no more in common use outside educational context). Linear algebra, which deals with linear equations and linear mappings, is used for modern presentations of geometry, and has many practical applications (in weather forecasting, for example). There are many areas of mathematics that belong to algebra, some having "algebra" in their name, such as commutative algebra, and some not, such as Galois theory. The word ''alge ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




Philosophical Transactions Of The Royal Society A
''Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences'' is a fortnightly peer-reviewed scientific journal published by the Royal Society. It publishes original research and review content in a wide range of physical scientific disciplines. Articles can be accessed online a few months prior to the printed journal. All articles become freely accessible two years after their publication date. The current editor-in-chief is John Dainton. Overview ''Philosophical Transactions of the Royal Society A'' publishes themed journal issues on topics of current scientific importance and general interest within the physical, mathematical and engineering sciences, edited by leading authorities and comprising original research, reviews and opinions from prominent researchers. Past issue titles include "Supercritical fluids - green solvents for green chemistry?", "Tsunamis: Bridging science, engineering and society", "Spatial transformations: from ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Gamas's Theorem
Gamas's theorem is a result in multilinear algebra which states the necessary and sufficient conditions for a tensor symmetrized by an irreducible representation of the symmetric group S_n to be zero. It was proven in 1988 by Carlos Gamas. Additional proofs have been given by Pate and Berget. Statement of the theorem Let V be a finite-dimensional complex vector space and \lambda be a partition of n. From the representation theory of the symmetric group S_n it is known that the partition \lambda corresponds to an irreducible representation of S_n. Let \chi^ be the character of this representation. The tensor v_1 \otimes v_2 \otimes \dots \otimes v_n \in V^ symmetrized by \chi^ is defined to be \frac \sum_ \chi^(\sigma) v_ \otimes v_ \otimes \dots \otimes v_, where e is the identity element of S_n. Gamas's theorem states that the above symmetrized tensor is non-zero if and only if it is possible to partition the set of vectors \{v_i\} into linearly independent sets whose ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Gram Matrix
In linear algebra, the Gram matrix (or Gramian matrix, Gramian) of a set of vectors v_1,\dots, v_n in an inner product space is the Hermitian matrix of inner products, whose entries are given by the inner product G_ = \left\langle v_i, v_j \right\rangle., p.441, Theorem 7.2.10 If the vectors v_1,\dots, v_n are the columns of matrix X then the Gram matrix is X^* X in the general case that the vector coordinates are complex numbers, which simplifies to X^\top X for the case that the vector coordinates are real numbers. An important application is to compute linear independence: a set of vectors are linearly independent if and only if the Gram determinant (the determinant of the Gram matrix) is non-zero. It is named after Jørgen Pedersen Gram. Examples For finite-dimensional real vectors in \mathbb^n with the usual Euclidean dot product, the Gram matrix is G = V^\top V, where V is a matrix whose columns are the vectors v_k and V^\top is its transpose whose rows are the vectors v_k ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Representation Theory Of The Symmetric Group
In mathematics, the representation theory of the symmetric group is a particular case of the representation theory of finite groups, for which a concrete and detailed theory can be obtained. This has a large area of potential applications, from symmetric function theory to quantum chemistry studies of atoms, molecules and solids. The symmetric group S''n'' has order ''n''!. Its conjugacy classes are labeled by partitions of ''n''. Therefore according to the representation theory of a finite group, the number of inequivalent irreducible representations, over the complex numbers, is equal to the number of partitions of ''n''. Unlike the general situation for finite groups, there is in fact a natural way to parametrize irreducible representations by the same set that parametrizes conjugacy classes, namely by partitions of ''n'' or equivalently Young diagrams of size ''n''. Each such irreducible representation can in fact be realized over the integers (every permutation acting by ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Schur Polynomial
In mathematics, Schur polynomials, named after Issai Schur, are certain symmetric polynomials in ''n'' variables, indexed by partitions, that generalize the elementary symmetric polynomials and the complete homogeneous symmetric polynomials. In representation theory they are the characters of polynomial irreducible representations of the general linear groups. The Schur polynomials form a linear basis for the space of all symmetric polynomials. Any product of Schur polynomials can be written as a linear combination of Schur polynomials with non-negative integral coefficients; the values of these coefficients is given combinatorially by the Littlewood–Richardson rule. More generally, skew Schur polynomials are associated with pairs of partitions and have similar properties to Schur polynomials. Definition (Jacobi's bialternant formula) Schur polynomials are indexed by integer partitions. Given a partition , where , and each is a non-negative integer, the functions a_ (x_ ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]