Matrix multiplication in the context of Special linear group


Matrix multiplication in the context of Special linear group

Matrix multiplication Study page number 1 of 1

Play TriviaQuestions Online!

or

Skip to study material about Matrix multiplication in the context of "Special linear group"


⭐ Core Definition: Matrix multiplication

In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the number of rows in the second matrix. The resulting matrix, known as the matrix product, has the number of rows of the first and the number of columns of the second matrix. The product of matrices A and B is denoted as AB.

Matrix multiplication was first described by the French mathematician Jacques Philippe Marie Binet in 1812, to represent the composition of linear maps that are represented by matrices. Matrix multiplication is thus a basic tool of linear algebra, and as such has numerous applications in many areas of mathematics, as well as in applied mathematics, statistics, physics, economics, and engineering.Computing matrix products is a central operation in all computational applications of linear algebra.

↓ Menu
HINT:

πŸ‘‰ Matrix multiplication in the context of Special linear group

In mathematics, the special linear group of degree over a commutative ring is the set of matrices with determinant , with the group operations of ordinary matrix multiplication and matrix inversion. This is the normal subgroup of the general linear group given by the kernel of the determinant

where is the multiplicative group of (that is, excluding when is a field).

↓ Explore More Topics
In this Dossier

Matrix multiplication in the context of Binary operation

In mathematics, a binary operation or dyadic operation is a rule for combining two elements (called operands) to produce another element. More formally, a binary operation is an operation of arity two.

More specifically, a binary operation on a set is a binary function that maps every pair of elements of the set to an element of the set. Examples include the familiar arithmetic operations like addition, subtraction, multiplication, set operations like union, complement, intersection. Other examples are readily found in different areas of mathematics, such as vector addition, matrix multiplication, and conjugation in groups.

View the full Wikipedia page for Binary operation
↑ Return to Menu

Matrix multiplication in the context of Product (mathematics)

In mathematics, a product is the result of multiplication, or an expression that identifies objects (numbers or variables) to be multiplied, called factors. For example, 21 is the product of 3 and 7 (the result of multiplication), and is the product of and (indicating that the two factors should be multiplied together).When one factor is an integer, the product is called a multiple.

The order in which real or complex numbers are multiplied has no bearing on the product; this is known as the commutative law of multiplication. When matrices or members of various other associative algebras are multiplied, the product usually depends on the order of the factors. Matrix multiplication, for example, is non-commutative, and so is multiplication in other algebras in general as well.

View the full Wikipedia page for Product (mathematics)
↑ Return to Menu

Matrix multiplication in the context of Matrix (mathematics)

In mathematics, a matrix (pl.: matrices) is a rectangular array of numbers or other mathematical objects with elements or entries arranged in rows and columns, usually satisfying certain properties of addition and multiplication.

For example, denotes a matrix with two rows and three columns. This is often referred to as a "two-by-three matrix", a 2 × 3 matrix, or a matrix of dimension 2 × 3.

View the full Wikipedia page for Matrix (mathematics)
↑ Return to Menu

Matrix multiplication in the context of Group representation

In the mathematical field of representation theory, group representations describe abstract groups in terms of bijective linear transformations of a vector space to itself (i.e. vector space automorphisms); in particular, they can be used to represent group elements as invertible matrices so that the group operation can be represented by matrix multiplication.

In chemistry, a group representation can relate mathematical group elements to symmetric rotations and reflections of molecules.

View the full Wikipedia page for Group representation
↑ Return to Menu

Matrix multiplication in the context of Triangular matrix

In mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called lower triangular if all the entries above the main diagonal are zero. Similarly, a square matrix is called upper triangular if all the entries below the main diagonal are zero.

Because matrix equations with triangular matrices are easier to solve, they are very important in numerical analysis. By the LU decomposition algorithm, an invertible matrix may be written as the product of a lower triangular matrix L and an upper triangular matrix U if and only if all its leading principal minors are non-zero.

View the full Wikipedia page for Triangular matrix
↑ Return to Menu

Matrix multiplication in the context of Associative algebra

In mathematics, an associative algebra A over a commutative ring (often a field) K is a ring A together with a ring homomorphism from K into the center of A. This is thus an algebraic structure with an addition, a multiplication, and a scalar multiplication (the multiplication by the image of the ring homomorphism of an element of K). The addition and multiplication operations together give A the structure of a ring; the addition and scalar multiplication operations together give A the structure of a module or vector space over K. In this article we will also use the term K-algebra to mean an associative algebra over K. A standard first example of a K-algebra is a ring of square matrices over a commutative ring K, with the usual matrix multiplication.

A commutative algebra is an associative algebra for which the multiplication is commutative, or, equivalently, an associative algebra that is also a commutative ring.

View the full Wikipedia page for Associative algebra
↑ Return to Menu

Matrix multiplication in the context of Algebra over a field

In mathematics, an algebra over a field (often simply called an algebra) is a vector space equipped with a bilinear product. Thus, an algebra is an algebraic structure consisting of a set together with operations of multiplication and addition and scalar multiplication by elements of a field and satisfying the axioms implied by "vector space" and "bilinear".

The multiplication operation in an algebra may or may not be associative, leading to the notions of associative algebras where associativity of multiplication is assumed, and non-associative algebras, where associativity is not assumed (but not excluded, either). Given an integer n, the ring of real square matrices of order n is an example of an associative algebra over the field of real numbers under matrix addition and matrix multiplication since matrix multiplication is associative. Three-dimensional Euclidean space with multiplication given by the vector cross product is an example of a nonassociative algebra over the field of real numbers since the vector cross product is nonassociative, satisfying the Jacobi identity instead.

View the full Wikipedia page for Algebra over a field
↑ Return to Menu

Matrix multiplication in the context of Representation theory

Representation theory is a branch of mathematics that studies abstract algebraic structures by representing their elements as linear transformations of vector spaces, and studies modules over these abstract algebraic structures. In essence, a representation makes an abstract algebraic object more concrete by describing its elements by matrices and their algebraic operations (for example, matrix addition, matrix multiplication).

The algebraic objects amenable to such a description include groups, associative algebras and Lie algebras. The most prominent of these (and historically the first) is the representation theory of groups, in which elements of a group are represented by invertible matrices such that the group operation is matrix multiplication.

View the full Wikipedia page for Representation theory
↑ Return to Menu

Matrix multiplication in the context of DFT matrix

In applied mathematics, a DFT matrix is a square matrix as an expression of a discrete Fourier transform (DFT) as a transformation matrix, which can be applied to a signal through matrix multiplication.

View the full Wikipedia page for DFT matrix
↑ Return to Menu

Matrix multiplication in the context of Linear algebraic group

In mathematics, a linear algebraic group is a subgroup of the group of invertible matrices (under matrix multiplication) that is defined by polynomial equations. An example is the orthogonal group, defined by the relation where is the transpose of .

Many Lie groups can be viewed as linear algebraic groups over the field of real or complex numbers. (For example, every compact Lie group can be regarded as a linear algebraic group over R (necessarily R-anisotropic and reductive), as can many noncompact groups such as the simple Lie group SL(n,R).) The simple Lie groups were classified by Wilhelm Killing and Γ‰lie Cartan in the 1880s and 1890s. At that time, no special use was made of the fact that the group structure can be defined by polynomials, that is, that these are algebraic groups. The founders of the theory of algebraic groups include Maurer, Chevalley, and KolchinΒ (1948). In the 1950s, Armand Borel constructed much of the theory of algebraic groups as it exists today.

View the full Wikipedia page for Linear algebraic group
↑ Return to Menu

Matrix multiplication in the context of Special orthogonal group

In mathematics, the orthogonal group in dimension n, denoted O(n), is the group of distance-preserving transformations of a Euclidean space of dimension n that preserve a fixed point, where the group operation is given by composing transformations. The orthogonal group is sometimes called the general orthogonal group, by analogy with the general linear group. Equivalently, it is the group of n Γ— n orthogonal matrices, where the group operation is given by matrix multiplication (an orthogonal matrix is a real matrix whose inverse equals its transpose). The orthogonal group is an algebraic group and a Lie group. It is compact.

The orthogonal group in dimension n has two connected components. The one that contains the identity element is a normal subgroup, called the special orthogonal group, and denoted SO(n). It consists of all orthogonal matrices of determinant 1. This group is also called the rotation group, generalizing the fact that in dimensions 2 and 3, its elements are the usual rotations around a point (in dimension 2) or a line (in dimension 3). In low dimension, these groups have been widely studied, see SO(2), SO(3) and SO(4). The other component consists of all orthogonal matrices of determinant βˆ’1. This component does not form a group, as the product of any two of its elements is of determinant 1, and therefore not an element of the component.

View the full Wikipedia page for Special orthogonal group
↑ Return to Menu

Matrix multiplication in the context of Unitary group

In mathematics, the unitary group of degree n, denoted U(n), is the group of n Γ— n unitary matrices, with the group operation of matrix multiplication. The unitary group is a subgroup of the general linear group GL(n, C), and it has as a subgroup the special unitary group, consisting of those unitary matrices with determinant 1.

In the simple case n = 1, the group U(1) corresponds to the circle group, isomorphic to the set of all complex numbers that have absolute value 1, under multiplication. All the unitary groups contain copies of this group.

View the full Wikipedia page for Unitary group
↑ Return to Menu

Matrix multiplication in the context of Linear group

In mathematics, a matrix group is a group G consisting of invertible matrices over a specified field K, with the operation of matrix multiplication. A linear group is a group that is isomorphic to a matrix group (that is, admitting a faithful, finite-dimensional representation over K).

Any finite group is linear, because it can be realized by permutation matrices using Cayley's theorem. Among infinite groups, linear groups form an interesting and tractable class. Examples of groups that are not linear include groups which are "too big" (for example, the group of permutations of an infinite set), or which exhibit some pathological behavior (for example, finitely generated infinite torsion groups).

View the full Wikipedia page for Linear group
↑ Return to Menu

Matrix multiplication in the context of Jacques Philippe Marie Binet

Jacques Philippe Marie Binet (French: [binΙ›]; 2 February 1786 – 12 May 1856) was a French mathematician, physicist and astronomer born in Rennes; he died in Paris, France, in 1856. He made significant contributions to number theory, and the mathematical foundations of matrix algebra which would later lead to important contributions by Cayley and others. In his memoir on the theory of the conjugate axis and of the moment of inertia of bodies he enumerated the principle now known as Binet's theorem. He is also recognized as the first to describe the rule for multiplying matrices in 1812, and Binet's formula expressing Fibonacci numbers in closed form is named in his honour, although the same result was known to Abraham de Moivre a century earlier.

View the full Wikipedia page for Jacques Philippe Marie Binet
↑ Return to Menu