# Linear Algebra/Introduction to Matrices and Determinants

The determinant is a function which associates to a square matrix an element of the field on which it is defined (commonly the real or complex numbers).

## Contents

## Matrices[edit]

Informally an *m*×*n* matrix (plural matrices) is a rectangular table of entries from a field (that is to say that each entry is an element of a field). Here m is the number of rows and n the number of the columns in the table. Those unfamiliar with the concept of a field, can for now assume that by a field of characteristic 0 (which we will denote by F) we are referring to a particular subset of the set of complex numbers.

An m×n matrix (read as m by n matrix), is usually written as:

The row is an element of , showing the n components . Similary the column is an element of showing the m components .

Here m and n are called the *dimensions* of the matrix. The dimensions of a matrix are always given with the number of rows first, then the number of columns. It is also said that an m by n matrix has an *order* of m×n.

Formally, an m×n matrix M is a function where A = {1,2...m} × {1,2...n} and F is the field under consideration. It is almost always better to visualize a matrix as a rectangular table (or array) then as a function.

A matrix having only one row is called a *row matrix* (or a *row vector*) and a matrix having only one column is called a *column matrix* (or a *column vector*). Two matrices of the same order whose corresponding entries are equal are considered equal. The (*i*,*j*)-entry of the matrix (often written as or ) is the element at the intersection of the row (from the top) and the column (from the left).

For example,

is a 3×3 matrix (said 3 by 3). The 2nd row is and the 3rd column is . The (2,3) entry is the entry at intersection of the 2nd row and the 3rd column, that is 11.

Some special kinds of matrices are:

- A
*square matrix*is a matrix which has the same number of rows and columns. A*diagonal matrix*is a matrix with non zero entries only on the main diagonal (ie at positions).

- The
*unit matrix*or*identity matrix*I_{n}, is the matrix with elements on the diagonal set to 1 and all other elements set to 0. Mathematically, we may say that for the identity matrix (which is usually written as and called Kronecker's delta) is given by:

For example, if *n* = 3:

- The
*transpose*of an m-by-n matrix A is the n-by-m matrix A^{T}formed by turning rows into columns and columns into rows, i.e. . An example is

- A square matrix whose transpose is equal to itself is called a
*symmetric matrix*; that is, A is symmetric if . An example is

- A square matrix whose transpose is equal to its negative is called
*skew-symmetric matrix*; that is, A is skew-symmetric if . An example is

Properties of these matrices are developed in the exercises.

## Determinants[edit]

To define a determinant of order n, suppose there are n^{2} elements of a field s_{ij} where i and j are less than or equal to n. Define the following function (this function is important in the definition):

S(a_{1},a_{2},a_{3},...,a_{n})=# of reversals, meaning the number of times a_{n1}<a_{n2} when n_{1}>n_{2}, for each possible combination.

Suppose you have a permutation of numbers from 1 to n {a_{1},a_{2},a_{3},...,a_{n}). Then define a term of the determinant to be equal to (-1)^{S(a1,a2,a3,...,an)}s_{1a1},s_{2a2},s_{3a3},...,s_{nan}. The sum of all possible terms (i. e. through all possible permutations) is called the determinant.

## Theorem[edit]

Definition: The transpose of a matrix A, A^{T} is the matrix resulting when the columns and rows are interchanged i. e. the matrix s_{ji} when A is the matrix s_{ij}
A matrix and its transpose have the same determinant:

### Proof[edit]

All terms are the same, and the signs of the terms are also unchanged since all reversals remain reversals. Thus, the sum is the same.

## Theorem[edit]

Interchanging two rows (or columns) changes the sign of the determinant:

- .

### Proof[edit]

To show this, suppose two adjacent rows (or columns) are interchanged. Then any reversals in a term would not be affected except for the reversal of the elements of that term within that row (or column), in which case adds or subtracts a reversal, thus changing the signs of all terms, and thus the sign of the matrix. Now, if two rows, the a^{th} row and the (a+n)^{th} are interchanged, then interchange successively the a^{th} row and (a+1)^{th} row, and then the (a+1)^{th} row and (a+2)^{th} row, and continue in this fashion until one reaches the (a+n-1)^{th} row. Then go backwards until one goes back to the a^{th} row. This has the same effect as switching the a^{th} row and the (a+n)^{th} rows, and takes n-1 switches for going forwards, and n-2 switches for going backwards, and their sum must then be an odd number, so it multiplies by -1 an odd number of times, so that its total effect is to multiply by -1.

### Corollary[edit]

A determinant with two rows (or columns) that are the same has the value 0. Proof: This determinant would be the additive inverse of itself since interchanging the rows (or columns) does not change the determinant, but still changes the sign of the determinant. The only number for which it is possible is when it is equal to 0.

## Theorem[edit]

It is linear on the rows and columns of the matrix.

### Proof[edit]

The terms are of the form a_{1}......a_{n}. Using the distributive law of fields, this comes out to be a_{1}......a_{n} + a_{1}......a_{n}, an thus its sum of such terms is the sum of the two determinants:

### Corollary[edit]

Adding a row (or column) times a number to another row (or column) does not affect the value of a determinant.

#### Proof[edit]

Suppose you have a determinant A with the k^{th} column added by another column times a number:

where a_{kb} are elements of another column. By the linear property, this is equal to

The second number is equal to 0 because it has two columns that are the same. Thus, it is equal to

which is the same as the matrix A.

- It is easy to see that and thus

- for all -by- matrices and all scalars .

- A matrix over a commutative ring
*R*is invertible if and only if its determinant is a unit in*R*. In particular, if*A*is a matrix over a field such as the real or complex numbers, then*A*is invertible if and only if det(*A*) is not zero. In this case we have

Expressed differently: the vectors *v*_{1},...,*v*_{n} in **R**^{n} form a basis if and only if det(*v*_{1},...,*v*_{n}) is non-zero.

The determinants of a complex matrix and of its conjugate transpose are conjugate: