1. Introduction
Block matrices arise in many fields of pure and applied mathematics and also applied sciences. Since these matrices are widely used in so many different fields, it is important to know about their algebraic properties. This paper aims to bring together the two fundamental concepts: inverses and determinants of
block matrices. Exercises on some inverse and determinant properties of block matrices can be found in standard linear algebra textbooks. However, in this study, we discuss these two concepts with a complete theory, hoping that it will serve as a primary reference for those interested in the subject or would like to use these matrices in their research. There are many published works in the literature concerning these two algebraic concepts. For example, the inverses of
block matrices have been studied by Lu and Shiou in [
1], the determinants of these matrices have been examined by Powell and Silvester in [
2] and [
3], respectively, and the generalized inverses and ranks for
block matrices are given in [
4]. Obtaining formula for the inverse and the determinant of a
block matrix is crucial, as these results can always be expanded to the block matrices of larger sizes by either splitting it into
blocks for the inverse case or by expressing the determinants in terms of the determinants of the
block matrices using the cofactor expansion in the determinant case. Therefore, in this paper, the two theories are first established for the
block matrices.
Block matrices are used in the proofs of many critical theorems in linear algebra. For example, the determinants of block upper triangular matrices are used to prove that the dimension of the eigenspace corresponding to an eigenvalue
is always less than or equal to the algebraic multiplicity of the given
. For a linear operator
T on a finite dimensional vector space
V, and for a
T-invariant subspace
W of
V, the proof of the theorem stating that the characteristic polynomial of
(
T restricted to
W) divides the characteristic polynomial of
T also uses the determinants of block matrices. This is an important result used in the proof of the famous Cayley–Hamilton theorem [
5]. Last but not least, the Jordan canonical form theory is also based on block diagonal matrices. Remember that a linear operator
T on
n dimensional vector space
V is invertible if and only if its standard matrix is invertible, and in fact, the matrix for the inverse operator is just the inverse of the standard matrix of
T [
5]. Therefore, the invertibility of a block matrix will give information on the invertibility of the so-called ‘block operators’; operators whose standard matrices are the block matrices (standard matrices are obtained by finding the images of the bases vectors of the given vector space
V under the linear operator). In functional analysis, these linear operators can help solve integral equations (see, for example, [
6], where the block operator diagrams are used, or [
7], where the strict positivity (invertibility and the positivity) of the operators on Hilbert spaces are identified with the strict positivity of the
block matrices). Other crucial problems under pure mathematics can also be solved using these linear operators. Hence, our paper can help to obtain information on them via their matrices.
Many algebraic operations can be simplified by using blocks in a matrix as elements instead of using matrices of much larger sizes. We give some references for the applications of block matrices in applied mathematics and general sciences. However, let us remind the reader that this is an algebraic paper, and the scope of it is not in the application direction. Generalised inverses of block matrices are used in [
8] to establish conditions for the values of two matrix functions to be equal, intuitionistic fuzzy block matrices (generalisations of subsets fuzzy block matrices) and their properties are studied in [
9], and the solution of linear systems using block tridiagonal coefficient matrices via the block cyclic reduction method is used to study the roots of the characteristic polynomials of matrices in [
10]. To give more examples on applications in the numerical analysis field, incomplete block-matrix factorisations are used to precondition the iterative methods in [
11], where applications for the solution of the Dirichlet problem of Laplace’s equation on a rectangle using the finite difference method with classical rectangular grids are given. Moreover, in [
12], an incomplete block factorisation for symmetric positive definite block tridiagonal matrices is given; this factorization is used to precondition the conjugate gradient method and is applied to solve the Dirichlet boundary value problem of the heat equation on a rectangle by developing a new difference method on a hexagonal grid.
The details on the inverse and determinant theories of block matrices are discussed in this paper. All proofs are independently produced by the authors unless explicitly stated and cited in this paper. For examples on the covered topics, one may refer to the Eastern Mediterranean University MSc Thesis of the second named author [
13]. The organization of this paper is as follows: in the next section, we start by giving the preliminaries. In
Section 3, we first give formulas for the inverses of
block diagonal and block triangular matrices; the techniques of proofs here can be generalised to block diagonal and block triangular matrices of higher dimensions. In the same section, we give the inverse formulas for the
block matrices in case one of the blocks is invertible. Proofs are provided via block Gaussian elimination and also block elementary matrices. LDU decompositions of the given block matrix are also provided in this section. Once the inverse formula for
block matrices is obtained, this can be generalised to
block matrices by diving it into four blocks (by producing a
block matrix).
Section 4 deals with the determinant concept in two different approaches. First, a special commutative case is revised, where the determinants of matrices with blocks belonging to a commutative subring of a field or a commutative ring are studied. The determinants of
block diagonal and block triangular matrices, together with the determinants of a general
block matrix, are provided. In this section, we also give a determinant formula for tensor products of two matrices. Next, the general formula existing in the literature is presented, which works for matrices in any ring or field. This formula can get very complex if the block matrix has a large size. Therefore, in specific useful cases, the determinants for
, and
block matrices are also provided. To demonstrate the
case, we also give a numerical example and compute the determinant of a
matrix. With this example, one can easily see the efficiency of the method compared to computing the determinant via a cofactor expansion or row reduction. Finally, we conclude this paper in
Section 5.
3. Inverses of Block Matrices
3.1. Inverses of Block Diagonal and Block Triangular Matrices
In this first section, we start by giving the inverses of block diagonal and block triangular matrices. The same techniques of the proofs can be applied to the block diagonal and block triangular matrices of larger sizes.
Proposition 1. If is a block diagonal matrix with square and invertible blocks , for , then D is invertible, and the inverse of D is given via Proof. One can easily observe that the multiplication of the two matrices will produce the block identity matrix. The Gauss elimination method is another way to see the result. □
Proposition 2. If is a block upper triangular matrix with square and invertible main blocks , for , then U is invertible and the inverse is given by Proof. Using the block Gauss Elimination, we obtain
□
Similarly, we provide the inverse of the lower triangular matrix below; the idea of the proof works in a very similar manner to the proof of Proposition 2.
Proposition 3. If is a block lower triangular matrix with square and invertible main blocks , for , then L is invertible and the inverse is given via 3.2. Inverses of Block Matrices
Assume that A is a non-singular square block matrix , and its inverse is Let and N be the partitioned matrices in A, with sizes , , , (with ), and and S be the submatrices in with sizes , , , and , respectively, for the multiplications to be compatible. We can verify ; here, we consider the following two cases. The case where all the blocks are square matrices (i.e., if ) is discussed in Remark 1.
Square matrices in diagonal positions in A and , implying and .
Square matrices in anti-diagonal positions of A and implying and .
The following two theorems deal with the first case where matrices in the diagonal positions are all square.
Theorem 1. Let T be non-singular. Then, exists if and only if the matrix is invertible and
Proof. First, we use the block Gauss elimination on block matrix A. The Gauss elimination method on the block matrices is not as straightforward as the one on standard matrices without blocks. At each stage of performing an elementary row operation, one must check that the sizes of the matrices are compatible.
Another way of proving this inverse is via realising that
A can be written as a product of four block elementary matrices. Note that
In
Section 3.1, the inverses of these block elementaries can easily be computed, which would give a second way of proving the theorem. Also, note that
, which would give the LDU decomposition of the matrix
A above. □
Theorem 2. Let N be non-singular. Then, exists if and only if the matrix is invertible, and Proof. Again, by using block Gauss elimination method, we obtain
As we illustrated above, we next express
A as a product of four block elementary matrices. These block elementary matrices can be observed by applying the inverses of the elementary row operations to the block identity matrix. Note that
From this equation, it is also very straightforward to observe . As we did above in the proof of Theorem 1, the two elementary matrices on the left can be multiplied to give a lower triangular matrix, which produces the LDU decomposition of the given matrix A. □
If the square blocks are now in the anti-diagonal positions of
A and
, a small trick can be applied to move these square blocks to the diagonal positions. Assume that
J is a matrix with 1’s in the reverse diagonal position and 0’s elsewhere. Then,
reverses the order of columns of
A, and
reverses the order of rows of
A.
Note that . Therefore, Theorems 1 and 2 can be used to obtain the inverse formulas for the block matrices when the square matrices are in the anti-diagonal positions.
Remark 1. Note that inverse formulas in Theorems 1 and 2 are equivalent, if T and N are both non-singular. Also note that if we have square matrices in all positions of A and , that is to say, if , then Theorems 1 and 2 and also the remaining two theorems coming from square matrices in anti-diagonal positions of A, must coincide and produce the same inverse formula, depending, of course, on the invertibility of the square blocks. This technique will always work by splitting a block matrix of any size into four blocks and using theorems above by considering the positions of the square blocks.
Remark 2. The inverses of the diagonal and triangular matrices in Section 3.1 can also be computed using Theorems 1 and 2.