A Matrix is an Orthogonal Matrix when the product of a matrix and its transpose gives an identity value. An orthogonal matrix is a square matrix where transpose of Square Matrix is also the inverse of Square Matrix.
Orthogonal Matrix in Linear Algebra is a type of matrices in which the transpose of matrix is equal to the inverse of that matrix. As we know, the transpose of a matrix is obtained by swapping its row elements with its column elements. For an orthogonal matrix, the product of the transpose and the matrix itself is the identity matrix, as the transpose also serves as the inverse of the matrix.
Let's know more about Orthogonal Matrix in detail below.

Orthogonal Matrix
Orthogonal Matrix is a square matrix in which all rows and columns are mutually orthogonal unit vectors, meaning that each row and column of the matrix is perpendicular to every other row and column, and each row or column has a magnitude of 1.
Orthogonal Matrix Definition
A Matrix is called Orthogonal Matrix when the transpose of Matrix is inverse of that matrix or the product of Matrix and it's transpose is equal to an Identity Matrix.
Mathematically, an n x n matrix A is considered orthogonal if
AT = A-1
OR
AAT = ATA = I
Where,
- AT is the transpose of the square matrix,
- A-1 is the inverse of the square matrix, and
- I is the identity matrix of same order as A.
AT = A-1 (Condition for an Orthogonal matrix)...(i)
Pre-multiply by A on both sides,
We get, AAT = AA-1,
We know this relation of the identity matrix, AA-1 = I, (of the same order as A).
So we can also write it as AAT = I. (From (i))
Similarly, we can derive the relation ATA = I.
So, from the above two equations, we get AAT = ATA = I.
Condition for Orthogonal Matrix
For any matrix to be an orthogonal Matrix, it needs to fulfil the following conditions:
- Every two rows and two columns have a dot product of zero, and
- Every row and every column has a magnitude of one.
Orthogonal Matrix in Linear Algebra
The condition of any two vectors to be orthogonal is when their dot product is zero. Similarly, in the case of an orthogonal matrix, every two rows and every two columns are orthogonal. Also, one more condition is that the length of every row (vector) or column (vector) is 1.
For Example, let's consider a 3×3 matrix, i.e., A = \begin{bmatrix} \frac{1}{3} & \frac{2}{3} & -\frac{2}{3}\\ -\frac{2}{3} & \frac{2}{3} & \frac{1}{3}\\ \frac{2}{3} & \frac{1}{3} & \frac{2}{3} \end{bmatrix}
Here, the dot product between vector 1 and vector 2 i.e. between row 1 and row 2
Row 1 â‹… Row 2 = (1/3)(-2/3)+(2/3)(2/3)+(-2/3)(1/3) =0
So, Row 1 and Row 2 are Orthogonal.
Also, the Magnitude of Row 1 = ((1/3)2+(2/3)2+(-2/3)2)0.5 = 1
Similarly, we can check for all other rows.
Thus, this matrix A is an example of Orthogonal Matrix.
Example of Orthogonal Matrix
If the transpose of a square matrix with real numbers or values is equal to the inverse matrix of the matrix, the matrix is said to be orthogonal.
Example of 2×2 Orthogonal Matrix
Let's consider the an 2×2 i.e., \bold{A = \begin{bmatrix} \cos x & \sin x\\ -\sin x & \cos x \end{bmatrix}}  Â
.
Let's check this using the product of the matrix and its transpose.
\bold{A^T = \begin{bmatrix} \cos x & -\sin x\\ \sin x & \cos x \end{bmatrix}}
Thus, A\cdot A^T = \begin{bmatrix} \cos x & \sin x\\ -\sin x & \cos x \end{bmatrix}\cdot \begin{bmatrix} \cos x & -\sin x\\ \sin x & \cos x \end{bmatrix}
⇒ A\cdot A^T = \begin{bmatrix} \cos^2 x + \sin^2 x& \sin x \cos x - \sin x \cos x\\ \cos x \sin x - \cos x \sin x & \sin^2 x +\cos^2 x \end{bmatrix}
⇒ A\cdot A^T = \begin{bmatrix} 1 & 0 \\ 0 & 1 \end{bmatrix}
Which is an Identity Matrix.
Thus, A is an example of an Orthogonal Matrix of order 2×2.
Example of 3×3 Orthogonal Matrix
Let us consider 3D Rotation Matrix i.e., 3×3 matrix such that A = \begin{bmatrix} 1 & 0 & 0 \\ 0 & \cos(\theta) & -\sin(\theta) \\ 0 & \sin(\theta) & \cos(\theta) \end{bmatrix}   Â
,
To check this matrix is an orthogonal matrix, we need to
Verify that each column and row matrix is a unit vector, i.e., vector with unit magnitude, and
Verify that the columns and rows are pairwise orthogonal i.e., the dot product between any two rows and columns is 0.
Let's check.
From the matrix, we get
Column 1: [1, 0, 0]
Column 2: [0, cos(θ), sin(θ)]
Column 3: [0, -sin(θ), cos(θ)]
Now,
|[1, 0, 0]| = √(12 + 02 + 02) = 1,
|[0, cos(θ), sin(θ)]| = √(02 + cos2(θ) + sin2(θ)) = 1, and
|[0, -sin(θ), cos(θ)]| = √(02 +  sin2(θ) + cos2(θ)) = 1
Thus, each column is a unit vector.
Column 1 ⋅ Column 2 = [1, 0, 0] ⋅ [0, cos(θ), sin(θ)] = 1*0 + 0*cos(θ) + 0*sin(θ) = 0
Similarly, we can check, other columns as well.
Thus, this satisfies all the conditions for a matrix to be orthogonal.
It follows that the provided matrix is an orthogonal matrix given the characteristics of orthogonal matrices.
Determinant of Orthogonal Matrix
Determinant of any Orthogonal Matrix is either +1 or -1. Here, let's demonstrate the same. Imagine a matrix A that is orthogonal.
For any orthogonal matrix A, we know
A · AT = I
Taking determinants on both sides,
det(A · AT) = det(I)
⇒ det(A) · det(AT) = 1
As, determinant of identity matrix is 1 and det(A) = det(AT)
Thus, det(A) · det(A) = 1
⇒ [det(A)]2 = 1
⇒ det(A) = ±1
Inverse of Orthogonal Matrix
The inverse of the orthogonal matrix is also orthogonal as inverse is same transpose for orthogonal matrix. As for any matrix to be an orthogonal, inverse of the matrix is equal to its transpose.
For an Orthogonal matrix, we know that
A-1 = AT
Also A · AT = AT · A = I . . . (i)
Let two matrix A and B and if they are inverse of each other then,
A · B = B · A = I . . . (ii)
From (i) and (ii),
B = AT which is same as A = AT
So, we conclude that the transpose of an orthogonal matrix is its inverse only.
Learn more
Properties of an Orthogonal Matrix
Some of the properties of Orthogonal Matrix are:
- Inverse and Transpose are equivalent. i.e., A-1 = AT.
- An identity matrix is the outcome of A and its transpose. That is, AAT = ATA = I.
- In light of the fact that its determinant is never 0, an orthogonal matrix is always non-singular.
- An orthogonal diagonal matrix is one whose members are either 1 or -1.
- AT is orthogonal as well. A-1 is also orthogonal because A-1 = AT.
- The eigenvalues of A are ±1 and the eigenvectors are orthogonal.
- As I × I = I × I = I, and IT = I. Thus, I an identity matrix (I) is orthogonal.
How to Identify Orthogonal Matrices?
If the transpose of a square matrix with real numbers or elements equals the inverse matrix, the matrix is said to be orthogonal. Or, we may argue that a square matrix is an orthogonal matrix if the product of the square matrix and its transpose results in an identity matrix.
Suppose A is a square matrix with real elements and of n x n order and AT is the transpose of A. Then according to the definition, if, AT = A-1 is satisfied, then,
A â‹… AT = I
Eigen Value of Orthogonal Matrix
The eigenvalues of an orthogonal matrix are always complex numbers with a magnitude of 1. In other words, if A is an orthogonal matrix, then its eigenvalues λ satisfy the equation |λ| = 1. Let's prove the same as follows:
Let A be an orthogonal matrix, and let λ be an eigenvalue of A. Also, let v be the corresponding eigenvector.
By the definition of eigenvalues and eigenvectors, we have:
Av = λv
Now, take the dot product of both sides of this equation with itself:
(Av) ⋅ (Av) = (λv) ⋅ (λv)
Since A is orthogonal, its columns are orthonormal, which means that AT (the transpose of A) is also its inverse:
AT â‹… A = I
Where I is the identity matrix.
Thus, (vTAT) ⋅ Av = (λv)T ⋅ (λv)
⇒ vT (AT A) v = (λv)T (λv)
⇒ vT I v = (λv)T (λv)
⇒ vT v = (λv)T (λv)
⇒ |v|2 = |λ|2 |v|2
Now, divide both sides of the equation by |v|2:
1 = |λ|2
⇒ |λ| = 1
Read more
Multiplicative Inverse of Orthogonal Matrices
The orthogonal matrix's inverse is also orthogonal. It is the result of the intersection of two orthogonal matrices. An orthogonal matrix is one in which the inverse of the matrix equals the transpose of the matrix.
Orthogonal Matrix Applications
Some of the most common applications of Orthogonal Matrix are:
- Used in multivariate time series analysis.
- Used in multi-channel signal processing.
- Used in QR decomposition.
Read More
Solved Examples on Orthogonal Matrix
Example 1: Is every orthogonal matrix symmetric?
Solution:
Every time, the orthogonal matrix is symmetric. Thus, the orthogonal matrix is a property of all identity matrices. An orthogonal matrices will also result from the product of two orthogonal matrices. The orthogonal matrix will likewise have a transpose that is orthogonal.
Example 2: Check whether the matrix X is an orthogonal matrix or not?
\bold{\begin{bmatrix} \cos x & \sin x\\ -\sin x & \cos x \end{bmatrix}}
Solution:
We know that the orthogonal matrix's determinant is always ±1.
The determinant of X = cos x · cos x - sin x · (-sin x)
⇒ |X| = cos2x + sin2x = 1
⇒ |X| = 1
Hence, X is an Orthogonal Matrix.
Example 3: Prove orthogonal property that multiplies the matrix by transposing results into an identity matrix if A is the given matrix.
Solution:
A = \begin{bmatrix} -1 & 0\\ 0 & 1 \end{bmatrix}
Thus, A^{T} = \begin{bmatrix} 1 & 0\\ 0 & -1 \end{bmatrix}
⇒ A \cdot A^{T} = \begin{bmatrix} 1 & 0\\ 0 & 1 \end{bmatrix}
Which is an identity matrix.
Thus, A is an Orthogonal Matrix.
Practice Problems on Orthogonal Matrix
Q1: Let A be a square matrix:
A = \begin{bmatrix} 0.6 & -0.8 \\ 0.8 & 0.6 \end{bmatrix}
Determine whether matrix A is orthogonal.
Q2: Given the matrix A:
A = \begin{bmatrix} 1 & 0 \\ 0 & -1 \end{bmatrix}
Is matrix A orthogonal?
Q3: Let Q be an orthogonal matrix. Prove that the transpose of Q is also its inverse, i.e., QT = Q-1
Q4: Consider the matrix C:
C = \begin{bmatrix} \frac{1}{\sqrt{3}} & -\frac{1}{\sqrt{3}} & \frac{1}{\sqrt{3}} \\ \frac{1}{\sqrt{2}} & 0 & -\frac{1}{\sqrt{2}} \\ -\frac{1}{\sqrt{6}} & \frac{2}{\sqrt{6}} & \frac{1}{\sqrt{6}} \end{bmatrix}
Is matrix C orthogonal?
Similar Reads
Non-linear Components In electrical circuits, Non-linear Components are electronic devices that need an external power source to operate actively. Non-Linear Components are those that are changed with respect to the voltage and current. Elements that do not follow ohm's law are called Non-linear Components. Non-linear Co
11 min read
Spring Boot Tutorial Spring Boot is a Java framework that makes it easier to create and run Java applications. It simplifies the configuration and setup process, allowing developers to focus more on writing code for their applications. This Spring Boot Tutorial is a comprehensive guide that covers both basic and advance
10 min read
Class Diagram | Unified Modeling Language (UML) A UML class diagram is a visual tool that represents the structure of a system by showing its classes, attributes, methods, and the relationships between them. It helps everyone involved in a projectâlike developers and designersâunderstand how the system is organized and how its components interact
12 min read
3-Phase Inverter An inverter is a fundamental electrical device designed primarily for the conversion of direct current into alternating current . This versatile device , also known as a variable frequency drive , plays a vital role in a wide range of applications , including variable frequency drives and high power
13 min read
Backpropagation in Neural Network Back Propagation is also known as "Backward Propagation of Errors" is a method used to train neural network . Its goal is to reduce the difference between the modelâs predicted output and the actual output by adjusting the weights and biases in the network.It works iteratively to adjust weights and
9 min read
What is Vacuum Circuit Breaker? A vacuum circuit breaker is a type of breaker that utilizes a vacuum as the medium to extinguish electrical arcs. Within this circuit breaker, there is a vacuum interrupter that houses the stationary and mobile contacts in a permanently sealed enclosure. When the contacts are separated in a high vac
13 min read
Polymorphism in Java Polymorphism in Java is one of the core concepts in object-oriented programming (OOP) that allows objects to behave differently based on their specific class type. The word polymorphism means having many forms, and it comes from the Greek words poly (many) and morph (forms), this means one entity ca
7 min read
CTE in SQL In SQL, a Common Table Expression (CTE) is an essential tool for simplifying complex queries and making them more readable. By defining temporary result sets that can be referenced multiple times, a CTE in SQL allows developers to break down complicated logic into manageable parts. CTEs help with hi
6 min read
Spring Boot Interview Questions and Answers Spring Boot is a Java-based framework used to develop stand-alone, production-ready applications with minimal configuration. Introduced by Pivotal in 2014, it simplifies the development of Spring applications by offering embedded servers, auto-configuration, and fast startup. Many top companies, inc
15+ min read
Python Variables In Python, variables are used to store data that can be referenced and manipulated during program execution. A variable is essentially a name that is assigned to a value. Unlike many other programming languages, Python variables do not require explicit declaration of type. The type of the variable i
6 min read