Continue **reading** Section 3.3.
**But we aren't covering elementary matrices.**
Work through recommended homework questions.

**Quiz 5** will focus on 3.1, 3.2 and the first half of 3.3 (up to
and including Example 3.26).

**Office hour:** today, 12:30-1:30 and Wednesday, 10:30-11:15, MC103B.

**Help Centers:** Monday-Friday 2:30-6:30 in MC 106.

For example, if $B$ is partitioned into columns as $B = [ \, \vb_1 \mid \vb_2 \mid \cdots \mid \vb_r ]$, then we have: $$ AB = [\, A\vb_1 \mid A\vb_2 \mid \cdots \mid A\vb_r ] . $$

Also, remember that if $A$ is partitioned into columns as $A = [ \, \va_1 \mid \va_2 \mid \cdots \mid \va_n ]$, then $$ A \vx = x_1 \va_1 + \cdots + x_n \va_n , $$ a linear combination of the columns of $A$.

After adding, subtracting and multiplying, what is missing?

We could do the same thing for a matrix equation $A \vx = \vb$ if we
could find a matrix $A'$ such that $A' A = I$. Then:
$$
\kern-4ex
A \vx = \vb
\qimplies
A' A \vx = A' \vb
\qimplies
\vx = A' \vb .
$$
So, *if* $A \vx = \vb$ has a solution, then it must be $A' \vb$.
On the other hand, let's check whether $A' \vb$ *is* a solution:
$$
\kern-4ex
A ( A' \vb ) = A A' \vb = \query{I \vb} = \vb ,
$$
where the last step only works if we know that $A A' = I$ as well.

So we require both conditions:

**Definition:** An **inverse** of an $n \times n$ matrix $A$ is an $n \times n$
matrix $A'$ such that
$$
A A' = I \qtext{and} A' A = I .
$$
If such an $A'$ exists, we say that $A$ is **invertible**.

(We'll talk about what happens when $A$ is not square tomorrow.)

**Example:** If $A = \bmat{rr} 1 & 2 \\ 3 & 7 \emat$,
then $A' = \bmat{rr} 7 & -2 \\ -3 & 1 \emat$ is an inverse of $A$.
(On board.)

**Example:**
Does $O = \bmat{rr} 0 & 0 \\ 0 & 0 \emat$ have an inverse?

No, since for any matrix $C$, we always have $CO$ equal to a zero matrix,
so it can't be equal to the identity matrix.

**Example:**
Does $B = \bmat{rr} -1 & 3 \\ 2 & -6 \emat$ have an inverse?

No. Suppose that $B'$ was an inverse to $B$. Then $B B' = I$.

In particular, if $\vb$ is the first column of $B'$, then $B \vb = \ve_1$.

But this means that $\ve_1$ is a linear combination of the columns of $B$,

which is not possible since the columns are parallel and point in a different direction.

(The book gives a different argument.)

In particular, if $\vb$ is the first column of $B'$, then $B \vb = \ve_1$.

But this means that $\ve_1$ is a linear combination of the columns of $B$,

which is not possible since the columns are parallel and point in a different direction.

(The book gives a different argument.)

We'll learn next class how to determine whether a matrix has an inverse, and how to find it when it does. Today we'll discuss some general properties, and also $2 \times 2$ matrices.

**Theorem 3.6:** If $A$ is an invertible matrix, then its inverse is unique.

**Proof:**
Suppose that $A'$ and $A''$ are **both** inverses of $A$. We'll show they must be equal:
$$
\kern-6ex
A' = A' I = A' (A A'') = (A' A) A'' = I A'' = A'' . \qquad\Box
$$
Because of this, we write $A^{-1}$ for **the** inverse of $A$,
when $A$ is invertible. We do *not* write $\frac{1}{A}$.

**Theorem 3.7:** If $A$ is an invertible matrix $n \times n$ matrix,
then the system $A \vx = \vb$ has the unique solution $\vx = A^{-1} \vb$
for any $\vb$ in $\R^n$.

This follows from the argument we gave earlier.

**Example on board:** Solve the systems
$$
\kern-4ex
\begin{aligned}
\ph x + 2 y &= 3 \\
3 x + 7 y &= 4
\end{aligned}
\qqtext{and}
\begin{aligned}
\ph x + 2 y &= \ph \, 2 \\
3 x + 7 y &= -1
\end{aligned}
$$

**Remark:** This is **not** in general an efficient way to solve
a system. Using row reduction is usually faster. And row reduction
works when the coefficient matrix is not square or not invertible.
The above method can be useful if you need to solve a lot of systems
with the same $A$ but varying $\vb$.

**Theorem 3.8:** The matrix $A = \bmat{cc} a & b \\ c & d \emat$ is
invertible if and only if $ad - bc \neq 0$. When this is the case,
$$
A^{-1} = \frac{1}{ad-bc} \, \bmat{rr} \red{d} & \red{-}b \\ \red{-}c & \red{a} \emat .
$$

We call $ad-bc$ the **determinant** of $A$, and write it $\det A$.

It *determines* whether or not $A$ is invertible, and also shows
up in the formula for $A^{-1}$.

**Example:** The determinant of $A = \bmat{rr} 1 & 2 \\ 3 & 7 \emat$
is $\det A = 1 \cdot 7 - 2 \cdot 3 = 1$, so
$$A^{-1} = \frac{1}{1} \bmat{rr} 7 & -2 \\ -3 & 1 \emat ,$$
as we saw before.

**Example:** The determinant of $B = \bmat{rr} -1 & 3 \\ 2 & -6 \emat$
is $\det B = (-1)(-6) - 3 \cdot 2 = 0$, so $B$ is not invertible (as we saw).

**Why the formula works:** Show on board that
$$
\bmat{cc} a & b \\ c & d \emat \bmat{rr} d & -b \\ -c & a \emat
= \det A \bmat{cc} 1 & 0 \\ 0 & 1 \emat .
$$
Therefore, if $\det A$ is nonzero,
$$
\kern-8ex
\bmat{cc} a & b \\ c & d \emat \left( \frac{1}{\det A} \bmat{rr} d & -b \\ -c & a \emat \right)
= \frac{\det A}{\det A} \bmat{cc} 1 & 0 \\ 0 & 1 \emat
= \bmat{cc} 1 & 0 \\ 0 & 1 \emat .
$$
A similar argument works for the other order of multiplication.

**Why $A$ is not invertible when $\det A = 0$:**
If $\det A = 0$, then
$$
A B = (\det A) I = O ,
$$
where we write $B$ for $\bmat{rr} d & -b \\ -c & a \emat$.
So if $A'$ is an inverse of $A$, then
$$
B = A' A B = A ' O = O
$$
But if the entries of $B$ are zero, then so are the entries of $A$,
and it's impossible to have $A' A = I$.

- $A^{-1}$ is invertible and $(A^{-1})^{-1} = \query{A}$
- If $c$ is a non-zero scalar, then $cA$ is invertible and $(cA)^{-1} = \query{\frac{1}{c} A^{-1}}$
- $AB$ is invertible and $(AB)^{-1} = \query{B^{-1} A^{-1}}$ (socks and shoes rule)
- $A^T$ is invertible and $(A^T)^{-1} = \query{(A^{-1})^T}$
- $A^n$ is invertible for all nonnegative integers $n$ and $(A^n)^{-1} = \query{(A^{-1})^n}$

To verify these, in every case you just check that the matrix shown is an inverse. All 5 done on the board.

**Remark:** Property (c) is the most important, and generalizes to
more than two matrices, e.g. $(ABC)^{-1} = C^{-1} B^{-1} A^{-1}$.

**Remark:** For $n$ a positive integer, we define $A^{-n}$ to
be $(A^{-1})^n = (A^n)^{-1}$. Then $A^n A^{-n} = I = A^0$,
and more generally $A^r A^s = A^{r+s}$ for all integers $r$ and $s$.

**Remark:** There is no formula for $(A+B)^{-1}$.
In fact, $A+B$ might not be invertible, even if $A$ and $B$ are.

We can sometimes use these properties to solve a matrix equation for an unknown matrix. Assume that $A$, $B$ and $X$ are invertible matrices of the same size.

**Example:** Solve $AXB^2 = B A B^{-1}$ for $X$.

**Solution:**
$$
\kern-9ex
\begin{aligned}
AXB^2 = B A B^{-1} &\implies A^{-1} (A X B^2) B^{-2} = A^{-1} (B A B^{-1}) B^{-2} \\
&\implies X = A^{-1} B A B^{-3}
\end{aligned}
$$

**Example:** Solve $(AX^TB)^{-1} = BA$ for $X$.

**Solution:**
$$
\kern-9ex
\begin{aligned}
(AX^TB)^{-1} = BA &\implies ((A X^T B)^{-1})^{-1} = (BA)^{-1} \\
&\implies A X^T B = A^{-1} B^{-1} \\
&\implies A^{-1} (A X^T B) B^{-1} = A^{-1} (A^{-1} B^{-1}) B^{-1} \\
&\implies X^T = A^{-2} B^{-2} \\
&\implies (X^T)^T = (A^{-2} B^{-2})^T \\
&\implies X = (A^{-2} B^{-2})^T
\end{aligned}
$$

**Example:** Solve $A X^2 + BX + C = 0$?
There is no easy method in general, using linear algebra.

**True/false:**
If $A$ is symmetric, then $A$ is invertible.

**False.** For example, the zero matrix is symmetric, but not invertible.

**True/false:**
If $A$ is symmetric and invertible, then $A^{-1}$ is symmetric.

**True,**
because $(A^{-1})^T = (A^T)^{-1} = A^{-1}$, using Theorem 3.9 (d).

**True/false:**
If $AB = AC$, then $B = C$.

**False.**
For example, if $A$ is the zero matrix, then $AB = AC$ for
any matrices $B$ and $C$.

**True/false:**
If $AB = AC$ and $A$ is invertible, then $B = C$.

**True.**
If $AB = AC$, then $A^{-1}AB = A^{-1}AC$ and so $B = C$.

**Challenge problem:**
Can you find a $2 \times 3$ matrix $A$ and a $3 \times 2$ matrix $A'$ such that
$A A' = I_2$ and $A' A = I_3$?

We'll discuss this one next class.