Chapter 2

Section 2.1: Matrix Operations

Definition of multiplication:
If $A$ is an $mxn$ matrix and $B$ is an $nxp$ matrix where $B=[\vec{b_1},\vec{b_2},\vec{b_3}...\vec{b_p}$ then $AB$ is the matrix

$AB=A[\vec{b_1},\vec{b_2},\vec{b_3}...\vec{b_p}]=[A\vec{b_1},A\vec{b_2},A\vec{b_3}...A\vec{b_p}]$

Example 1:

Let $A=\begin{bmatrix}1&2\\0&-1\end{bmatrix}$ and $B=\begin{bmatrix}2&0&1\\1&-1&2\end{bmatrix}$

$AB=\begin{bmatrix}1&2\\0&-1\end{bmatrix}\begin{bmatrix}2&0&1\\1&-1&2\end{bmatrix}=\begin{bmatrix}A\vec{b_1}&A\vec{b_2}&A\vec{b_3}\end{bmatrix}$

$A\vec{b_1}=\begin{bmatrix}1&2\\0&-1\end{bmatrix}\begin{bmatrix}2\\1\end{bmatrix}=\begin{bmatrix}4\\-1\end{bmatrix}$

$A\vec{b_2}=\begin{bmatrix}1&2\\0&-1\end{bmatrix}\begin{bmatrix}0\\-1\end{bmatrix}=\begin{bmatrix}-2\\1\end{bmatrix}$

$A\vec{b_3}=\begin{bmatrix}1&2\\0&-1\end{bmatrix}\begin{bmatrix}1\\2\end{bmatrix}=\begin{bmatrix}-2\\5\end{bmatrix}$

$AB=\begin{bmatrix}4&-2&5\\-1&1&-1\end{bmatrix}$

Theorem 1
Let $A$, $B$, and $C$ be matrices of the same size, and let $r$ and $s$ be scalars.
$A + B = B + A$ $r(A + B) = rA + rB$
$(A + B) + C = A + (B + C)$ $(r + s)A = rA + sA$
$A + 0 = A$ $r(sA) = (rs)A$
Theorem 2

Let $A$ be an $m \times n$ matrix, and let $B$ and $C$ have sizes for which the indicated sums and products are defined.

$A(BC) = (AB)C$ Associative Law
$A(B + C) = AB + AC$ Left Distributive Law
$(B + C)A = BA + CA$ Right Distributive Law
$r(AB) = (rA)B = A(rB)$ for any scalar $r$
$I_mA = A = AI_n$ Identity Matrix
Theorem 3

Let A and B denote matrices whose sizes are appropriate for the following sums and products.

a. $(A^T)^T = A$
b. $(A+B)^T = A^T+B^T$
c. For any scalar r, $(rA)^T = rA^T$
d. $(AB)^T = B^TA^T$

Section 2.2: The Inverse of a Matrix

Definitions:

Invertible (also known as nonsingular) = Matrices such that there exists a multiplicative inverse.
Not invertible (also known as singular) = Matrices such that there does not exist a multiplicative inverse.
*When differentiating between the two, remember Dr.V's reference to singular and nonsingular aka nonsingular matrices have a buddy and singular matrices do not.

Theorem 4

Let $A=\begin{bmatrix}a&b\\c&d\end{bmatrix}$. If $ad-bc\not=0$, then $A$ is invertible and
$A^{-1} =\frac{1}{ad-bc} \begin{bmatrix}d&-b\\-c&a\end{bmatrix}$
If $ad-bc = 0$, then $A$ is not invertible.

Theorem 5

Let $A$ be an $n \times n$ matrix, then for each $\overrightarrow{b}$ in $\mathbb{R}^{n}$ , the equation $A\overrightarrow{x} = \overrightarrow{b}$ has the unique solution $\overrightarrow{x} = A^{-1}\overrightarrow{b}$.

Theorem 6
  • If $A$ is an invertible matrix, then $A^{-1}$ is invertible and

$(A^{-1})^{-1} = A$

  • If $A$and $B$ are $n \times n$ invertible matrices, then so is $AB$, and the inverse of $AB$ is the product of the inverses of $A$ and $B$ in reverse order.

That is, $(AB)^{-1} = B^{-1}A^{-1}$

  • If $A$ is an invertible matrix, then so is $A^{T}$, and the inverse of $A^{T}$ is the transpose of $A^{-1}$. That is,

$(A^{T})^{-1} = (A^{-1})^{T}$

Theorem 7

An $n \times n$ matrix $A$ is invertible if and only if $A$ is row equivalent to $I_n$, and in this case, any sequence of elementary row operations that
reduces $A$ to $I_n$ also transforms $I_n$ into $A^{-1}$.

Theorem 8

Let $A$ be a square $n \times n$ matrix. Then the following statements are equivalent. That is, for a given $A$, the statements are either all true or all false.

a. $A$ is an invertible matrix
b. $A$ is row equivalent to the $n \times n$ identity matrix.
c. $A$ has n pivot position
d. The equation $Ax = 0$ has only the trivial solution
e. The columns of A form a linearly independent set.
f. The linear transformation $x$ to $Ax$ is one-to one.
g. The equation $Ax = b$ has at least one solution for each $b$ in $R^n$
h. The columns of $A$ span $R^n$
i. The linear transformation $x$ to $Ax$ maps $R^n$ onto $R^n$
j. There is an $n \times n$ matrix $C$ such that $CA = I$
k. There is an $n \times n$ matrix $D$ such that $AD = I$
l. $A^T$ is an invertible matrix

Theorem 8: Explained
Equivalencies Theorem, Definition, etc.
a $\Leftrightarrow$ l Section 2, Theorem 6.c, (p.105)
d $\Leftrightarrow$ e Fact of Linear Independence, (p. 57)
g $\Leftrightarrow$ h Section 1, Theorem 4, (p. 37)
cell-content cell-content
cell-content cell-content
cell-content cell-content
cell-content cell-content
cell-content cell-content
cell-content cell-content
cell-content cell-content
cell-content cell-content
Theorem 9

Let $T : R^n —> R^n$ be a linear transformation and let A be the standard matrix for T. Then T is invertible if and only if A is an invertible matrix.
In that case, the linear transformation S given by $S(x) = A^{-1}x$ is the unique function satisfying equations (1) and (2).

Homework Problems

Section 2.1
23) If $CA=I_n$, show that $A\vec{x} = 0$ has only the trivial solution. Why must $A$ not have more columns then rows?
Notice that
$\vec{x} = (I_n)\vec{x} = CA(\vec{x}) = C(A\vec{x})=C(\vec{0})=\vec{0}$.
This implies only trivial solution exists.
If $A$ had more columns then rows, the system would be linearly dependent, and non-trivial solutions would exist. This would contradict our previous logic.

Section 2.2
17) If $A$, $B$, and $C$ are invertible, show that $ABC$ is invertible.
Notice if $ABC$ was invertible, then there would exist a matrix $D$ such that $(ABC)D = D(ABC) = I$, where $I$ is the identity matrix.
Consider $D=C^{-1}B^{-1}A^{-1}$. Notice $(ABC)(C^{-1}B^{-1}A^{-1}) = I$ and $(C^{-1}B^{-1}A^{-1})(ABC) = I$. Additionally, notice by theorem 6, $C^{-1}B^{-1}A^{-1}=(ABC)^{-1}$.
Thus, $ABC$ truly is invertible.

Unless otherwise stated, the content of this page is licensed under Creative Commons Attribution-ShareAlike 3.0 License