(a) Computing the transpose of a matrix
Let \(A\) be an \(m\times n\) matrix with entry \(a_{ij}\) in the \(i\)-th row and the \(j\)-th column. The transpose of \(A\), denoted \(A^\top\), is the \(n\times m\) matrix with entry \(a_{ji}\) in the \(i\)-th row and the \(j\)-th column. In other words, the rows of \(A\) become the columns of \(A^\top\) (equivalently, the columns of \(A\) become the rows of \(A^\top\)).
Compute the transpose of the following matrices:
(b) Computing matrix products with transposes
Suppose that
\[ A = \begin{bmatrix} -1 & 3 \\ 3 & 4 \end{bmatrix}, \quad \mathbf{u} = \begin{bmatrix} 1 \\ 2 \end{bmatrix}, \quad \mathbf{v} = \begin{bmatrix} 0 \\ -2 \end{bmatrix}. \]
Definition
Given a symmetric \(n\times n\) matrix \(A\), the function \[ q: \mathbb{R}^n \to \mathbb{R}, \quad q(\mathbf{x}) = \mathbf{x}^\top A \mathbf{x} \]
is called a quadratic form. The matrix \(A\) is called the matrix of the quadratic form.
If you expand the product \(\mathbf{x}^\top A \mathbf{x}\), you will see that \(q(\mathbf{x})\) is a polynomial of degree 2 in the entries of \(\mathbf{x}\): \[ q(\mathbf{x}) = \sum_{i=1}^n\sum_{j=1}^n a_{ij} x_i x_j. \]
You can use this formula to reverse engineer a quadratic form to find is matrix.
Find the matrices of the following quadratic forms.
Let’s suppose for simplicity that \(f: \mathbb{R}^2 \to \mathbb{R}\), so that we may visualize the graph of \(f\) embedded in \(\mathbb{R}^3\).
Given a position vector \(\mathbf{v} = \begin{bmatrix} x \\ y \end{bmatrix}\) and a step vector \(\mathbf{h}\), the definition of the derivative \(f'(\mathbf{v})\) is \[ f'(\mathbf{v}) \mathbf{h} = \lim_{\lambda \to 0} \frac{f(\mathbf{v} + \lambda \mathbf{h}) - f(\mathbf{v})}{\lambda}. \]
On the other hand, we may consider the function \[ g: \mathbb{R} \to \mathbb{R}, \quad g(t) = f(\mathbf{v} + t \mathbf{h}) \]
The graph of \(g\) is a cross section obtained by slicing the graph of \(f\) with a vertical plane containing the step vector \(\mathbf{h}\).
Note that \[ g'(0) = \lim_{t \to 0} \frac{g(t) - g(0)}{t} = \lim_{t \to 0} \frac{f(\mathbf{v} + t \mathbf{h}) - f(\mathbf{v})}{t}. \]
But if you replace \(t\) with \(\lambda\) in the previous limit, you see that \(g'(0) = f'(\mathbf{v}) \mathbf{h}\).
Theorem
Suppose \(f: \mathbb{R}^n \to \mathbb{R}\) is differentiable at \(\mathbf{v} \in \mathbb{R}^n\). Then for any step vector \(\mathbf{h} \in \mathbb{R}^n\), we have
\[ f'(\mathbf{v}) \mathbf{h} = \frac{d}{d\lambda} f(\mathbf{v} + \lambda \mathbf{h}) \Big|_{\lambda = 0}. \]
Compute the derivatives of the following functions using the theorem on the previous slide. Check your answer by computing the derivatives using partial derivatives.
Definition
Let \(f: \mathbb{R}^n \to \mathbb{R}\) be a function. The second derivative of \(f\) at a point \(\mathbf{v} \in \mathbb{R}^n\) with step vector \(\mathbf{h}\) is the \(n\times n\) matrix \(f''(\mathbf{v})\) such that
\[ \mathbf{h}^\top f''(\mathbf{v}) \mathbf{h} = \frac{d^2}{d\lambda^2} f(\mathbf{v} + \lambda \mathbf{h}) \Big|_{\lambda = 0}, \]
provided that the derivative on the right hand side exists.
Theorem/Definition
Let \(f: \mathbb{R}^n \to \mathbb{R}\) be a function, and let \(\mathbf{v} \in \mathbb{R}^n\) be a point.
(a) Computing second derivatives
Compute the second derivatives of the following functions using the definition of the second derivative.
(b) Computing curvatures
Now compute the curvatures of the functions in part (a) at the following points and in the following directions.