Multivariable functions and how to optimize them

- Multivariate functions
- Partial Derivatives
- The Gradient
- The Jacobian
- Second-order and cross-partial derivatives
- The Hessian
- Unconstrainted Optimization
- Using
`sympy`

to walk through the multivariate optimization steps

In [1]:

```
import numpy as np
import numpy.linalg as la
import requests
from sympy import *
init_printing(use_unicode=True)
# Define sympy symbols
x,y = symbols('x y')
```

In [2]:

```
# Read in Visualization code from Github (requires plotly module)
exec(requests.get('https://raw.githubusercontent.com/edunford/ppol564/master/lectures/visualization_library/plotly_graph_functions.py').content)
mv = multivar()
```

$$f(x): A \to B$$

Where values from set $A$, where $x \in A$, are mapped to values in set $B$

However, as we saw in linear algebra, sets need not be confined to one dimension. That is, set $A$ need no be limited to $\mathbb{R}$ but rather can encompass $\mathbb{R}^n$.

$$\textbf{f(x)} : \mathbb{R}^n \to \mathbb{R}^m$$

**scalar multi-variable value** functions take the following forms:

$$f(x,y,z) = 3xy - y^2z + 2$$

$$f(\textbf{x}) = f(x_1,x_2,x_3) =\frac{x_1 x_2}{x_3}$$

where $\textbf{x}$ is a vector of values.

**Vector-value multi-variable functions** that take in some $n$ number of variables and spit out $m$ number of mapped locations: $\mathbb{R}^n \to \mathbb{R}^m$. These vector functions can be thought of as being composed of multiple components $f_i(\cdot)$.

$$ \textbf{f}(x,y,z) = \begin{pmatrix} f_1(x,y,z) \\ f_2(x,y,z) \\ f_3(x,y,z) \\ f_4(x,y,z) \end{pmatrix} = \begin{pmatrix} xyz \\ 2x+\frac{y}{z} \\ z-y^x \\ y^2 + x^2 + z^2 \end{pmatrix}$$

As we've seen if $\textbf{f(x)}$ is linear (review "linear combination"), then we can rewrite it as $\textbf{Ax}$ for some transformation matrix $\textbf{A}$.

However, we are often dealing with non-linear functions. As such, we need to employ either quadratic approximations or think about concepts like "local-linearity".

In short, like vectors and matrices, beyond a point, it's hard and or impossible to visualize higher dimensional space. That said, I'll introduce two types of plots in the bivariate setting that might help us think through some concepts: 3D surface plots and contour plots (with heat map coloring elements).

In [3]:

```
# For this example, consider the following function
f = x**2 + y**2
f
```

Out[3]:

In [4]:

```
mv.plot_3d(f)
```

In [5]:

```
mv.plot_contour(f)
```

What is the change in the function in one of the input variables, *holding all else constant*?

Consider the following function:

$$ f(x,y) = 2x+y^2 $$

How should we think about the output of $f(x,y)$ when we nudge $x$ a little bit? We'll nudge it and just leave the other variable $y$ be.

$$ \frac{\partial f(x,y) }{\partial x} = \lim_{h\to0} \frac{f(x+h,y) - f(x,y)}{h}$$

$$ \lim_{h\to0} \frac{ (2x+h)+y^2 - (2x+y^2)}{h} $$

$$ \lim_{h\to0} \frac{ (2x+h) - 2x + y^2 - y^2}{h} $$

$$ \lim_{h\to0} \frac{ 2(x+h - x) }{h} $$

$$ \lim_{h\to0} \frac{ 2h }{h} $$

$$ \frac{\partial f(x,y) }{\partial x} = 2 $$

Note that: Everything that is ** not** $x$ effectively gets

**Notation**

$$ \frac{\partial f }{\partial x} = \frac{\partial }{\partial x} f = \partial_x f = f_x $$

The first order derivative of a function is now a component of many things. Sure, it's $\frac{\partial }{\partial x} f$ but it's also $\frac{\partial }{\partial y} f$ and so on. The function is *multivariate* and each of those variables corresponds with a dimension. Thus, we want information about how the function changes given a nudge along all those dimensions.

Luckily, there is a way of capturing and storing this information as a vector. We call this the **gradient vector** and it is denoted with the symbol $\nabla$.

$$ \nabla = \begin{pmatrix} \partial_{x_1} \\ \partial_{x_2} \\ \vdots \\ \partial_{x_m} \end{pmatrix} $$

The gradient vector is an operator where each element of the vector is the partial derivative with respect to one of the available variables. We write $\nabla f$ for the gradient of the scalar function $f$

Again, let's use the scalar value function from above:

$$f(x,y) = 2x+y^2$$

We know that

$$f_x = 2$$

And we can use the same method to calculate $f_y$

$$f_y = 2y$$

Now we stack the results

$$\nabla f = \begin{pmatrix} 2 \\ 2y \\ \end{pmatrix}$$

Note that the gradient is outputs a vector, which we can visualize.

In [6]:

```
mv.plot_gradient(f)
```

$\nabla f $ points in the direction of steepest ascent: the direction in which the function of $f$ **increases most rapidly**, and its **magnitude is the rate of this increase**. It tells us how to climb the hill.

**Another example**

Say we have function

$$g(x,y) = x + .0001y $$

The gradient would be

$$ \nabla g = \begin{pmatrix} 1 \\ .0001 \\ \end{pmatrix} $$

the gradient of $g$ points most prominently in the direction of the x-axis. This means that it is moving most rapidly in that direction.

Say we want to know how fast the function increases in some other direction $\vec{v}$. To figure this out, we can just take the dot product of the gradient of the function with a vector in the direction of interest.

$$ \nabla f \cdot \vec{v} = \begin{pmatrix} 2 \\ 2y \\ \end{pmatrix} \cdot \vec{v}$$

This tells us the rate at which $f$ changes as one moves in the direction of $\vec{v}$. Given how we can conceptualize movement in a multivariate setting is more nuanced, the directional derivative offers a way for us to make particular movements in particular directions.

For example, say

$$ \textbf{x} = \begin{pmatrix} 1 \\ 1 \end{pmatrix} $$

$$ \textbf{v} = \begin{pmatrix} 3 \\ 2 \end{pmatrix} $$

$$\nabla f (x) \cdot \textbf{v} = \begin{pmatrix} 2 \\ 2(1) \end{pmatrix} \cdot \begin{pmatrix} 3 \\ 2 \end{pmatrix} = 2(3) + 2(2) = 10$$

The rate of change starting at $\textbf{x}$ and changing to $\textbf{v}$ is 10.

For vector-valued functions, isolating the first-order partial derivatives is more involved but nothing new from what we've seen before.

Consider the following vector-valued function from before:

$$ \textbf{f}(x,y) = \begin{pmatrix} xy \\ x^2 + y^2 \\ y-2x \end{pmatrix}$$

We capture information about the first order derivatives of $\textbf{f}$ using a special matrix called a **Jacobian Matrix**. Note that $\textbf{f}$ takes in two variable inputs and outputs a 3-dimensional vector. The dimensions of this matrix will correspond with this 2 in, 3 out arrangement.

$$ \textbf{f}(x,y) = \begin{pmatrix} f^1(x,y) \\ f^2(x,y) \\ f^3(x,y) \end{pmatrix}$$

$$ \textbf{J} = \begin{pmatrix} f_x^1(x,y) & f_y^1(x,y)\\ f_x^2(x,y) & f_y^2(x,y) \\ f_x^3(x,y) & f_y^3(x,y) \end{pmatrix} = \begin{pmatrix} y & x\\ 2x & 2y \\ -2 & 1 \end{pmatrix} $$

All in all, the Jacobian is just a generalization of the gradient for a vector-valued function. You'll see the Jacobian used for linear approximations of functions (see reading).

As we saw last time, taking the higher order derivatives of a function can tell us important information about the shape of a curve.

Taking **second-order partial derivations** follows the same logic we encountered in the univariate context (with slightly different notation).

$$ \frac{\partial^2}{ \partial x^2 } = f_{xx}$$

Consider the function

$$ f(x,y) = 2y^2x^3$$

$$ f_x = 6y^2x^2$$

$$ f_{xx} = 12y^2x$$

Likewise,

$$ f_y = 4x^3y$$

$$ f_{yy} = 4x^3 $$

We can also take the **cross-partial derivative** (or the **mixed-partial derivative**): that is, we can first take the derivative with respect to one variable and then take that partial derivate with respect to another variable.

$$ f_x = 6y^2x^2$$

$$ f_{xy} = \frac{\partial^2 f}{\partial_y \partial_x} = 12yx^2$$

Likewise,

$$ f_y = 4x^3y$$

$$ f_{yx} = \frac{\partial^2 f}{\partial_x \partial_y} = 12x^2y $$

Note something interesting here

$$ f_{yx} = f_{xy}$$

$$ 12yx^2 = 12x^2y $$

The order in which we differentiate doesn't matter!

A multidimensional analog to the second derivative, the Hessian provides us a way to record information about the second-order partial derivatives and the cross-partial derivatives.

Again, consider:

$$ f(x,y) = 2y^2x^3$$

We can concisely store this information as a matrix.

$$ \textbf{H} =\begin{pmatrix} f_{xx} & f_{yx} \\f_{xy} & f_{yy}\end{pmatrix}$$

$$ \textbf{H} =\begin{pmatrix} 12y^2x & 12yx^2 \\12yx^2 & 4x^3\end{pmatrix}$$

where the diagonal of $\textbf{H}$ describes the second-order partials and the off-diagonal describes the cross-partials. Note that the Hessian is both a square and symmetric matrix. As we'll see below, this means we can decompose it. Moreover, its eigenvalues will be instrumental when evaluating minima/maxima.

<font color = "grey" > Disclaimer: note that we are discussing

unconstrained optimization, so we'll avoid thinking about the boundaries of some functional domain for now. Later on, we'll introduce this topic as constraints on our optimization problems. </font>

Consider the function:

$$ f(x,y) = x^2 + y^2 + xy $$

As was the case in the univariate setting, we want to find our extrema candidates by isolating the stationary points in the function. Thus, let's hunt for the locations where the first-order derivatives are 0. We can do this by setting the gradient to 0.

$$\nabla f = \textbf{0} $$

where $\textbf{0}$ equals the zero vector.

$$ \nabla f = \begin{pmatrix} 2x + y \\ 2y + x \end{pmatrix} $$

$$ \begin{pmatrix} 2x + y \\ 2y + x \end{pmatrix} = \textbf{0} $$

$$ 2x + y = 0 \\ 2y + x = 0 $$

$$ y = -2x \\ 2(-2x) + x = 0 $$

$$ y = -2x \\ x = 0 $$

$$ y = 0 \\ x = 0 $$

So we only have one candidate entry

$$ \textbf{x*} = \begin{pmatrix} x_0^* \\ y_0^* \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \end{pmatrix}$$

Recall that stationary points are where the "action happens" but we only care about certain kinds of actions. In particular, we want to know where the function is maximized/minimized. Thus, we must classify if a function is a local maxima or minima.

Recall that we leveraged the second-order derivatives to figure out if a function was convex (minimum), concave (maximum), or an inflection (saddle) around a stationary point. We'll we do exactly the same thing here, except we have to use our more nuanced conceptualization of the second derivative: the **Hessian**!

$$ \textbf{H} =\begin{pmatrix} f_{xx} & f_{yx} \\f_{xy} & f_{yy}\end{pmatrix}$$

In fact, the determinant of the **Hessian** tells us what we want to know.

$$ |\textbf{H}| =f_{xx}f_{yy} - f_{xy}^2 $$

If $|\textbf{H}| > 0 $, then we know a maxima/minima exists at the stationary point.

If $|\textbf{H}| < 0 $, then we know the stationary point is a saddle point.

If $|\textbf{H}| = 0 $, then it's undefined (insufficient information)

- $f_{xx} \to $ is the function concave/convex in the $x$ direction?

- $f_{yy} \to $ is the function concave/convex in the $y$ direction?

- $f_{xx}f_{yy} \to $ is there agreement with regard to the concavity/convexity along the two dimensions?
- if they agree, then this product will be positive.
- if they disagree, i.e. one axis says one thing, the other says another thing, then this product will be negative.

- $f_{xy}^2 \to $ how much does the interaction of the dimensions muddle the story?
- if the interaction is really pronounced, then this influence can outweight the agreement along the first term.

In [7]:

```
c = 0 # There is no interaction in the y*x dimension
f = x**2 + y**2 + c*(x*y)
mv.plot_3d(f)
# Why is this??
```

In [8]:

```
c = 5 # increase c; no there is a significant intraction
f = x**2 + y**2 + c*(x*y)
mv.plot_3d(f)
# Why is this??
```

Let's actually apply this...

$$ \nabla f = \begin{pmatrix} 2x + y \\ 2y + x \end{pmatrix} $$

$$ \textbf{H} =\begin{pmatrix} 2 & 1 \\1 & 2\end{pmatrix}$$

$$ |\textbf{H}| = (2)(2) - (1)(1) = 3 $$

Okay, so it's an extrema, but which type? We'll to make this call, we just need to evaluate the signs of the second-order partial derivatives.

Given $|\textbf{H}| > 0$, then

if $f_{xx} = +$ and $f_{yy} = +$, $\textbf{x}^*$ is a

**local minima**.if $f_{xx} = -$ and $f_{yy} = -$, $\textbf{x}^*$ is a

**local maxima**.

So in the case of $f$, our stationary point is *local minima*. To actually get this minimum value, we plug the points back into $f$,

$$ f(x_0^*,y_0^*) = f(0,0) = 0^2 + 0^2 + 0(0) = 0 $$

The above is straightforward enough when dealing with two variables, but as we know, calculating the determinant for a matrix becomes increasingly more involved as the dimensions of the matrix grows: that is, as we begin to consider more and more variables.

Luckily, there is a familiar solution: leverage the **eigenvalues**!

Specifically, given that $|\textbf{H}| \ne 0$, we want to determine if the Hessian matrix is:

**positive definite**: if all eigenvalues for $\textbf{H}$ evaluated at $\textbf{x}^*$ are positive, then the matrix is positive definite and we can conclude that $\textbf{x}^*$ is a local minimum.**negative definite**: if all eigenvalues for $\textbf{H}$ evaluated at $\textbf{x}^*$ are positive, then the matrix is negative definite and we can conclude that $\textbf{x}^*$ is a local maxima.**indeterminate**: if the eigenvalues are a mix of positive and negative values.

Again if $|\textbf{H}| = 0$ then the second derivative test is inconclusive.

**Given our intuition of eigenvalues and vectors, why might this make sense?**

In summary, the steps:

- find $\nabla f$
- find stationary points: set $\nabla f = \textbf{0}$
- calculate the Hessian $\textbf{H}$
- calculate the determinant of $\textbf{H}$ for each stationary point.
- if $|\textbf{H}_{x*}| > 0$, a min/max exists
- if $|\textbf{H}_{x*}| < 0$, saddle point
- if $|\textbf{H}_{x*}| = 0$, undetermined

- calculate the eigenvalues of $\textbf{H}$ for all remaining stationary points.
- if all $\lambda > 0$, local minima.
- if all $\lambda < 0$, local maxima.
- if all $\lambda > 0$ and $\lambda < 0$, indeterminant .
- if all $\lambda = 0$, undefined.

- plug the remaining values back into $f$, find the highest or lowest.

`sympy`

in a multivariate optimization problem¶As we've seen before, sympy requires that we initialize a unique data type for its class of symbols.

In [9]:

```
x,y = symbols('x y') # This is repeat from above...
```

Let's find all critical points and determine whether they are minima, maxima, or neighter for the function

$$ g(x,y) = x^2 + 6xy + y^2 - 18x - 22y +5 $$

In [10]:

```
g = x**2 + 6*x*y + y**2 - 18*x - 22*y + 5
g
```

Out[10]:

In [11]:

```
# Step 1: Calculate the gradient
g_x = g.diff(x)
g_y = g.diff(y)
grad = [g_x,g_y]
grad
```

Out[11]:

In [12]:

```
# Step 2: Set gradient to 0 and locate the stationary points
crit_points = solve(grad,[x,y])
crit_points
```

Out[12]:

In [13]:

```
# Step 3: Calculate the Hessian
g_xx = g_x.diff(x)
g_xy = g_x.diff(y)
g_yy = g_y.diff(y)
g_yx = g_y.diff(x)
H = np.matrix([[g_xx,g_xy],[g_yx,g_yy]],dtype='float')
H
```

Out[13]:

In [14]:

```
# Step 4: Calculate the determinant of H
la.det(H)
```

Out[14]:

The determinant is negative! This means we're dealing with a saddle point. Let's inspect visually to see if this is, in fact, the case.

In [15]:

```
mv.plot_3d(g)
```

Sure enough there doesn't appear to be any emerging extrema

Let's find all critical points and determine whether they are minima, maxima, or neighter for the function

$$ t(x,y) = - x^{2} - 3 y^{2} + y $$

In [16]:

```
t = -x**2+y+-3*y**2
t
```

Out[16]:

In [17]:

```
grad = [t.diff(term) for term in [x,y]]
grad
```

Out[17]:

In [18]:

```
cp = solve(grad,[x,y])
cp # One critical point
```

Out[18]:

In [19]:

```
H = [t.diff(term).diff(term2) for term in [x,y] for term2 in [x,y]]
H = np.matrix(H,dtype='float').reshape(2,2)
H
```

Out[19]:

In [20]:

```
# Positive! Let's continue...
la.det(H)
```

Out[20]:

In [21]:

```
evals = la.eigvals(H)
evals
```

Out[21]:

The eigenvalues are all negative, so we can conclude that $x^* = (0,\frac{1}{6})$ is a global maxima.

Plug that value back into $t$

In [22]:

```
t.evalf(subs={x:0,y:(1/6)})
```

Out[22]:

Does that seem right?

In [23]:

```
mv.plot_3d(t)
```