Dual Gradient Descent

TagsConstrained

Dual Gradient Descent

This is how you can turn a constrained optimization problem into an unconstrained.

Let’s say you wanted to do

maxxf(x)s.t. g(x)0\max_x f(x) \\ \text{s.t.} ~g(x) \geq 0

You can set up the lagrangian:

L(x,λ)=f(x)λg(x)\mathcal{L}(x,\lambda) = f(x) - \lambda g(x)

Now, we propose that we can yield a constrained solution with the following objective:

maxxminλ0[f(x)λg(x)]\max_x \min_{\lambda \geq 0} [f(x) -\lambda g(x)]

Why? Well, let’s set up two scenarios

Therefore, we see that using this objective, the only viable solution will follow the constraint

Implementation

In practice, we just alternate a gradient WRT λ\lambda and the policy.