Hamilton–Jacobi–Bellman equation
The Hamilton–Jacobi–Bellman (HJB) equation is a partial differential equation which is central to optimal control theory. The solution of the HJB equation is the 'value function' which gives the minimum cost for a given dynamical system with an associated cost function.