Constrained optimization is all about finding the best solution while playing by the rules. It's like trying to make the most money possible, but you can't spend more than you have or break any laws.
This topic builds on earlier optimization concepts, adding real-world limitations to the mix. We'll explore how to solve problems with constraints, using methods like penalties and barriers to keep solutions in check.
Constrained Optimization Problems
Defining Constrained Optimization
- Constrained optimization problems involve finding the optimal solution to an objective function subject to one or more constraints on the decision variables
- Objective function represents the quantity to be maximized or minimized, typically denoted as , where is a vector of decision variables
- Constraints limit the feasible region of the optimization problem, often represented as equalities or inequalities
- Feasible region encompasses all possible solutions that satisfy all constraints in the optimization problem
- Standard form of a constrained optimization problem includes the objective function, equality constraints, and inequality constraints:
- Minimize (or Maximize)
- Subject to:
- for (equality constraints)
- for (inequality constraints)
- (domain constraints)
- Real-world applications include resource allocation (budgeting), portfolio optimization (finance), and engineering design problems (structural optimization)
- Solution to a constrained optimization problem must satisfy both the optimality criteria for the objective function and all specified constraints
Components and Characteristics
- Decision variables () represent the quantities to be optimized (production levels, investment allocations)
- Objective function () quantifies the goal to be optimized (maximize profit, minimize cost)
- Constraints define limitations on decision variables:
- Resource constraints (budget limitations)
- Physical constraints (capacity restrictions)
- Logical constraints (binary decisions)
- Feasible region shape depends on constraint types:
- Linear constraints create a polyhedron
- Nonlinear constraints may result in a non-convex region
- Global optimum may not be achievable due to constraints, leading to local optima
- Constraint qualification ensures the existence of Lagrange multipliers at the optimal solution
- Sensitivity analysis examines how changes in constraints affect the optimal solution
Constraint Types
Equality and Inequality Constraints
- Equality constraints expressed as , where is a function of the decision variables that must equal zero at the optimal solution
- Example: (total production must equal 100 units)
- Inequality constraints expressed as or , defining regions where the solution must lie above or below a certain threshold
- Example: (resource usage must not exceed 50 units)
- Active constraints satisfied as equalities at the optimal solution, while inactive constraints satisfied as strict inequalities
- Active constraint example: g(x^*) = 0$ at the optimal point $x^*
- Inactive constraint example: g(x^*) < 0$ at the optimal point $x^*
- Linear constraints expressed as linear functions of the decision variables
- Example: (linear equality constraint)
- Nonlinear constraints involve nonlinear functions
- Example: (circular constraint region)
Bound Constraints and Convexity
- Bound constraints directly limit the range of individual decision variables, often expressed as
- Example: (production of item 1 between 0 and 10 units)
- Convex constraints define a convex feasible region, crucial for guaranteeing global optimality in convex optimization problems
- Convex set property: line segment between any two points in the set lies entirely within the set
- Number and type of constraints significantly impact the complexity and solvability of the optimization problem
- Highly constrained problems may have a very small feasible region
- Conflicting constraints may result in an infeasible problem
- Constraint classification affects algorithm choice:
- Linear constraints allow for efficient linear programming methods
- Nonlinear constraints may require more complex nonlinear programming techniques
Optimality Conditions
Karush-Kuhn-Tucker (KKT) Conditions
- Karush-Kuhn-Tucker (KKT) conditions represent first-order necessary conditions for a solution to be optimal in nonlinear programming
- KKT conditions include:
- Stationarity
- Complementary slackness
- Dual feasibility
- Primal feasibility
- Lagrange multipliers ( and ) introduced in KKT conditions to account for the impact of constraints on the objective function
- Stationarity condition relates the gradient of the objective function to the gradients of the active constraints at the optimal point:
- Complementary slackness ensures either an inequality constraint is active or its corresponding Lagrange multiplier is zero:
- for all
Additional Optimality Criteria
- Dual feasibility requires non-negative Lagrange multipliers for inequality constraints:
- for all
- Primal feasibility ensures all constraints are satisfied:
- for all
- for all
- Second-order sufficient conditions, involving the Hessian of the Lagrangian, can be used to verify that a KKT point is indeed a local minimum
- Lagrangian function:
- Second-order condition: Hessian of Lagrangian is positive definite on the tangent space of active constraints
- Fritz John conditions provide an alternative set of necessary conditions when constraint qualification fails
- Slater's condition ensures strong duality in convex optimization problems, guaranteeing zero duality gap
Handling Constraints in Optimization
Penalty and Barrier Methods
- Penalty methods transform constrained optimization problems into unconstrained problems by adding a penalty term to the objective function for constraint violations
- Quadratic penalty method uses a quadratic function of constraint violations, with the penalty parameter increasing iteratively to enforce feasibility:
- Penalized objective:
- Augmented Lagrangian methods combine penalty terms with Lagrange multiplier estimates to improve convergence properties:
- Augmented Lagrangian:
- Barrier methods, such as the logarithmic barrier method, add terms to the objective function that prevent solutions from violating inequality constraints:
- Barrier function:
- Interior point method uses barrier functions to solve linear and nonlinear programming problems efficiently
- Primal-dual interior point methods simultaneously optimize primal and dual variables
Projection and Active Set Methods
- Projection methods handle constraints by projecting infeasible points onto the feasible region after each iteration of an unconstrained optimization algorithm
- Example: Projected gradient method for bound-constrained problems
- Active set methods identify the set of active constraints at each iteration and solve a sequence of equality-constrained subproblems
- Working set: estimate of the active set at each iteration
- Equality-constrained subproblem solved using Lagrange multipliers
- Choice of constraint-handling strategy depends on:
- Problem structure (linear vs. nonlinear)
- Constraint types (equality, inequality, bounds)
- Desired trade-offs between solution accuracy and computational efficiency
- Hybrid methods combine multiple techniques to leverage their strengths:
- Sequential Quadratic Programming (SQP) uses active set approach with quadratic programming subproblems
- Augmented Lagrangian methods with bound constraints handled separately