Sample Research Questions
•
Designing a geometrical 3D shape (cylinder, cone, etc.) to maximize volume for fixed surface area
•
Minimizing the surface area required to construct a 3D shape that can hold a given volume of contents.
Problem Definition
An optimization problem can be formulated by incorporating the following elements.
Decision Variables
The controllable inputs of an objective function that represent decisions to be made within the problem.
Objective Function
Mathematical expression that calculates the objective value.
Objective Value
The value that the optimization problem seeks to maximize or minimize; evaluating the objective function by substituting decision variables yields the objective value.
Constraints
Logical conditions or real-world limitations that restrict the possible values of decision variables.
Substituting decision variables into and evaluating the objective function yields the objective value.
Optimization Process
The goal of constrained optimization is to fine-tune the decision variables so that they will yield the highest (for a maximization problem) or the lowest (for a minimization problem) objective value when substituted into the objective function.
Constraints screen for feasible solutions, which satisfy all equality/inequality conditions. Solutions that fail to satisfy even one of the constraints are considered infeasible and therefore excluded from consideration.
The objective function then evaluates and compares objective values corresponding to all feasible solutions, eventually selecting the optimal solution (set of decision variables) that yields the most desirable objective value.
In the mathematics IA, students should strive to define and solve an optimization problem that is more difficult than typical optimization problems they might encounter in the HL syllabus. For example, multivariate objective functions whose objective values will be simultaneously influenced by more than one input variable significantly raises the difficulty of the optimization problem.
Single Variate Optimization
The simplest approach to optimizing multivariate problems is using substitution to represent a multivariate function as a single variate function.
Suppose we seek to maximize the volume of a cylinder with fixed surface area by manipulating its radius and height.
Decision variables:
Objective Function:
Objective Value:
Constraint: , where is a constant value.
Maximize
subject to
Algebraically manipulate the constraint to express in terms of , then use substitution to convert into a single variate function.
Solve for to find radius value that maximizes volume. Once this value is found, substitute into to find the corresponding height.
Gradient Descent
Numerical methods may also be used to find the optimal value of a function.
Gradient
A continuously differentiable real-valued function increases the fastest in the direction of , while decreasing the fastest in the direction of .
Gradient Descent Algorithm
Gradient descent algorithm for single variate function:
Gradient descent algorithm for multivariate function:
Repeat until convergence (where movement from to offers negligible improvement to objective value that is below a predefined threshold value) while, simultaneously updating variable values, , , with each iteration of algorithm.
Investigation
Start from initial point , then compute gradient . For a problem that seeks to maximize , nudge the point towards the direction of maximal increase, , by an incremental step size . For a minimization problem, successive iterations should guide the point towards the direction of maximal decrease, , by step size .
After one iteration of the gradient descent algorithm the current solution is updated to . Compute the new gradient and move towards this direction by step size . Repeating these steps allows the initial point to incrementally crawl towards the optimal point as illustrated below.
At the optimal point, the gradient may be zero: indicates that the objective value can no longer be improved. The direction of maximal increase of being a zero vector would imply that the current position of is the maximum (or minimum) point.
Extension:
•
What are the pitfalls of gradient descent or other numerical methods for optimization?
•
Discuss how gradient descent algorithms may return a local maximum/minimum instead of the true global maximum/minimum.
•
Explore how the gradient descent algorithm may result in different outcomes depending on the initial point chosen.
Lagrange Multipliers Method
Constrained optimization problems in the form:
Maximize (or Minimize)
subject to
where and are smooth functions and is a scalar constant,
can be solved by finding that satisfy , where is a scalar constant.
Geometric Interpretation of Lagrange Multipliers
If the domain of was unrestricted, the maximum value of can be found at the unconstrained optimal point.
However, the constraint restricts the scope of search for possible solutions to points along the bounded curve (green).
Level curves (red) are analogous to contour lines; encompasses all that yield the same objective value (altitude) when substituted into the objective function .
The constrained optimal point is found at the intersection between the constraint curve (green) and one of the possible level curves (red) where the gradients are parallel to one another.
Mathematically speaking,