SecondOrderConeOptimization
SecondOrderConeOptimization[f,cons,vars]
finds values of variables vars that minimize the linear objective f subject to secondorder cone and/or linear constraints cons.
SecondOrderConeOptimization[c,{{a_{1},b_{1},α_{1},β_{1}},…,{a_{k},b_{k},α_{k},β_{k}}}]
finds a vector that minimizes subject to the constraints .
SecondOrderConeOptimization[c,…,{dom_{1},dom_{2},…}]
takes to be in the domain dom_{i}, where dom_{i} is Integers or Reals.
SecondOrderConeOptimization[…,"prop"]
specifies what solution property "prop" should be returned.
Details and Options
 Secondorder cone optimization is also known as secondorder cone programming (SOCP).
 Secondorder cone optimization is used in problems such as parameter fitting and geometric distance problems.
 Secondorder cone optimization is a convex optimization problem that can be solved globally and efficiently with real, integer or complex variables.
 Secondorder cone optimization finds that solves the primal problem:

minimize subject to constraints where  Mixedinteger secondorder cone optimization finds and that solve the problem:

minimize subject to constraints where  When the objective function is real valued, SecondOrderConeOptimization solves problems with by internally converting to real variables , where and .
 The variable specification vars should be a list with elements giving variables in one of the following forms:

v variable with name and dimensions inferred v∈Reals real scalar variable v∈Integers integer scalar variable v∈ℛ vector variable restricted to the geometric region v∈Vectors[n,dom] vector variable in or v∈Matrices[{m,n},dom] matrix variable in or  Note that linear constraints may be expressed using secondorder constraints. For convenience, a single 0 is interpreted as needed as a matrix or vector with zero entries:

{a_{i},b_{i},0,0} linear equality constraints {0,0,α_{i},β_{i}} linear inequality constraint  The constraints cons can be specified by:

LessEqual scalar inequality GreaterEqual scalar inequality VectorLessEqual vector inequality VectorGreaterEqual vector inequality Equal scalar or vector equality Element convex domain or region element  With SecondOrderConeOptimization[f,cons,vars], parameter equations of the form parval, where par is not in vars and val is numerical or an array with numerical values, may be included in the constraints to define parameters used in f or cons.
 The primal minimization problem has a related maximization problem that is the Lagrangian dual problem. The dual maximum value is always less than or equal to the primal minimum value, so it provides a lower bound. The dual maximizer provides information about the primal problem, including sensitivity of the minimum value to changes in the constraints.
 The secondorder cone optimization has a dual:

maximize subject to constraints where  For secondorder cone optimization, strong duality always holds, meaning that if there is a solution to the primal minimization problem, then there is a solution to the dual maximization problem, and the dual maximum value is equal to the primal minimum value.
 The possible solution properties "prop" include:

"PrimalMinimizer" a list of variable values that minimizes the objective function "PrimalMinimizerRules" values for the variables vars={v_{1},…} that minimize "PrimalMinimizerVector" the vector that minimizes "PrimalMinimumValue" the minimum value "DualMaximizer" the maximizing dual vector "DualMaximumValue" the dual maximizing value "DualityGap" the difference between the dual and primal optimal values "Slack" vector that converts inequality constraints to equality "ConstraintSensitivity" sensitivity of to constraint perturbations "ObjectiveVector" the linear objective vector "SecondOrderConeConstraints" the list of coefficients for the constraints "LinearEqualityConstraints" the linear equality constraint matrix and vector {"prop_{1}","prop_{2}",…} several solution properties  The following options may be given:

MaxIterations Automatic maximum number of iterations to use Method Automatic the method to use PerformanceGoal $PerformanceGoal aspects of performance to try to optimize Tolerance Automatic the tolerance to use for internal comparisons  The option Method>method may be used to specify the method to use. Available methods include:

Automatic choose the method automatically "SCS" SCS splitting conic solver "CSDP" CSDP semidefinite optimization solver "DSDP" DSDP semidefinite optimization solver "PolyhedralApproximation" approximate constraints using polyhedra "MOSEK" commercial MOSEK convex optimization solver "Gurobi" commercial Gurobi linear and quadratic optimization solver "Xpress" commercial Xpress linear and quadratic optimization solver  Computations are limited to MachinePrecision.
Examples
open allclose allBasic Examples (4)
Minimize subject to the constraint :
The optimal point lies in a region defined by the constraints and where is smallest within the region:
Minimize subject to the constraints :
The optimal point lies on the smallest value of in a region bounded by the intersection of two disks:
Minimize subject to the constraints . Specify the inputs in matrixvector form:
Solve the problem using matrixvector form:
Scope (27)
Basic Uses (9)
Minimize subject to the constraints :
Get the minimizing value using solution property "PrimalMinimiumValue":
Minimize subject to the constraints :
Get the minimizing value and minimizing vector using solution properties:
Minimize subject to the constraints :
Define the objective as and the constraints as :
Solve using matrixvector inputs:
Minimize subject to the constraints :
Define the objective as and the constraints as :
Specify the equality constraint as:
Solve using matrixvector inputs:
Minimize subject to the constraint using constant parameter equations for a and b:
Minimize subject to the constraint using vector variable :
Minimize subject to the constraint :
Specify the constraints using VectorGreaterEqual () and VectorLessEqual ():
Minimize subject to . Use Inactive[Plus] to avoid unintentional threading:
Specify the matrix and vector as parametrized equations to avoid using Inactive:
Minimize subject to . Use NonNegativeReals to specify constraint :
Integer Variables (4)
Specify integer domain constraints using Integers:
Specify integer domain constraints on vector variables using Vectors[n,Integers]:
Specify nonnegative integer domain constraints using NonNegativeIntegers ():
Specify nonpositive integer domain constraints using NonPositiveIntegers ():
Complex Variables (4)
Primal Model Properties (3)
Minimize subject to the constraints :
Get the primal minimizer as a vector:
Obtain the primal minimum value using symbolic inputs:
Extract the matrixvector inputs of the optimization problem:
Solve using matrixvector form:
Recover the symbolic primal value by adding the objective constant:
Find the slack associated with a minimization problem:
Dual Model Properties (3)
Minimize subject to : »
The dual problem is to maximize subject to and :
Solve the resulting maximization problem:
The primal minimum value and the dual maximum value coincide because of strong duality:
That is the same as having a duality gap of zero. In general :
Construct the dual problem using constraint matrices extracted from the primal problem: »
The dual problem is to maximize subject to and :
Solve the resulting maximization problem:
Get the dual maximum value directly using solution properties:
Sensitivity Properties (4)
Use "ConstraintSensitivity" to find the change in optimal value due to constraint relaxations:
Each element of the sensitivity is , where is a vector and is a scalar:
Consider the new constraints , where is the relaxation for constraint :
The sensitivity element is associated with relaxation of . The expected new optimal value is:
Compare by directly solving the relaxed problem:
Consider the new constraints :
The sensitivity element is associated with relaxation of . The expected new optimal value is:
Compare by directly solving the relaxed problem:
Each sensitivity is associated with an inequality or equality constraint:
Extract the secondorder cone constraints. The constraint is given as :
Linear constraints are given as . The associated sensitivity will have :
The constraints and their associated sensitivities:
The change in optimal value due to constraint relaxation is proportional to the sensitivity value:
Compute the minimal value and constraint sensitivity:
A zero sensitivity will not change the optimal value if the constraint is relaxed:
A negative sensitivity will decrease the optimal value:
A positive sensitivity will increase the optimal value:
The "ConstraintSensitivity" is related to the dual maximizer of the problem:
The sensitivity satisfies , where is the dual inequality maximizer:
Options (8)
Method (4)
The method "SCS" uses the splitting conic solver method:
"CSDP" and "DSDP" reduce to semidefinite optimization and use the interior point method:
"PolyhedralApproximation" approximates secondorder constraints using polyhedra that have representations in terms of linear constraints:
The method "SCS" may give less accurate results than "CSDP" or "DSDP" due to relaxed tolerance:
Use a smaller tolerance to get better results for "SCS":
The method "PolyhedralApproximation" may have longer computation times than other methods:
Different methods may give different results for problems with more than one optimal solution:
PerformanceGoal (1)
Tolerance (3)
A smaller Tolerance setting gives a more precise result:
Find the error between the computed and exact minimum value using different Tolerance settings:
Visualize the change in minimum value error with respect to Tolerance:
A smaller Tolerance setting gives a more precise answer, but typically takes longer to compute:
A smaller tolerance takes longer:
The tighter tolerance gives a more precise answer:
Different methods have different tolerances, which affects the accuracy and precision:
Applications (35)
Basic Modeling Transformations (11)
Maximize subject to . Solve the maximization problem by negating the objective function:
Negate the primal minimum value to get the corresponding maximum value:
Minimize subject to the constraints . Using the auxiliary variable , transform the problem to minimize subject to :
Recover the original minimum value by performing direct substitution into :
Minimize subject to the constraints . Using the auxiliary variable , transform the problem to minimize subject to :
Recover the original minimum value by performing direct substitution into :
Minimize subject to . Using two auxiliary variables , transform the problem to minimize subject to :
Minimize subject to . Using the auxiliary variable , transform the problem to minimize subject to :
Minimize subject to . For , is equivalent to :
This is also equivalent to minimizing subject to :
SecondOrderConeOptimization directly does the transformation for Abs:
Minimize subject to the constraints . Using the auxiliary variable , transform the problem to minimize subject to :
SecondOrderConeOptimization directly does the transformation for Abs:
Minimize . Using auxiliary variables , convert the problem to minimize subject to the constraints :
Directly using Abs in the constraints will be faster and more accurate:
Minimize . Using auxiliary variable , convert the problem to minimize subject to the constraints :
Minimize subject to , where is a nondecreasing function, by instead minimizing . The primal minimizer will remain the same for both problems. Consider minimizing subject to :
The true minimum value can be obtained by applying the function to the primal minimum value:
Since , an additional constraint of must be applied:
The true minimum value can be obtained by applying the function to the primal minimum value:
Geometric Problems (6)
Find the minimum distance between two disks of radius 1 centered at (0,0) and (2,1). Using auxiliary variable , the transformed problem is to minimize subject to :
Visualize the positions of the two points:
The auxiliary variable gives the distance between the points:
Find the minimum distance between two nonintersecting convex polygons:
Let be a point on . Let be a point on . The objective is to minimize . Using auxiliary variable , the transformed objective is to minimize subject to :
Visualize the position of the points and :
The auxiliary variable gives the distance between the points:
Find the plane that separates two nonintersecting convex polygons:
Let be a point on . Let be a point on . The objective is to minimize . Using auxiliary variable , the transformed objective is to minimize subject to :
According to the separating hyperplane theorem, the dual associated with the constraint will give the normal of the hyperplane:
The constraint internally has the form . is the only constraint containing a matrix :
The hyperplane is constructed as:
Visualize the plane separating the two polygons:
Find the plane that separates two nonintersecting convex polygons:
Let be a point on . Let be a point on . The objective is to minimize . Using auxiliary variable , the transformed objective is to minimize subject to :
According to the separating hyperplane theorem, the dual associated with the constraint will give the normal of the hyperplane:
The constraint internally has the form . is the only constraint containing a matrix :
The hyperplane is constructed as:
Visualize the plane separating the two polygons:
Find the center and radius of the smallest circle that encloses a set of points:
The objective is to minimize the radius subject to the constraints :
The center and radius of the disk are:
The minimal enclosing disk can be found efficiently using BoundingRegion:
Find the radius and center of a minimal enclosing ball that encompasses a given region:
Minimize the radius subject to the constraints :
The minimal enclosing ball can be found efficiently using BoundingRegion:
Facility Location Problems (3)
A company wants to open a new factory. The factory needs raw materials from five warehouses. The transportation cost per unit distance between the new factory and warehouses is:
Let be the distance between factory and warehouses The objective is to minimize :
The five warehouses are located at:
The new factory must be located such that :
Find the optimal distances between the new factory and the warehouses:
The new factory location is closer to the warehouses where the transportation costs are higher:
A company wants to open a two new factories. The factories need raw materials from five warehouses. The transportation cost per unit distance between the new factories and warehouses is:
Let be the distance between factory and warehouse The objective is to minimize :
The five warehouses are located at:
The new factories must be located such that :
Find the optimal distances between the new factory and the warehouses:
The new factory location is closer to the warehouses where the transportation costs are higher:
Find the minimum number and location of warehouses to construct that can supply the company's five retail stores. There are five potential sites. The cost of transporting one unit of product from a given warehouse to a given store is given by a table:
The positions of the retail stores are:
The demand in the retail stores is:
For every warehouse, there is a $1000 overhead cost:
Each warehouse can hold a maximum of 500 units of product:
Let be the number of units that is transported from warehouse to store . The demand constraint is:
Let be a binary decision vector such that if , then warehouse is constructed. The supply constraint is:
The new warehouses must be located such that :
The objective is to minimize the number of warehouses, the transportation cost of goods and the distance of the warehouse from the retail store:
Solve the minimum number and positions of the warehouses:
The number of units to be transported from warehouse to store is:
DataFitting Problems (6)
Find a linear fit to discrete data by minimizing subject to the constraints :
Using auxiliary variable , minimize subject to :
Fit a fifthorder polynomial to discrete data by minimizing :
Construct the input data matrix using DesignMatrix:
Find the coefficients of the quadratic curve by minimizing subject to :
Find an approximating function to noisy data using bases such that the first and last points lie on the curve:
The approximating function will be :
Find the coefficients of the approximating function by minimizing subject to :
Find a robust linear fit to discrete data by minimizing , where is a known regularization parameter:
Fit the line by minimizing subject to :
A plot of the line shows that it is robust to the outliers:
The robust fitting can also be obtained using the function Fit:
Excess regularization will lead to the fit becoming increasingly insensitive to data:
Cardinality constrained least squares: minimize such that has at most nonzero elements:
Let be a decision vector such that if , then is nonzero. The decision constraints are:
To model constraint when , chose a large constant such that :
Using epigraph transformation, the objective is to minimize subject to :
Solve the cardinality constrained leastsquares problem:
The subset selection can also be done more efficiently with Fit using regularization. First, find the range of regularization parameters that uses at most basis functions:
Find the nonzero terms in the regularized fit:
Find the fit with just these basis terms:
Find the best subset of functions from a candidate set of functions to approximate given data:
The approximating function will be :
A maximum of 5 basis functions are to be used in the final approximation:
The coefficients associated with functions that are not chosen must be zero:
Using epigraph transformation, the objective is to minimize such that :
Classification Problems (2)
Find a quartic polynomial that robustly separates two sets of points in the plane:
The polynomial is defined as , with :
For separation, set 1 must satisfy and set 2 must satisfy :
The objective is to minimize , which gives twice the thickness between and . Using the auxiliary variable , the objective function is transformed to the constraint :
Find the coefficients by minimizing :
Visualize the separation of the sets using the polynomial:
The value is the width in terms of level values of , where there are no points:
Find a quartic polynomial that robustly separates two sets of points in 3D:
The polynomial is defined as , with :
For separation, set 1 must satisfy and set 2 must satisfy :
The objective is to minimize . Using transformation , the objective is now to minimize :
Structural Optimization Problems (1)
Find the shape of a hanging chain formed by spring links under a vertical load at the end of each link. The objective is to find the link positions :
The potential energy due to gravity is , where is the vertical load at each end and is gravity:
The potential energy due to the tension in the spring link due to stretching is , where is the stretch in spring link and is the stiffness of the spring. Using , the energy is transformed to :
An additional constraint must be added due to the transformation:
The ends of the linked chain are clamped at positions (0,0) and (2,1):
Each link has to satisfy the condition , where is the resting length of each spring:
The final objective function is the sum of gravity and spring potential energy that must be minimized:
Find the end points of each of the spring links:
Visualize the shape of the resulting spring chain:
The stretch is greatest for the links near the ends of the link chain. Links 11 and 12 have the least elongation:
Portfolio Optimization (1)
Find the distribution of capital to invest in six stocks to maximize return while minimizing risk:
Let be the percentage of total investment made in stock . The return is given by where is a vector of expected return value of each individual stock:
The risk is given by , is a riskaversion parameter and :
The objective is to maximize return while minimizing risk for a specified riskaversion parameter:
The percentage of investment must be greater than 0 and must add to 1:
Compute the returns and corresponding risk for a range of riskaversion parameters:
The optimal over a range of gives an upperbound envelope on the tradeoff between return and risk:
Compute the percentage of investment for a specified number of riskaversion parameters:
Increasing the riskaversion parameter causes the investment to be diversified:
Increasing the riskaversion parameter causes the return on investment to decrease:
Nonnegative Matrix Factorization (1)
Investment Problems (3)
Find the number of stocks to buy from four stocks such that a minimum $1000 dividend is received and risk is minimized. The expected return value and the covariance matrix associated with the stocks are:
The unit price for the four stocks is $1. Each stock can be allocated a maximum of $2500:
The investment must yield a minimum of $1000:
A negative stock cannot be bought:
Using epigraph transformation, the risk, given by , is converted into a constraint:
The total amount to spend on each stock is found by minimizing the risk given by:
The total investment to get a minimum of $1000 is:
Find the number of stocks to buy from four stocks with an option to shortsell such that a minimum dividend of $1000 is received and the overall risk is minimized:
The capital constraint, return on investment constraints and risk constraint are:
A shortsale option allows the stock to be sold. The optimal number of stocks to buy/shortsell is found by minimizing the risk given by the objective :
The second stock can be shortsold. The total investment to get a minimum of $1000 due to the shortselling is:
Without shortselling, the initial investment will be significantly more:
Find the best combination of six stocks to invest in out of a possible 20 candidate stocks, so as to maximize return while minimizing risk:
Let be the percentage of total investment made in stock . The return is given by , where is a vector of the expected return value of each individual stock:
The risk is given by , is a riskaversion parameter and :
The objective is to maximize return while minimizing risk for a specified riskaversion parameter:
Let be a decision vector such that if , then that stock is bought. Six stocks have to be chosen:
The percentage of investment must be greater than 0 and must add to 1:
Find the optimal combination of stocks that minimizes the risk while maximizing return:
The optimal combination of stocks is:
The percentages of investment to put into the respective stocks are:
Trajectory Optimization Problems (1)
Find the shortest path between two points while avoiding obstacles. Specify the obstacle:
Extract the halfspaces that form the convex obstacle:
Specify the start and end points of the path:
The path can be discretized using points. Let represent the position vector:
The objective is to minimize . Using epigraph transformation, the objective can be transformed to subject to the constraints :
Specify the end point constraints:
The distance between any two subsequent points should not be too large:
A point is outside the object if at least one element of is less than zero. To enforce this constraint, let be a decision vector and be the element of such that , then and is large enough such that :
Find the minimum distance path around the obstacle:
To avoid potential crossings at the edges, the region can be inflated and the problem solved again:
Properties & Relations (8)
SecondOrderConeOptimization gives the global minimum of the objective function:
Visualize the objective function:
Minimize gives global exact results for quadratic optimization problems:
NMinimize can be used to obtain approximate results using global methods:
FindMinimum can be used to obtain approximate results using local methods:
LinearOptimization is a special case of SecondOrderConeOptimization:
QuadraticOptimization is a special case of SecondOrderConeOptimization:
Use auxiliary variable and minimize with the additional constraint :
SemidefiniteOptimization is a generalization of SecondOrderConeOptimization:
ConicOptimization is a generalization of SecondOrderConeOptimization:
Possible Issues (5)
Constraints specified using strict inequalities may not be satisfied for certain methods:
The reason is that SecondOrderConeOptimization solves :
The minimum value of an empty set or infeasible problem is defined to be :
The minimizer is Indeterminate:
The minimum value for an unbounded set or unbounded problem is :
The minimizer is Indeterminate:
Dual related solution properties for mixedinteger problems may not be available:
Constraints with complex values need to be specified using vector inequalities:
Just using Less will not work even when both sides are real in theory:
Text
Wolfram Research (2019), SecondOrderConeOptimization, Wolfram Language function, https://reference.wolfram.com/language/ref/SecondOrderConeOptimization.html (updated 2020).
CMS
Wolfram Language. 2019. "SecondOrderConeOptimization." Wolfram Language & System Documentation Center. Wolfram Research. Last Modified 2020. https://reference.wolfram.com/language/ref/SecondOrderConeOptimization.html.
APA
Wolfram Language. (2019). SecondOrderConeOptimization. Wolfram Language & System Documentation Center. Retrieved from https://reference.wolfram.com/language/ref/SecondOrderConeOptimization.html