Before an optimizer actually performs the optimization the problem is preprocessed using the so-called presolve. The purpose of the presolve is to
- remove redundant constraints,
- eliminate fixed variables,
- remove linear dependencies,
- substitute out (implied) free variables, and
- reduce the size of the optimization problem in general.
After the presolved problem has been optimized the solution is automatically postsolved so that the returned solution is valid for the original problem. Hence, the presolve is completely transparent. For further details about the presolve phase, please see [AA95] and [AGMX96].
It is possible to fine-tune the behavior of the presolve or to turn it off entirely. If presolve consumes too much time or memory compared to the reduction in problem size gained it may be disabled. This is done by setting the parameter
presolvemode.off. The two most time-consuming steps of the presolve are
- the eliminator, and
- the linear dependency check.
Therefore, in some cases it is worthwhile to disable one or both of these.
Numerical issues in the presolve
During the presolve the problem is reformulated so that it hopefully solves faster. However, in rare cases the presolved problem may be harder to solve then the original problem. The presolve may also be infeasible although the original problem is not. If it is suspected that presolved problem is much harder to solve than the original, we suggest to first turn the eliminator off by setting the parameter
iparam.presolve_eliminator_max_num_tries to 0. If that does not help, then trying to turn entire presolve off may help.
Since all computations are done in finite precision, the presolve employs some tolerances when concluding a variable is fixed or a constraint is redundant. If it happens that MOSEK incorrectly concludes a problem is primal or dual infeasible, then it is worthwhile to try to reduce the parameters
dparam.presolve_tol_s. However, if reducing the parameters actually helps then this should be taken as an indication that the problem is badly formulated.
The purpose of the eliminator is to eliminate free and implied free variables from the problem using substitution. For instance, given the constraints
y is an implied free variable that can be substituted out of the problem, if deemed worthwhile. If the eliminator consumes too much time or memory compared to the reduction in problem size gained it may be disabled. This can be done by setting the parameter
iparam.presolve_eliminator_max_num_tries to 0. In rare cases the eliminator may cause that the problem becomes much hard to solve.
Linear dependency checker
The purpose of the linear dependency check is to remove linear dependencies among the linear equalities. For instance, the three linear equalities
contain exactly one linear dependency. This implies that one of the constraints can be dropped without changing the set of feasible solutions. Removing linear dependencies is in general a good idea since it reduces the size of the problem. Moreover, the linear dependencies are likely to introduce numerical problems in the optimization phase. It is best practice to build models without linear dependencies, but that is not always easy for the user to control. If the linear dependencies are removed at the modelling stage, the linear dependency check can safely be disabled by setting the parameter
All linear, conic, and convex optimization problems have an equivalent dual problem associated with them. MOSEK has built-in heuristics to determine if it is more efficient to solve the primal or dual problem. The form (primal or dual) is displayed in the MOSEK log and available as an information item from the solver. Should the internal heuristics not choose the most efficient form of the problem it may be worthwhile to set the dualizer manually by setting the parameters:
iparam.intpnt_solve_form: In case of the interior-point optimizer.
iparam.sim_solve_form: In case of the simplex optimizer.
Note that currently only linear and conic quadratic problems may be automatically dualized.
Problems containing data with large and/or small coefficients, say \(1.0e+9\) or \(1.0e-7\), are often hard to solve. Significant digits may be truncated in calculations with finite precision, which can result in the optimizer relying on inaccurate data. Since computers work in finite precision, extreme coefficients should be avoided. In general, data around the same order of magnitude is preferred, and we will refer to a problem, satisfying this loose property, as being well-scaled. If the problem is not well scaled, MOSEK will try to scale (multiply) constraints and variables by suitable constants. MOSEK solves the scaled problem to improve the numerical properties.
The scaling process is transparent, i.e. the solution to the original problem is reported. It is important to be aware that the optimizer terminates when the termination criterion is met on the scaled problem, therefore significant primal or dual infeasibilities may occur after unscaling for badly scaled problems. The best solution of this issue is to reformulate the problem, making it better scaled.
By default MOSEK heuristically chooses a suitable scaling. The scaling for interior-point and simplex optimizers can be controlled with the parameters