In our case we may use only phase I or phase I and II
by considering the optimum problems which are to
determine the minimum and maximum of the
unknowns under the
constraints
and update the interval for an
unknown if the simplex applied to minimize or maximize enable to
improve the range. It may be seen that this is a
recursive procedure: an improvement on one variable change the
constraint equations and may thus change the result of the simplex
method applied for determining the extremum of a variable which has
already been considered.
This procedure, proposed in [25], enable to correct one of the drawback of the general solving procedures: each equation is considered independently and for given intervals for the unknowns two equations may have an interval evaluation that contain 0 although these equations cannot be canceled at the same time. The previous method enable to take into account at least partly the dependence of the equations. Clearly it will more efficient if the functions has a large number of linear terms and a "small" non-linear part.
In all of the following procedures the various storage mode and bisection mode of the general solving procedures may be used and inequalities are handled.