Contents
TogglePost-optimal sensitivity analysis
When the basic fix optimal of the PL problem is analyzed to answer questions about changes in its formulation, the study is called post-optimal sensitivity analysis.
We call post-optimization all the techniques making it possible to obtain the optimum of the PL problem when certain data have undergone modifications.
We consider the problem of linear programming general in its stand art form:
This study can be motivated by several reasons:
- the data of the problem is not known with exactitude, in which case it is important to determine to what extent this affects the proposed solution;
- we want to assess the consequences of a policy change that would modify the facts of the problem.
Marginal costs
If the problem posed consists in transforming goods to sell a production with a better profit and the maximum increase in income which results from the possibility of having an additional unit of one of the goods, is the marginal value of this good . Very often, the qualifier marginal cost is also used in this case.
If a difference variable is not zero, in the optimal solution, it means that the corresponding good is already surplus. Therefore, having an additional unit of this good will have no influence on the income. The slack variable has zero marginal value.
On the other hand, if a variance variable is zero in the optimal solution, it is because the corresponding good is totally used. Subsequently, a variation in availability will generally have an influence on income. This is why this zero deviation variable in the optimal solution has a non-zero marginal value, and this marginal value specifies the variation of the economic function resulting from the use of an additional unit of the associated good.
with the solution vector x* = (2,6). Be careful here the line shows the value of Z and not -Z (hence the positive values).
We can measure the sensitivity of the optimal solution to a change in a line term or a coefficient in the objective.
Study 1: variation in objective function
The variation of a coefficient in the objective function over a certain interval does not lead to a modification of the optimal solution. Outside this interval, we have a new solution which itself remains optimal on another interval. We can thus highlight a finite number of variation intervals for cj, with on each of them an invariant solution.
Let's change the objective function by max z' =4*x1 + 5 * x2. The value of the objective function will change by x *1 = 2, while the solution vector will not be modified as shown in graphics resolution :
Likewise if c1 goes from 3 to 2, only the value of the objective function will be modified. To calculate the interval over which the coefficient x*1 is valid, we need to scale the objective function until it is parallel to the other constraints.
That is, when the slope of the objective function is equal to the slope of the saturated constraints for the solution vector s*:
- z = c1* x1 + 5 * x2 and 2 * x2 = 12 so -c1/ 5 = 0, c1 = 0;
- z = c1* x1 + 5 * x2 and 3 * x1 + 2 * x2 = 18 so -c1/ 5 = -3/2, c1 =15/2.
The coefficient x *1 is therefore valid for c1 between 0 and 15/2.
When the problem is large, it is possible to calculate the cost variation using the simplex by adding a delta on the cost to vary as in the following example:
The solution remains optimal as long as the line of -Z has negative numbers so if:
Study 2: variation in the second limb
When the second member of a constraint varies (in a certain interval), if this constraint was not saturated, then the solution does not change and neither does the optimal value of the objective function. This result is obvious since the optimal solution does not satisfy the constraint with equality, we can vary (a little) the second member without “touching” the optimal solution.
On the other hand, if the constraint were checked with equality with the optimum, one has an interval of variation for the second member such as:
- The solution changes but the zero variables remain zero and the non-zero variables remain non-zero: the structure of the solution does not change.
- The variation of the second member i causes a variation of the optimal value of the objective function equal to ui* di, therefore proportional to di.
If we leave the interval, we have a new dual cost. We can thus highlight a finite number of intervals of variation for the second member with, on each of them, a value for the dual cost. On the different intervals, the sensitivity analysis does not give the optimal solution since the numerical values of the variables depend on the exact value of the second member.
Consider in the example that b1 =4 becomes b'1 = 5. Let's perform a graphic resolution of the new problem:
When is the decrease in b1 ? As the explanation on marginal costs explains, decreasing the second member by 1 causes the value of the objective function to decrease by an amount equal to marginal cost. So the decrease of 1 will not lead to modification.
To know the possibilities of evolution of the stock without changing the value of the optimal solution, it is necessary to add a delta in the second member studied as shown in the following example:
The solution remains optimal as long as the simplex is not degenerate, that is to say that the second member is positive:
Now consider an augmentation of the third right hand side to b'3 = 19. Since the marginal cost is not zero, the optimal solution will be modified as shown in the graphical resolution.
Study 3: variation of non-base variables
Let's go back to the previous example with a new constraint and a new variable:
We have the following reduced cost: d3 = -2. This means that to build a unit of x3 would decrease the value of the objective function by 2 (since it is out of basis x*3 = 0).
Study 4: variation in production
Indeed by adding a delta variable in the cost of the variable based on the targeted constraint then it suffices to inject the optimal vector and solve the equation as in the following example:
