APP下载

Process optimization with consideration of uncertainties—An overview

2018-09-28YingChenZhihongYuanBingzhenChen

Ying Chen,Zhihong Yuan,Bingzhen Chen*

Department of Chemical Engineering,Tsinghua University,Beijing 10084,China

Keywords:Optimization under uncertainty Robust optimization Stochastic programming Chance constrained programming Data-driven optimization

ABSTRACT Optimization under uncertainty is a challenging topic of practical importance in the Process Systems Engineering.Since the solution of an optimization problem generally exhibits high sensitivity to the parameter variations,the deterministic model which neglects the parametric uncertainties is not suitable for practical applications.This paper provides an overview of the key contributions and recent advances in the field of process optimization under uncertainty over the past ten years and discusses their advantages and limitations thoroughly.The discussion is focused on three specific research areas,namely robust optimization,stochastic programming and chance constrained programming,based on which a systematic analysis of their applications,developments and future directions are presented.It shows that the more recent trend has been to integrate different optimization methods to leverage their respective superiority and compensate for their drawbacks.Moreover,data-driven optimization,which combines mathematical programming methods and machine learning algorithms,has become an emerging and competitive tool to handle optimization problems in the presence of uncertainty based on massive historical data.

1.Introduction

Over the past several decades,people have realized the significance of accounting for uncertainty in the process optimization.Uncertainties are inevitable in the realistic chemical industries,ranging from process design and synthesis,to operational planning,production scheduling and supply chain optimization.An optimal production strategy derived from the deterministic model may become suboptimal or even totally infeasible owing to a slight “drift”around nominal values.When confronted with numerous parameters subjected to uncertainty and fluctuation,decision makers suffer a lot.An over-conservative decision always leads to an unnecessary deterioration of profit expectation,while an aggressive decision may result in severe constraint violations.Process optimization under uncertainty,which helps us to seek a satisfactory compromise between reliability and profitability,has received extensive attention for decades.

Three of the most popular optimization methodologies which incorporate uncertainty into the mathematical models are stochastic programming,robust optimization and chance constrained programming.Stochastic programming can be regarded as a scenario-based mathematical model,which allows recourse to hedge against in feasibility.Robust optimization is a set-induced approach that we do not need to know accurate probability distribution information about uncertainties.Its main feature is to ensure absolute feasibility over a specified uncertainty set.Chance constrained programming is a generation of robust optimization.It allows constraint violation to a certain degree.Constraints are guaranteed to be satisfied with at least a specified probability at the optimum found.

Each of them has its own advantages,limitations,applicability and main mathematical formulation,which are listed in Table 1.

Great progress has been made in this filed,however,its application in Process Systems Engineering still faces many barriers.The key difficulties in process optimization under uncertainty lie in the over conservativeness of robust optimization,the computational intractability of stochastic programming and chance constrained programming.

The scope of this paper is to provide an overview of recent advances in process optimization under uncertainty over the past ten years.It is worth noting that optimization under uncertainty has a long history of development in Process Systems Engineering and more details and information can be found in some previous excellent reviews[1–5].The content of this paper is organized as follows.Following these introductory remarks,Sections 2,3,4 present the characteristics,developments and applications of robust optimization,stochastic programming,chance constrained programming respectively.Section 5 describes the recent trends of this field and indicates the future directions.Finally,remaining challenges are discussed and conclusions are drawn in Section 6.

2.Robust Optimization

Robust optimization has emerged as a powerful technique to address uncertainty owing to its high reliability to immune against uncertainty and its computational tractability.It ensures feasibility against all possible realizations of uncertain parameters among a predefined uncertainty set.However,traditional robust optimization has alwaysbeen criticized by its over-conservativeness since the worst scenario is considered.In fact,the worst case is a rare occurrence,which may lead to an extremely poor economic objective.Therefore,many attempts have been made to reduce the conservativeness of robust optimization by avoiding the worst-case analysis.Here it should be noticed that flexibility analysis is also closely related to the solution of process optimization under uncertainty.Flexibility is concerned with the problem of ensuring feasible steady-state operation over a given range of uncertain operating conditions.Although flexibility analysis and robust optimization are derived from different fields of research,they share some knowledge in common.Both of them use the polyhedral sets to describe uncertainty and ensure feasibility by considering the worst case.Zhang et al.[6]made a comprehensive comparison between flexibility analysis and robust optimization and revealed the strong connection between them.In this paper,the discussion is focused on robust optimization rather than flexibility analysis.

Table 1 Comparison among three optimization methods

Recent advances and methods to reduce the conservatism of robust optimization can be summarized as follows.

2.1.Utilize the probability information to adjust the uncertainty set

The size of the uncertainty set reflects the decision maker's risk tolerance,while the shape of it indicates detailed knowledge the modeler may have toward uncertainties such as correlations among uncertain parameters.The robustness and conservatism of robust optimization models are severely affected by the size of uncertainty set.Since a certain degree of constraint violation is allowed in reality,the uncertainty set is not necessary to cover the whole region of the uncertainty space.To avoid overly conservative solutions,the uncertainty set should be as small as possible on condition that the probability requirement on constraint satisfaction is satisfied.Li et al.[7]presented a systematic research on the set-induced robust counterpart optimization based on different types of uncertainty sets(i.e.,interval;ellipsoidal;polyhedral;the combination of interval,ellipsoidal,and polyhedral set)and compared their conservatism.Li et al.[8,9]and Guzman et al.[10–12]demonstrated that the performance of robust optimization can be greatly improved as the probabilistic bounds that determine probabilistic guarantees on constraint violation are tightened by utilizing the probability distribution information.In the real-world optimization problems,some uncertainties have greater effects on the objective than others,so Zhang et al.[13]proposed a flexible uncertainty set where uncertainties are considered at different weights to get tighter bounds on every dimension.Yuan et al.[14]also illustrated that incorporating more accurate correlation information through the covariance matrix into robust optimization framework will deform the uncertainty set and reduce its conservativeness.

2.2.Introduce a budget parameter to adjust the degree of conservatism

Bertsimas and Sim[15,16]introduced a budge parameter Γl(not necessarily integer,that takes the value ranging from 0 to the number of uncertain parameters)to control the degree of conservatism.The corresponding formulation takes the following form.

It is assumed that some of the parameters almare subjected to uncertainty and fluctuate within the range of[alm−^alm,alm+],where almrepresents the nominal value of the parameters and^almdenotes the perturbation amplitude.Slis the subset that containsuncertain parameters,and if Γlis not an integer,tlis regarded as an index to describe an additional uncertain parameter.The intention of this formulation is to restrict up toparameters that are permitted to reach their worst case simultaneously.Hence,by adjusting Γl,we can control the level of conservatism of the solution.Besides,we have the flexibility to assign appropriate budget parameters to the different constraints according to a priori which parameters are allowed to take the worst-case value.Not only does it avoid intractable nonlinear formulation,but it also provides a trade-off between performance and conservatism.

2.3.Incorporate recourse decisions——adaptive robust optimization

The traditional robust optimization(also called static robust optimization,SRO),which requires all decisions to be made once and for all without the consideration of recourse,may lead to an overly conservative solution.Borrowing from the philosophy of multi-stage stochastic programming,adaptive robust optimization is proposed and has received much attention in recent years.Recourse decisions are introduced into static robust optimization so that adjustments can be implemented after the uncertainties are gradually revealed.

General formulation of two-stage adaptive robust optimization is presented in Eq.(2)[17],and it can be easily extended to multi-stage adaptive robust optimization framework.

Two-stage adaptive robust optimization is a sequential decision making process,which comprises a design stage and an operational stage.x represents the first-stage decision variables that have to be made before the uncertainty u reveals,while y represents the second stage(recourse)decisions which can be postponed in a “wait and see”mode based on the first-stage decisions and uncertainty realization.The set U represents the uncertainty set.The matrix W is generally called recourse matrix and T stands for technology matrix.

“You poor child,” said the prince and princess; then they praised the crows, and said they were not angry for what they had done, but that it must not happen again, and this time they should be rewarded.

To retain computational tractability,the adjustable variables are generally restricted to be affine functions of the uncertain parameters,which is shown in Eq.(3).It means the recourse variables can tune themselves to the disclosure of uncertainties.

By using linear decision rules,adaptive robust optimization not only remains a tractable LP structure,but also accounts for recourse to a certain extent.

Recent advances of robust optimization applications in Process Systems Engineering are shown in Table 2.

3.Stochastic Programming

Stochastic programming is the scenario-based model with fixed recourse that optimizes the expectation of a number of discrete scenarios using either discrete probability distributions or the discretization of continuous probability distributions.There are two types of decision variables in stochastic programming.“Here and now”decisions are those that have to be fixed for the entire horizon before uncertain parameters are revealed.“Wait and see”decisions are recognized as corrective measures that can be made at different stages based on the realization of uncertainties in sequence[2].

Eq.(4)presents the general formulation of two-stage stochastic programming[34]and it can be extended to multi-stage stochastic programming.

where x denotes the first-stage decision variables that need to be decided prior to the disclosure of uncertain variables ω,and y is regarded as the second-stage(recourse)decisions which can be postponed until the actual value of ξ becomes known.Eξrepresents the mathematical expectation with respect to ξ.T is called the technology matrix while W constitutes the recourse matrix.

The deterministic equivalent program is presented as follows.

Uncertainties in the stochastic programming can be classified into two types.Exogenous uncertainties generally correspond to market uncertainties,while endogenous uncertainties refer specifically to technical uncertainties.The realization of exogenous uncertainties is independent of process decisions.Since exogenous uncertainties are revealed automatically in each time period(e.g.oil prices),the structure of the exogenous scenario tree is known in advance.On the contrary,process decisions have important influence on the realization of endogenous uncertainties.Given that the timing of realizations relies heavily on the process decisions,the endogenous scenario-tree representation is more complicated and conditional,thus a superstructure is used to describe all potential outcomes.When copying the exogenous scenario tree for each possible combination of realizations of the endogenous uncertainties,we can get a ‘composite’scenario tree[35].Taking account of both the exogenous and endogenous uncertainties simultaneously is still an issue with great challenge.

Stochastic programming has played an important role in long-term production planning and strategic design,such as supply chain design and planning[36–42],process networks synthesis and planning[43–45],conceptual design and optimization of chemical processes[46–48],oil- field and gas- field development planning[49–51]and capacity planning for power-intensive continuous processes[52,53].Stochastic programming is generally assumed to be a risk-neutral,however,modelers have different levels of risk aversion.Therefore,incorporating risk management into stochastic programming will help the decision makers find the most satisfactory solution based on their risk tolerance and preference[40,42].

Table 2 Recent advances of robust optimization applications in Process Systems Engineering

A challenging aspect of applying stochastic programming to practical use is that the scale of the problem grows exponentially as the number of scenarios increases.Scenario generation is the basis and prerequisite of scenario-based stochastic programming,which can be done by sampling based method,statistic moment matching method and stochastic process simulation method.It is of vital importance to select an acceptable number of scenarios to well represent an approximate discretized representation of the uncertainty.Li and Floudas[54,55]proposed an optimal scenario reduction framework based on distance of uncertainty distribution and output performance,which can select a small number of representative scenarios capable of characterizing the probability density function.Calfa et al.[56]incorporated historical data to generate scenario tree by using moment matching along with distribution matching.The aforementioned methods provide a promising way to reduce the number of scenarios and computational complexity while ensuring a good-quality solution of stochastic programming.

To tackle the computational challenge of stochastic programming,integrating tailored solution strategies and decomposition algorithms is an efficient and widely used technique.Decomposition schemes take advantage of the special decomposable structure of stochastic programming(e.g.,a block angular structure),largely reducing the scale of the problem and computational expenses.It deals with the problems in which only a relatively small set of complicating variables and constraints connects the scenarios while most of the variables and constraints have no direct interaction with each other.If these complicating variables and constraints are removed by dualizing them,then the full model will be decomposed into many smaller scenario subproblems that can be solved independently.L-shaped method and Lagrangean decomposition are two of the most common decomposition algorithms to cope with large-scale stochastic programming problems and there have been many novel variants of them to improve the efficiency of the algorithms in recent literatures[35,39,42,50,53,57].

Approximating the nonlinear term in the objective is the basic idea of the L-shaped method.It is desirable to avoid numerous function evaluations for the nonlinear objective term(the recourse function)because it involves a solution of all second-stage recourse linear programs.The first-stage variables are generally selected as complicating variables in this method.L-shaped method,which applies Benders decomposition to two-stage linear stochastic programming problems with continuous recourse,is guaranteed to converge to the optimal solution of the original problem in a finite number of iterations.

Lagrangean decomposition tends to be more appropriate for two-stage and multistage stochastic integer programming problems with mixed integer variables in all stages.The scenario tree is split into independent scenarios through the relaxation of nonanticipativity constraints(NACs).The barrier of this method is that it is not always guaranteed to find feasible solutions to the original problem because the solution may violate the relaxed constraints.To eliminate the duality gap between upper bound and lower bound,R.M.Apap et al.[35]combined Lagrangean decomposition and a heuristic which helps to find feasible solutions in the branch-andbound algorithm.

The aforementioned methods are based on the discrete probability distributions of the uncertain parameters.As for the stochastic problems with continuous distributions of the uncertainties,discretizing the probability distributions through approximation or Monte Carlo sampling is the first step we should do.

4.Chance Constrained Optimization

Chance constrained optimization, first proposed by Charnes and Cooper[58,59],is a competitive tool to deal with parameter uncertainty.It allows constraint violation to some extent and does not exert recourse or penalty on the objective function,but the probability of complying with constraints must be satisfied with a specified confidence level.The probability distribution functions of the uncertainties need to be known so that the chance constrained optimization problem can be transformed into an equivalent deterministic formulation.Its main feature is that the decision makers can achieve a satisfactory trade-off between profitability and reliability by adjusting the confidence level.

A typical chance constrained optimization formulation is presented as follows.Eq.(6)represents the individual chance constrained optimization while Eq.(7)is the joint chance constrained optimization.

where f represents the objective function,the vector h denotes the inequality constraints.x,u,ξare the vectors of state,control and uncertain variables,respectively.Pr means probability andαis the confidence level specified by the modeler according to profit profiles vs.confidence level.

In the individual chance constraint,different confidence levels are assigned to different constraints based on their respective requirement of reliability.It is generally used when some constraints are more critical than others and should be paid more attention.Joint chance constraint requires all constraints to be satisfied simultaneously with the same predefined probability.When considering the safety and reliability of the process operation,joint chance constraint will be a better choice[60].When relaxing them into equivalent deterministic problems,the individual chance constraint is reformulated to a linear programming(LP)problem that can be solved with commercial LP solvers.Unfortunately,the relaxation of the joint chance constraint results in a complicated NLP problem.Especially when the uncertain parameters are correlated with each other,it will be converted to a multivariate integration formulation that is generally solved through approximation approach.Undoubtedly,the computational intractability limits its widespread applications in practice.

Major applications of chance constrained programming in recent years are shown as follows.Mitra et al.[61]used the chance constrained programming to address the multi-objective mid-term supply chain planning problem.Barz et al.[62]proposed a two-layer chance constrained optimization framework to get the optimal and reliable decisions,which is experimentally verified by a high-pressure distillation pilot plant.Yang et al.[63]employed chance constrained programming to achieve the trade-off between economic interest and reliability of decision results for a refinery supply chain operation under yield uncertainty.

To overcome the bottleneck in computational intractability,many efforts have been made in recent years.Klöppel et al.[64]used the sparse-grid integration method to solve nonlinear dynamic chance constrained optimization,which largely decreased the computational expense.Yuan et al.[65]and Li et al.[66]investigated the computational tractable robust optimization approximation framework for solving the joint chance constrained problem.And data-driven chance constrained programming used kernel-based density estimation to approximate unknown true probability density functions,reformulating chance constraints into algebraic forms[67–69].

5.Future Directions

5.1.Stochastic robust optimization

There is a tendency that a few recent studies have begun to integrate different optimization methods under uncertainties to leverage their respective superiority and compensate for their drawbacks.When coping with multi-scale uncertainties,stochastic robust optimization can achieve a low expected of cost while ensuring the system robustness by taking advantage of both of them[70].Since decision makers have different levels of conservativeness toward strategic and operational uncertainties,stochastic programming and robust optimization are integrated into a nested framework where stochastic programming is used to handle strategic uncertainties while robust optimization guarantees the robustness of operations[71].

Stochastic robust optimization model is presented as follows.

where x,y represent the binary variables and the continuous variables for strategic decisions respectively,and z denotes the continuous operational decisions.ω means a certain realization of strategic uncertainties,and uωdenotes a certain reveal of operational uncertainties.It should be noted that the operational uncertainty set Uωis indexed by the realization of the strategic uncertainty ω.Cstr(x,y)stands for strategic stage cost andrepresents the worst-case operational cost under the realization of strategic uncertainty ω.

Liu et al.[72]proposed a stochastic robust model to address two stage power system optimization problems with uncertainty,and it has a stochastic min-max structure where the inner problem of the second stage reaches the worst cases and the outer problem of the second stage optimizes the expected economic performance under various worst-case scenarios.Yue and You[71]presented a nested stochastic robust optimization model for supply chain design and operation.Amaran et al.[73]also applied it to medium-term maintenance turnaround planning for integrated chemical sites to reflect different degrees of robustness toward production levels and maintenance manpower allocation.

5.2.Data-driven optimization under uncertainty

Big data has been an active area of research in recent years,which injects fresh vitality into traditional mathematical programming with consideration of uncertainties.The past ten years witnessed an explosion in the availability of data for Process Systems Engineering applications.Massive amounts of data are collected in chemical industries(e.g.,historical demand profiles and historical yield variations).Data-driven optimization are motivated by dramatic advances of machine-learning algorithms and increasing availability of data.

Bertsimas et al.[74]utilized data to devise uncertainty sets for robust optimization through statistical hypothesis tests.Campbell et al.[75]presented a Bayesian nonparametric,data-driven,non convex uncertainty set constructed from a union of simple ellipsoidal sets for robust optimization,which can capture the nature of uncertainties in practical applications and achieve the desired degree of conservatism.Since the outliers caused by anomalous measurements significantly affect the area of the uncertainty set,robust optimization with conventional uncertainty sets could be overly conservative.Ning et al.[33]proposed a data-driven adaptive nested robust optimization framework which has two layers of robustness.A well-designed data-driven uncertainty set is immune against anomalous measurements,and the adaptive robust optimization model is robust to parameter variations.

Moreover,historical data also leverage great power in stochastic programming.The data-driven approach helps to generate the scenario tree through statistical methods and reduce its size while retaining its statistical properties.Calfa et al.[56]proposed a data-driven multistage scenario tree generation approach via moment matching and distribution matching,which can assign reasonable probabilities and outcomes in scenario trees by taking advantage of historical data.Feng et al.[76]proposed a novel heuristic scenario reduction method based on wait-and-see clustering and selected one representative scenario from each cluster.And Xu et al.[77]presented a multi-stage scenario tree generation approach by utilizing simulation,clustering,nonlinear time series and moment matching methods simultaneously.

Another challenge for process optimization under uncertainty lies in estimating probability density/distribution of uncertain parameters.Jiang et al.[68]proposed a data-driven chance-constrained approach and constructed a density-based confidence set for the uncertainty based on a general φ-divergence measure.Several data-driven nonparametric estimation techniques,such as kernel density estimation(KDE),can be used to estimate unknown probability density/distribution functions of uncertainties so that we can reformulate individual and joint chance constraints into algebraic forms[67].

In short,machine learning sits at the intersection of computer science,mathematical statistics,inference and decision-making under uncertainty [78]. Data-driven optimization, which integrates mathematical programming methods with machine learning algorithms,has become an emerging and competitive tool to cope with optimization problems under uncertainty utilizing massive historical data.

6.Conclusions and Outlook

An overview presented in the preceding sections demonstrates that dramatic progress has been made in the field of process optimization under uncertainties over the past ten years.Different optimization methods focus on specific application areas.Robust optimization is more suitable for short-term scheduling problems.As for long-term production planning and strategic design decisions,such as the development of oil fields,the capacity expansion of process networks and supply chain design and planning,stochastic programming leverages its superiority because it allows recourse decisions to be made at later time points after the gradual realization of the uncertainties.The above detailed discussion indicates that the major challenge of robust optimization is how to reduce the degree of conservatism and to seek a trade-off between robustness and profitability.Several methods,such as adjusting the size and the shape of the uncertainty set by utilizing the probability information,introducing a budget parameter and incorporating recourse decisions,provide greater flexibility for robust optimization and make the solutions of robust optimization less conservative.The key limitation of stochastic programming and chance constrained programming lies in their computational intractability.To solve large-scale stochastic programming problems efficiently,scenario reduction strategies and decomposition algorithms are widely used.Parallel computation and Multi-Agent System will also make a contribution to the large-scale mathematical models in the future.As for chance constrained programming,many scholars have studied the application of the robust counterpart optimization to approximate the joint chance constrained problem in recent years and indicate that robust optimization can provide safe and tractable analytical approximation for solving joint chance constrained problem.

In the era of big data and with the rapid development of advanced machine-learning algorithms,process optimization under uncertainty benefits from valuable information embedded within massive historical data.It is inspiring to see that the machine-learning model is seamlessly integrated with optimization approaches.Since the well-designed data-driven uncertainty set tends to be tighter than classical ones,it can greatly reduce the conservatism of robust optimization solutions.Also,the obtained data-driven probabilities and outcomes can be used in the scenario tree generation for stochastic programming.Undoubtedly, data-driven machine-learning approaches can help us to capture the characterization of uncertain data and advance the development of process optimization under uncertainty.Despite the recent progress in the data-driven optimization,there are still many problems and challenges left.Not all the uncertainties can be predicted and described based on the historical data(e.g.,the oil price suffers dramatic fluctuations due to some sudden political and economic factors,so the historical data do not have the reference value).In addition,the data obtained directly from the industries are not“clean”enough and have to be preprocessed before we use it.

Current trend in this field also demonstrates that different optimization methods have begun to integrate with each other.Since a“one size- fits-all”approach is not sufficient to reflect the modeler's different levels of conservatism toward multi-scale uncertainties,stochastic robust optimization is proposed.Furthermore,in the presence of uncertainties,decision makers not only pursue the maximization of profits,but they also care about the minimization of risks.Hence,incorporating effective risk management measures into optimization models is another promising direction in the future.