Researchers and industrial practitioners are now interested in combining machine learning (ML) and operations research and management science to develop prescriptive analytics frameworks. By and large, a single value or a discrete distribution with a finite number of scenarios is predicted using an ML model with an unknown parameter; the value or distribution is then fed into an optimization model with the unknown parameter to prescribe an optimal decision. In this paper, we prove a deficiency of prescriptive analytics, i.e., that no perfect predicted value or perfect predicted distribution exists in some cases. To illustrate this phenomenon, we consider three different frameworks of prescriptive analytics, namely, the predict-then-optimize framework, smart predict-then-optimize framework and weighted sample average approximation (w-SAA) framework. For these three frameworks, we use examples to show that prescriptive analytics may not be able to prescribe a full-information optimal decision, i.e., the optimal decision under the assumption that the distribution of the unknown parameter is given. Based on this finding, for practical prescriptive analytics problems, we suggest comparing the prescribed results among different frameworks to determine the most appropriate one.
Citation: Shuaian Wang, Xuecheng Tian, Ran Yan, Yannick Liu. A deficiency of prescriptive analytics—No perfect predicted value or predicted distribution exists[J]. Electronic Research Archive, 2022, 30(10): 3586-3594. doi: 10.3934/era.2022183
[1] | Xuecheng Tian, Ran Yan, Shuaian Wang, Yannick Liu, Lu Zhen . Tutorial on prescriptive analytics for logistics: What to predict and how to predict. Electronic Research Archive, 2023, 31(4): 2265-2285. doi: 10.3934/era.2023116 |
[2] | Ran Yan, Ying Yang, Yuquan Du . Stochastic optimization model for ship inspection planning under uncertainty in maritime transportation. Electronic Research Archive, 2023, 31(1): 103-122. doi: 10.3934/era.2023006 |
[3] | Sanqiang Yang, Zhenyu Yang, Leifeng Zhang, Yapeng Guo, Ju Wang, Jingyong Huang . Research on deformation prediction of deep foundation pit excavation based on GWO-ELM model. Electronic Research Archive, 2023, 31(9): 5685-5700. doi: 10.3934/era.2023288 |
[4] | Ju Wang, Leifeng Zhang, Sanqiang Yang, Shaoning Lian, Peng Wang, Lei Yu, Zhenyu Yang . Optimized LSTM based on improved whale algorithm for surface subsidence deformation prediction. Electronic Research Archive, 2023, 31(6): 3435-3452. doi: 10.3934/era.2023174 |
[5] | Shengming Hu, Yongfei Lu, Xuanchi Liu, Cheng Huang, Zhou Wang, Lei Huang, Weihang Zhang, Xiaoyang Li . Stability prediction of circular sliding failure soil slopes based on a genetic algorithm optimization of random forest algorithm. Electronic Research Archive, 2024, 32(11): 6120-6139. doi: 10.3934/era.2024284 |
[6] | Dewang Li, Meilan Qiu, Jianming Jiang, Shuiping Yang . The application of an optimized fractional order accumulated grey model with variable parameters in the total energy consumption of Jiangsu Province and the consumption level of Chinese residents. Electronic Research Archive, 2022, 30(3): 798-812. doi: 10.3934/era.2022042 |
[7] | Zhizhou Zhang, Zhenglei Wei, Bowen Nie, Yang Li . Discontinuous maneuver trajectory prediction based on HOA-GRU method for the UAVs. Electronic Research Archive, 2022, 30(8): 3111-3129. doi: 10.3934/era.2022158 |
[8] | Ilyоs Abdullaev, Natalia Prodanova, Mohammed Altaf Ahmed, E. Laxmi Lydia, Bhanu Shrestha, Gyanendra Prasad Joshi, Woong Cho . Leveraging metaheuristics with artificial intelligence for customer churn prediction in telecom industries. Electronic Research Archive, 2023, 31(8): 4443-4458. doi: 10.3934/era.2023227 |
[9] | Xinyi Xu, Shaojuan Ma, Cheng Huang . Uncertainty prediction of wind speed based on improved multi-strategy hybrid models. Electronic Research Archive, 2025, 33(1): 294-326. doi: 10.3934/era.2025016 |
[10] | Yiming Zhang, Zhiwei Pan, Shuyou Zhang, Na Qiu . Probabilistic invertible neural network for inverse design space exploration and reasoning. Electronic Research Archive, 2023, 31(2): 860-881. doi: 10.3934/era.2023043 |
Researchers and industrial practitioners are now interested in combining machine learning (ML) and operations research and management science to develop prescriptive analytics frameworks. By and large, a single value or a discrete distribution with a finite number of scenarios is predicted using an ML model with an unknown parameter; the value or distribution is then fed into an optimization model with the unknown parameter to prescribe an optimal decision. In this paper, we prove a deficiency of prescriptive analytics, i.e., that no perfect predicted value or perfect predicted distribution exists in some cases. To illustrate this phenomenon, we consider three different frameworks of prescriptive analytics, namely, the predict-then-optimize framework, smart predict-then-optimize framework and weighted sample average approximation (w-SAA) framework. For these three frameworks, we use examples to show that prescriptive analytics may not be able to prescribe a full-information optimal decision, i.e., the optimal decision under the assumption that the distribution of the unknown parameter is given. Based on this finding, for practical prescriptive analytics problems, we suggest comparing the prescribed results among different frameworks to determine the most appropriate one.
With the advancement of ML technologies and the accessibility of rich data, recent studies in data-driven optimization have illustrated the advantages of using rich feature data to reduce uncertainty in decision-making problems with uncertain parameters and thus improve the decision making performance [1,2]. Following the formulation of Bertsimas and Kallus [1], consider an optimization problem with a given cost function c(y;z), where y∈Y⊂Rdy denotes the uncertain parameter vector affecting the value of the cost function and z∈Z⊂Rdz denotes the decision vector. Assume that there is an observation of auxiliary data x0∈X⊂Rdx, which is associated with y and will be used to predict y. The optimization problem can be mathematically formulated as follows:
z∗(x0)∈argminz∈ZEy[c(y;z)∣x=x0]. | (1) |
To solve Optimization problem (1), a historical dataset denoted as {(xi,yi)}ni=1 is made accessible, where xi∈X is the historical auxiliary data vector, and yi∈Y is the corresponding historical realization of y.
To solve Optimization problem (1), the traditional method is to first use {(xi,yi)}ni=1 to build an ML model and then use the ML model to predict the value of y from the new example x0 [1,3]. That is, for every x, the ML model generates a predicted y, as denoted by ˆy(x). A commonly used loss function to train the ML model with a continuous prediction target is the mean squared error (MSE) loss, which is expressed as follows:
LMSE=1nn∑i=1(yi−ˆy(xi))2. | (2) |
After obtaining the predicted ˆy(x0), we then solve
minz∈Zc(ˆy(x0);z) | (3) |
to prescribe an optimal decision. This method is generally termed a predict-then-optimize (PO) framework or a two-stage framework [3].
Nonetheless, a good prediction may not lead to a good decision [1]. This is because Eq (2) does not consider the impact of ˆy(x) on the downstream optimization problem. As a result, a more natural and appropriate method is to plug the optimization problem into the training of the ML model. That is, instead of focusing on minimizing the prediction error, we train the ML model by minimizing the decision error. This method is generally termed a smart PO (SPO) framework, an end-to-end learning framework, or a decision-focused learning framework [3,4]. A commonly used loss function, for ML models under these frameworks, namely, SPO loss, can be formulated as follows:
LSPO=1nn∑i=1[c(yi;z(ˆyi))−c(yi;z(yi))], | (4) |
where the first term in square brackets represents the cost derived from the decision z(ˆyi)∈argminz∈Zc(ˆyi;z) and the second term in square brackets denotes the cost of the full-information optimal decision z(yi)∈argminz∈Zc(yi;z). However, training ML models using SPO loss might be impossible computationally because of the nonconvex and discontinuous characteristics of the SPO loss function for combinatorial optimization problems.
In a recent study, Bertsimas and Kallus [1] considered another approach to take the uncertainty of y into account when solving Eq (1). In their method, they claimed that, for many traditional ML methods, predicting y from a new example x0 can take the following form:
ˆy(x0)=n∑i=1w(xi,x0)yi, | (5) |
where w(xi,x0) measures the similarity (closeness) between Example xi in the historical data and the new example x0; and, its format depends on the ML model used (i.e., random forest, k-nearest neighbor (kNN)) [5]. For example, if we use a kNN model, w(xi,x0)=1/k if xi is a kNN of x0, and w(xi,x0)=0 otherwise. This is because w(xi,x0) can be regarded as an approximation of the conditional distribution of y given x=x0, i.e., the approximate distribution of y has n scenarios y1,…,yn with probabilities w(x1,x0),…(xn,x0) (and it is very likely that some probabilities are 0, meaning that the approximate distribution of y has less than n scenarios). Bertsimas and Kallus [1] then used
minzn∑i=1w(xi,x0)c(yi;z) | (6) |
as an approximation of Eq (1) [5]. This method has been formally termed as a weighted sample average approximation (w-SAA) framework by Notz and Pibernik [6] because it combines local predictive ML methods (i.e., kNN) and traditional techniques for data-driven optimization (i.e., SAA).
For these prescriptive analytics frameworks, the main aim is to predict a perfect value of y (i.e., using a PO framework and an SPO framework) or approximate a perfect conditional distribution of y (i.e., using a w-SAA framework) to minimize the uncertain cost c(y;z) after observing x=x0. In other words, we hope to prescribe a perfect decision that, ideally, ought to be the same as the full-information optimal decision z∗(x0) if we can predict a perfect ˆy(x0) or a perfect conditional distribution of y given x0. However, does the perfect predicted value or the perfect predicted distribution really exist? In this paper, we aim to answer this question by proving a surprising fact:
No perfect predicted value exists in PO and SPO frameworks, and no perfect predicted distribution exists in a w-SAA framework.
To illustrate this phenomenon, we first design two examples under the PO framework and the SPO framework in Section 2. We then design two examples under the w-SAA framework in Section 3 and show that the w-SAA framework may perform better than the PO/SPO framework. At last, Section 4 concludes this paper.
For most uncertain optimization problems with auxiliary data, a commonly used PO framework entails first predicting the value of uncertain parameters and then plugging predictions into the optimization problem to derive optimal decisions. As this method does not consider the impact of predictions on the downstream decisions, it may lead to sub-optimal decisions. Hence, Elmachtoub and Grigas [3] proposed an SPO framework for a broad class of decision-making problems with uncertainty to deal with this issue. Either under the PO framework or the SPO framework, the aim is to predict a perfect uncertain parameter with the use of auxiliary data, so as to prescribe a full-information optimal decision if the perfect full information (i.e., the marginal distribution) of the uncertain parameter is known. However, it is possible that no perfect predicted value exists in PO and SPO frameworks, which shall be illustrated using two examples. We first show that this deficiency exists in a classification task under the PO/SPO framework.
Example 1. Consider that we have an odd number n of historical records (examples) in the auxiliary dataset {(xi,yi)}ni=1. Suppose that the features of all examples are identical (i.e., x1=…=xn). For example, xi indicates whether it is a work day, where xi=1 indicates that it is a work day and xi=0 otherwise, thus, xi∈{0,1}. Additionally, suppose that the uncertain parameter y∈{0,1} follows a discrete uniform distribution independent of x and that the optimization problem with the cost function c(y;z) is as follows:
min−1≤z1,z2≤1[(y−0.7)z1+(y−0.3)z2], | (7) |
where z=(z1,z2) is the decision vector.
Suppose x0=x1=…=xn. After observing x0, the resulting decision problem is
min−1≤z1,z2≤1Ey[(y−0.7)z1+(y−0.3)z2∣x=x0]. | (8) |
If we know in advance that the uncertain parameter y follows a discrete uniform distribution with two values 0 and 1 independent of x, the above model can be written as:
min−1≤z1,z2≤1{0.5[(1−0.7)z1+(1−0.3)z2]+0.5[(0−0.7)z1+(0−0.3)z2]}. | (9) |
Therefore, the unique full-information optimal decision for Optimization problem (8) is z∗(x0)=(z∗1(x0),z∗2(x0))=(1,−1).
Suppose we use a classification ML model to predict the value of y (0 or 1) given the new example x0. Since we have n records satisfying x0=x1=…=xn and n is an odd number, if n∑i=1I(yi=1)>n∑i=1I(yi=0) (where I(q) is an indicator function which takes the value of 1 if Condition q is true and 0 if Condition q is false), we can obtain ˆy=1 for the new example x0 by using a classification ML model. This is because predicting y to be 1 for all examples in the historical dataset can minimize Loss function (2) under the PO framework and minimize Loss function (4) under the SPO framework. Given that ˆy=1, Optimization model (7) can be written as:
min−1≤z1,z2≤1[(1−0.7)z1+(1−0.3)z2], | (10) |
and the prescribed optimal decision is z∗(ˆy)=(z∗1(ˆy),z∗2(ˆy))=(−1,−1). Similarly, if n∑i=1I(yi=0)>n∑i=1I(yi=1), we can obtain ˆy=0 for the new example x0. Given that ˆy=0, Optimization model (7) can be written as
min−1≤z1,z2≤1[(0−0.7)z1+(0−0.3)z2], | (11) |
and the prescribed optimal decision is z∗(ˆy)=(z∗1(ˆy),z∗2(ˆy))=(1,1). So, for this optimization problem, z∗(x0)≠z∗(ˆy), which means that we can never predict a perfect ˆy to prescribe the full-information optimal decision z∗(x0)=(1,−1).
From Example 1, we find that under the PO/SPO framework, we may not be able to predict a perfect value to prescribe a full-information optimal decision with a classification ML model. This phenomenon can also be seen in a regression task, as illustrated in the following example.
Example 2. Suppose that we have an odd number n of historical records (examples) in the auxiliary dataset {(xi,yi)}ni=1. Suppose that the features of all examples are identical (i.e., x1=…=xn). For example, xi indicates whether it is a work day and xi∈{0,1}. Additionally, suppose that the uncertain parameter y is a real number following a uniform distribution independent of x with the following probability density function:
f(y)=12b,−b<y<b, | (12) |
where b is a positive real value. The optimization problem with the cost function c(y;z) is established as follows:
min−1≤z1,z2≤1[(I(y>0)−0.7)z1+(I(y>0)−0.3)z2], | (13) |
where z=(z1,z2) is the decision vector.
Suppose that x0=x1=…=xn. After observing x0, the resulting decision problem is
min−1≤z1,z2≤1Ey[(I(y>0)−0.7)z1+(I(y>0)−0.3)z2∣x=x0]. | (14) |
If we know in advance that the uncertain parameter y follows a uniform distribution independent of x and that its probability density function is as shown in Eq (12), the above model can be written as:
min−1≤z1,z2≤10.5[(1−0.7)z1+(1−0.3)z2]+0.5[(0−0.7)z1+(0−0.3)z2], | (15) |
and the unique full-information optimal decision for Optimization problem (14) is z∗(x0)=(z∗1(x0),z∗2(x0))=(1,−1).
Suppose we use a regression ML model to predict the value of y (a real number) given the new example x0. Since we have n records satisfying x0=x1=…=xn and n is an odd number, if n∑i=1I(yi>0)>n∑i=1I(yi≤0), we can obtain ˆy>0 for the new example x0 by using a regression ML model. This is because predicting y to be bigger than 0 for all examples in the historical dataset can minimize Loss function (2) under the PO framework and minimize Loss function (4) under the SPO framework. Given that ˆy>0, Optimization model (7) can be written as:
min−1≤z1,z2≤1[(1−0.7)z1+(1−0.3)z2], | (16) |
and the prescribed optimal decision is z∗(ˆy)=(z∗1(ˆy),z∗2(ˆy))=(−1,−1). Similarly, if n∑i=1I(yi≤0)>n∑i=1I(yi>0), we can obtain ˆy≤0 for the new example x0. Given that ˆy=0, Optimization model (7) can be written as:
min−1≤z1,z2≤1[(0−0.7)z1+(0−0.3)z2], | (17) |
and the prescribed optimal decision is z∗(ˆy)=(z∗1(ˆy),z∗2(ˆy))=(1,1). So, for this optimization problem, z∗(x0)≠z∗(ˆy), which means that we can never predict a perfect ˆy to prescribe the full-information optimal decision z∗(x0)=(1,−1).
From Examples 1 and 2, we have proved that, under the PO/SPO framework for both regression and classification tasks, there may not be a perfect predicted value that will prescribe the full-information optimal solution. There are, of course, situations in which there is a perfect predicted value that will prescribe a full-information optimal solution, which is shown in Proposition 1, as follows.
Proposition 1. If c(y;z) is linear in y, then there exists a perfect predicted value that will prescribe a full-information optimal solution if y can take any value in Rdy in the predictive ML model.
Proof: If c(y;z) is linear in y, the model given by Eq (1) can be represented by minz∈ZEy[c(y;z)∣x=x0]=minz∈ZEy[y∣x=x0]z. Hence, ˆy=E[y∣x=x0] can lead to the full-information optimal decision.
If there exists a perfect predicted value that will prescribe a full-information optimal solution, the perfect predicted value is not necessarily the conditional mean value, which is shown in Proposition 2, as follows.
Proposition 2. In the traditional newsvendor problem with a continuous order quantity (denoted by Z) and demand quantity (denoted by y) [7,8],
c(y;z)=omax(z−y,0)+umax(y−z,0) | (18) |
is the random cost, where o and u are the overage cost and underage cost, respectively. If the marginal demand distribution of y (with a given x0, e.g., whether it is a workday) is known in advance and is denoted as Fx0, the full-information optimal decision is z∗(x0)=F−1x0(u/(u+o)). Hence, the predicted value ˆy=F−1x0(u/(u+o)) will prescribe a full-information optimal solution. Furthermore, in the traditional newsvendor problem with an integer demand quantity and integer order quantity, the predicted value ˆy, which is the smallest integer satisfying Fx0(ˆy)≥u/(u+o), will prescribe a full-information optimal solution.
In Section 2, we presented two examples to show the deficiency of PO and SPO frameworks, demonstrating that it might be impossible to predict a perfect value to prescribe a full-information optimal decision. A similar phenomenon also exists in the w-SAA framework. The w-SAA framework was initially proposed by Bertsimas and Kallus [1], and it is based on the idea of deriving weights from features by developing local predictive ML methods and optimizing the decision against a reweighting of the data [6]. According to the format of Eq (5), this method aims to estimate the conditional distribution of y given the new example x0. Taking the kNN method as an example, we assume that the predicted target ˆy follows a discrete uniform distribution with no more than K values (neighbors). Here, K is not an infinite positive number. However, with full information, the real y may either follow a discrete distribution with more than K values, or a continuous distribution. Then, given the above information, we use the following example to show that a deficiency of the w-SAA framework also exists, that is, no perfect predicted distribution exists under the w-SAA framework.
Example 3. Consider that we have an auxiliary dataset {(xi,yi)}ni=1. Suppose that the features of all historical data are identical (i.e., x1=…=xn). For example, xi indicates whether it is a work day and xi∈{0,1}. Additionally, suppose that the uncertain parameter y is a real number following a discrete uniform distribution independent of x and y∈{1,2,3,…,K+1}. Let 0<ε<1, which is a parameter used to formulate the following optimization problem. The optimization problem with the cost function c(y;z) is established as follows:
min−1≤zk,1,zk,2≤1∑k=1,…,K+1[(I(y=k)−(1+ε)K+1)zk,1+(I(y=k)−(1−ε)K+1)zk,2]. | (19) |
Suppose that x0=x1=…=xn. After observing x0, the resulting decision problem is
min−1≤zk,1,zk,2≤1Ey{∑k=1,…,K+1[(I(y=k)−(1+ε)K+1)zk,1+(I(y=k)−(1−ε)K+1)zk,2∣x=x0]}. | (20) |
If we know in advance that the uncertain parameter y follows a discrete uniform distribution, the above model can be written as:
min−1≤zk,1,zk,2≤1∑k=1,…,K+1{1K+1(1−(1+ε)K+1)zk,1+KK+1(0−(1+ε)K+1)zk,1+1K+1(1−(1−ε)K+1)zk,2+KK+1(0−(1−ε)K+1)zk,2}, | (21) |
and the unique full-information optimal decision for Optimization problem (20) is z∗k,1(x0)=1,∀k=1,…,K+1 and z∗k,2(x0)=−1,∀k=1,…,K+1.
Suppose we use the w-SAA framework with a kNN model to predict a conditional distribution of y with no more than K values given the new example x0. These values are denoted by y1,…,yk′,…,yK and each value has a weight w(yk′)=1/K,∀k′=1,…,K. For other data examples that do not have a target value in {y1,…,yk′,…,yK}, they will be assigned a weight of 0. In order to obtain the full-information optimal decision, the following condition must be satisfied when we solve Optimization problem (6): ∀k=1,…,K+1,1−εK+1≤∑k′∈{yk′=k}w(yk′)≤1+εK+1. However, since yk′ only has no more than K values, the above condition cannot be met because ∃k=1,…,K+1,∑k′∈{yk′=k}w(yk′)=0. That is, we are not able to prescribe a decision, which, ideally, ought to be the same as the full-information optimal decision in the w-SAA framework. Similarly, this phenomenon can be extended when the uncertain parameter y follows a continuous distribution.
Note that the prediction of a single value in the PO/SPO framework can be considered as a special case of the w-SAA framework that involves applying a single scenario to approximate the marginal distribution. Based on this insight, the following Example 4 shows that the w-SAA framework can prescribe better solutions than the PO/SPO framework.
Example 4. (Example 3 continued) Suppose we set K=2 for Example 3, that is, the uncertain parameter y is a real number following a discrete uniform distribution where y∈{1,2,3}; then, we conduct a three-class classification task. Suppose that x0=x1=…=xn. After observing x0, the full-information optimal decision for Optimization problem (19) is z∗k,1(x0)=1,∀k=1,2,3 and z∗k,2(x0)=−1,∀k=1,2,3.
Let us use the w-SAA framework with a kNN model to find the nearest two examples of the new example x0 denoted by {(xj,yj)}2j=1. These two examples are assigned a weight of wi=1/2,∀i=1,2, and the rest of the examples (xj,yj),∀j=3,…,n are assigned a weight of 0. There are six possible combinations for (y1,y2), which are (1,1), (2,2), (3,3), (1,2), (1,3) and (2,3). For the first three combinations, following the same procedures shown in Example 3, the prescribed cost will be 0. Similarly, for the last three combinations, the prescribed cost will be −(4ε/3).
If we use a PO/SPO framework to predict the value of y, there are three possible outcomes, i.e., ˆy=1, ˆy=2, and ˆy=3. No matter under which outcome, if we plug the prediction into Optimization problem (19), the prescribed cost will be 0. Since 0≤0 and −(4ε/3)<0, it shows that the w-SAA framework can prescribe a better solution.
Note further that the w-SAA framework approximates the conditional distribution of y in a local manner (i.e., by using a portion of the training data). Wang and Yan [9,10] proposed two methods to approximate the conditional distribution of y in a global manner (i.e., by using all of the training data). Similarly, we can show that there are cases for which no perfect predicted distribution exists under the conditions of the methods proposed by Wang and Yan [9,10].
From the above examples, we have shown that it is possible that prescriptive analytics cannot predict a perfect value or a perfect discrete distribution for finite scenarios. Nonetheless, different frameworks may prescribe solutions with different qualities, and our aim is to find a better framework that can prescribe a near-optimal solution.
Using abundant auxiliary data and ML models, researchers and industrial practitioners can make better decisions. However, using current prescriptive analytics frameworks, we may not be able to prescribe a full-information optimal decision. This paper thus makes the following contributions. First, this paper proves a deficiency of prescriptive analytics, i.e., that no perfect predicted value or predicted distribution exists; this was demonstrated by presenting three examples under different frameworks, namely the PO framework, SPO framework, and w-SAA framework. From these three examples, we show that prescriptive analytics may not be able to prescribe a perfect decision which ideally ought to be the same as the full-information optimal decision.
Second, this paper inspires researchers and practitioners to check the existence of the perfect predicted value or the perfect predicted distribution while using prescriptive analytics frameworks. If there is no perfect predicted value or predicted distribution for an uncertain optimization problem, we may need to try more than one framework and compare the prescribed results obtained from different frameworks. As indicated by Example 4, it is possible for us to prescribe a better solution by using a w-SAA framework rather than a PO/SPO framework. To summarize, our research demonstrates that neither PO/SPO nor w-SAA is perfect, and that, consequently, more efforts should be devoted to the field of prescriptive analytics.
There are several limitations to this research. First, this paper only includes a theoretical analysis, as the available practical data are insufficient. In the future, practical problems with real data can be used to verify the deficiency demonstrated in this paper. Second, when analyzing the superiority of different prescriptive analytics frameworks, it should be taken into consideration that this paper does not prove a theoretical gap between their decision-making qualities, which should be explored further in the future.
The authors thank the two reviewers for their valuable comments.
The authors declare that there is no conflict of interest.
[1] | D. Bertsimas, N. Kallus, From predictive to prescriptive analytics. Manage. Sci., 66 (2020), 1025-1044. https://doi.org/10.1287/mnsc.2018.3253 |
[2] | T. Olovsson, T. Svensson, J. Wu, Future connected vehicles: Communications demands, privacy and cyber-security, Commun. Transp. Res., 2 (2022), 100056. https://doi.org/10.1016/j.commtr.2022.100056 |
[3] | A. N. Elmachtoub, P. Grigas, Smart "predict, then optimize", Manage. Sci., 68 (2021), 9-26. https://doi.org/10.1287/mnsc.2020.3922 |
[4] | M. Mulamba, J. Mandi, M. Diligenti, M. Lombardi, V. Bucarey, T. Guns, Contrastive losses and solution caching for predict-then-optimize, in Proceedings of 2021 International Joint Conference on Artificial Intelligence, (2021), 2833-2840. https://arXiv.org/abs/2011.05354v2 |
[5] |
D. Bertsimas, N. Koduri, Data-driven optimization: A reproducing kernel Hilbert space approach, Oper. Res., 70 (2021), 454-471. https://doi.org/10.1287/opre.2020.2069 doi: 10.1287/opre.2020.2069
![]() |
[6] |
P. Notz, R. Pibernik, Prescriptive analytics for flexible capacity management, Manage. Sci., 68 (2022), 1756-1775. https://doi.org/10.1287/mnsc.2020.3867 doi: 10.1287/mnsc.2020.3867
![]() |
[7] |
L. Chen, D. Long, G. Perakis, The impact of a target on newsvendor decisions, Manuf. Serv. Oper. Manage., 17 (2015), 78-86. https://doi.org/10.1287/msom.2014.0500 doi: 10.1287/msom.2014.0500
![]() |
[8] |
G. Y. Ban, C. Rudin, The big data newsvendor: Practical insights from machine learning, Oper. Res., 67 (2019), 90-108. https://doi.org/10.1287/opre.2018.1757 doi: 10.1287/opre.2018.1757
![]() |
[9] | S. Wang, R. Yan, A global method from predictive to prescriptive analytics considering prediction error for "Predict, then optimize" with an example of low-carbon logistics, Cleaner Log. Supply Chain, 4 (2022), 100062. https://doi.org/10.1016/j.clscn.2022.100062 |
[10] | S. Wang, R. Yan, "Predict, then optimize" with quantile regression: A global method from predictive to prescriptive analytics and applications to transportation, Multi. Transp., (2022), in press. |
1. | Yiwei Wu, Yadan Huang, Hans Wang, Lu Zhen, Wei Shao, Green Technology Adoption and Fleet Deployment for New and Aged Ships Considering Maritime Decarbonization, 2022, 11, 2077-1312, 36, 10.3390/jmse11010036 | |
2. | Jingwen Qi, Shuaian Wang, LNG Bunkering Station Deployment Problem—A Case Study of a Chinese Container Shipping Network, 2023, 11, 2227-7390, 813, 10.3390/math11040813 | |
3. | Haoqing Wang, Wen Yi, Shuaian Wang, Facility planning and schedule design in the pandemic: Eliminating contacts at construction workplace, 2023, 395, 09596526, 136394, 10.1016/j.jclepro.2023.136394 | |
4. | Wei Wang, Yannick Liu, Lu Zhen, H. Wang, How to Deploy Electric Ships for Green Shipping, 2022, 10, 2077-1312, 1611, 10.3390/jmse10111611 | |
5. | Xuecheng Tian, Ran Yan, Shuaian Wang, Yannick Liu, Lu Zhen, Tutorial on prescriptive analytics for logistics: What to predict and how to predict, 2023, 31, 2688-1594, 2265, 10.3934/era.2023116 | |
6. | Simon Tian, Xinyi Zhu, Data analytics in transport: Does Simpson's paradox exist in rule of ship selection for port state control?, 2023, 31, 2688-1594, 251, 10.3934/era.2023013 | |
7. | Xuecheng Tian, Ran Yan, Yannick Liu, Shuaian Wang, A smart predict-then-optimize method for targeted and cost-effective maritime transportation, 2023, 172, 01912615, 32, 10.1016/j.trb.2023.03.009 | |
8. | Haoqing Wang, Ran Yan, Man Ho Au, Shuaian Wang, Yong Jimmy Jin, Federated learning for green shipping optimization and management, 2023, 56, 14740346, 101994, 10.1016/j.aei.2023.101994 | |
9. | Yiwei Wu, Yao Lu, Shuaian Wang, Lu Zhen, New challenges in fleet deployment considering EU oil sanctions, 2023, 31, 2688-1594, 4507, 10.3934/era.2023230 | |
10. | Yiwei Wu, Hongyu Zhang, Shuaian Wang, Lu Zhen, Mathematical Optimization of Carbon Storage and Transport Problem for Carbon Capture, Use, and Storage Chain, 2023, 11, 2227-7390, 2765, 10.3390/math11122765 | |
11. | Wenfeng Li, Lei Cai, Lijun He, Wenjing Guo, Scheduling techniques for addressing uncertainties in container ports: A systematic literature review, 2024, 162, 15684946, 111820, 10.1016/j.asoc.2024.111820 |