Skip to main content
Log in

Using multiattribute utility theory to avoid bad outcomes by focusing on the best systems in ranking and selection

  • Original Article
  • Published:
Journal of Simulation

Abstract

When making decisions under uncertainty, it seems natural to use constraints on performance to avoid the selection of a particularly bad system. However that intuition has been shown to impair good recommendations as demonstrated by some well-known results in the stochastic optimization literature. Our work on multiattribute ranking and selection procedures demonstrates that Pareto and constraint-based approaches could be used as part of a successful decision process; but a tradeoff-based approach, like multiattribute utility theory, is required to identify the true best system in all but a few special cases. We show that there is no guaranteed strategic equivalence between utility theory and constraint-based approaches when constraints on the means of the performance measures are used in the latter. Hence, a choice must be made as to which is appropriate. In this paper, we extend well-known results in the decision analysis literature to ranking and selection.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Figure 1
Figure 2
Figure 3

Similar content being viewed by others

References

  • Abbas AE and Matheson JE (2005). Normative target-based decision making. Managerial and Decision Economics 26 (6): 373–385.

    Article  Google Scholar 

  • Andradóttir S, Goldsman D and Kim S (2005). Procedures for feasibility detection in the presence of multiple constraints. In: Kuhl ME, Steiger NM, Armstrong FB, and Jones JA (eds). Proceedings of the 2005 Winter Simulation Conference, IEEE: Piscataway, NJ, pp 692–698.

  • Andradóttir S and Kim S (2010). Fully sequential procedures for comparing constrained systems via simulation. Naval Research Logistics 57 (3): 403–421.

    Google Scholar 

  • Batur D and Kim S (2005). Finding the best in the presence of multiple constraints. In: Kuhl ME, Steiger NM, Armstrong FB and Jones JA (eds). Proceedings of the 2005 Winter Simulation Conference, IEEE: Piscataway, NJ, pp 732–738.

  • Batur D and Kim S (2010). Finding feasible systems in the presence of constraints on multiple performance measures. ACM TOMACS 20 (3)13:1–13:26.

  • Bechhofer RE, Santner TJ and Goldsman DM (1995). Design and Analysis of Experiments for Statistical Selection, Screening, and Multiple Comparisons. J. Wiley & Sons: New York.

    Google Scholar 

  • Blau RA (1974). Stochastic programming and decision analysis: An apparent dilemma. Management Science 21 (2): 271–276.

    Article  Google Scholar 

  • Bordley RF and Kirkwood CW (2004). Multiattribute preference analysis with performance targets. Operations Research 52 (6): 823–835.

    Article  Google Scholar 

  • Bordley RF and LiCalzi M (2000). Decision analysis using targets instead of utility functions. Decisions in Economics and Finance 23 (1): 53–74.

    Article  Google Scholar 

  • Bordley RF and Pollock SM (2009). A decision-analytic approach to reliability-based design optimization. Operations Research 57 (5): 1262–1270.

    Article  Google Scholar 

  • Butler JC, Jia J and Dyer JS (1997). Simulation techniques for the sensitivity analysis of multi-criteria decision models. European Journal of Operational Research 103 (3): 531–546.

    Article  Google Scholar 

  • Butler JC, Merrick JRW and Morrice DJ (2011). Assessing oil spill risk in port tanker operations using a multiattribute utility approach to ranking and selection. In: Jain S, Creasey RR, Himmelspach J, White KP and Fu M (eds). Proceedings of the 2011 Winter Simulation Conference, IEEE: Piscataway, NJ, pp 1696–1707.

  • Butler JC, Morrice DJ and Mullarkey P (2001). A multiple attribute utility theory approach to ranking and selection. Management Science 47 (6): 800–816.

    Article  Google Scholar 

  • Charnes A and Cooper WW (1963). Deterministic equivalents for optimizing and satisficing under chance constraints. Operations Research 11 (1): 18–39.

    Article  Google Scholar 

  • Charnes A and Cooper WW (1975). A comment on Blau’s dilemma in stochastic programming and Bayesian decision analysis. Management Science 22 (4): 498–500.

    Article  Google Scholar 

  • Chick SE and Frazier PI (2012). Sequential sampling with economics of selection procedures. Management Science 58 (3): 550–569.

    Article  Google Scholar 

  • Chick SE and Gans N (2009). Economic analysis of simulation selection problems. Management Science 55 (3): 421–437.

    Article  Google Scholar 

  • Chick SE and Inoue K (2001). New two-stage and sequential procedures for selecting the best simulated system. Operations Research 49 (5): 732–743.

    Article  Google Scholar 

  • Frazier PI, Powell WB and Dayanik S (2008). A knowledge gradient policy for sequential information collection. SIAM Journal on Control and Optimization 47 (5): 2410–2439.

    Article  Google Scholar 

  • Frazier PI, Powell WB and Dayanik S (2009). The knowledge-gradient policy for correlated normal rewards. INFORMS Journal on Computing 21 (4): 599–613.

    Article  Google Scholar 

  • Healey C, Andradóttir S and Kim S-H (2010). Minimal switching procedure for constrained ranking and selection. In: Johansson B, Jain S, Montoya-Torres J, Hugan J and Yucesan E (eds). Proceedings of the 2010 Winter Simulation Conference, IEEE, Piscataway, NJ, pp 1145–1151.

  • Healey C, Andradóttir S and Kim S-H (2013). A dormancy framework for efficient comparison of constrained systems. European Journal of Operations Research 224 (2): 340–352.

    Article  Google Scholar 

  • Hogan AJ, Morris JG and Thompson HE (1981). Decision problems under risk and chance constrained programming: Dilemmas in the transition. Management Science 27 (6): 698–716.

    Article  Google Scholar 

  • Hunter SR and Pasupathy R (2013). Optimal sampling laws for stochastically constrained simulation optimization on finite sets. INFORMS Journal on Computing 27 (3): 527–542.

    Article  Google Scholar 

  • Hunter SR, Pujowidianto NA, Chen C-H, Lee LH, Pasupathy R and Yap CM (2011). Optimal sampling laws for constrained simulation optimization on finite sets: The bivariate normal case. In: Jain S, Creasey RR, Himmelspach J, White KP and Fu M (eds). Proceedings of the 2011 Winter Simulation Conference, IEEE: Piscataway, NJ, pp 4294–4302.

  • Johnstone DJ and Lindley DV (2011). Elementary proof that mean-variance implies quadratic utility. Theory and Decision 70 (2): 149–155.

    Article  Google Scholar 

  • Kabirian A and Ólafsson S (2009). Selection of the best with stochastic constraints. In: Rossetti MD, Hill RR, Johansson B, Dunkin A and Ingalls RG (eds). Proceedings of the 2009 Winter Simulation Conference, IEEE: Piscataway, NJ, pp 574–583.

  • Kahneman DH and Tversky A (1979). Prospect theory: An analysis of decision under risk. Econometrica 47 (2): 263–290.

    Article  Google Scholar 

  • Keeney RL (1977). The art of assessing multiattribute utility functions. Organizational Behavior and Human Performance 19 (2): 267–310.

    Article  Google Scholar 

  • Keeney RL (2002). Common mistakes in making value trade-offs. Operations Research 50 (6): 935–945.

    Article  Google Scholar 

  • Keeney RL and Raiffa H (1976). Decisions with Multiple Objectives. Wiley: New York, NY.

    Google Scholar 

  • Kim S-H and Nelson BL (2001). A fully sequential procedure for indifference-zone selection in simulation. ACM TOMACS 11 (3): 251–273.

    Article  Google Scholar 

  • Lavalle IH (1987). Response to “Use of sample information in stochastic recourse and chance-constrained programming models”: On the “Bayesability” of CCP's. Management Science 33 (10): 1224–1228.

    Article  Google Scholar 

  • Law AM (2014). Simulation Modeling and Analysis, 5th edn McGraw-Hill: New York, NY.

    Google Scholar 

  • Lee LH, Chey EP, Chen SY and Goldsman D (2010). Finding the non-dominated Pareto set for multi-objective simulation models. IIE Transactions 42 (9): 656–674.

    Article  Google Scholar 

  • Lee LH, Pujowidianto NA, Li L-W, Chen CH and Yap CM (2012). Approximate simulation budget allocation for selecting the best system in the presence of stochastic constraints. IEEE Transactions on Automatic Control 57 (11): 2940–2945.

    Article  Google Scholar 

  • Merrick JRW, van Dorp JR, Mazzuchi T, Harrald J, Spahn J and Grabowski M (2002). The Prince William Sound risk assessment. Interfaces 32 (6): 25–40.

    Article  Google Scholar 

  • Merrick JRW, van Dorp JR, Spahn J, Harrald J, Mazzuchi T and Grabowski M (2000). A systems approach to managing oil transportation risk in Prince William Sound. Systems Engineering 3: 128–142.

    Article  Google Scholar 

  • Morrice DJ and Butler JC (2006). Ranking and selection with multiple “targets”. In: Perrone LF, Wieland FP, Liu J, Lawson BG, Nicol DM and Fujimoto RM (eds). Proceedings of the Winter Simulation Conference, IEEE: Piscataway, NJ, pp 222–230.

  • Mulvey JM, Vanderbei RJ and Zenios SA (1995). Robust optimization of large-scale systems. Operations Research 43 (2): 264–281.

    Article  Google Scholar 

  • Pietrzykowski T (1969). An exact potential method for constrained maxima. SIAM Journal on Numerical Analysis 6 (2): 299–304.

    Article  Google Scholar 

  • Pujowidianto NA, Hunter SR, Pasupathy R, Lee LH and Chen CH (2012). Closed form sampling laws for stochastically constrained simulation optimization on large finite sets. In: Laroque C, Himmelspach J, Pasupathy R, Rose O and Uhrmacher AM (eds). Proceedings of the 2012 Winter Simulation Conference, IEEE, Piscataway, NJ, pp 168–177.

  • Rinott Y (1978). On two-stage procedures and related probability-inequalities. Communications in Statistics: Theory and Methods 7 (8): 799–811.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Appendix A

Appendix A

A.1. Examples of the issues associated with using constraints to capture preferences

Example of Blau’s dilemma: the negative value of perfect information

Suppose we are choosing between two configurations 1 and 2 based on two performance criteria Xm1 and Xm2 and we have the following information about the distributions of attribute performance.

illustration

figure c

Further assume our objective is to select the configuration based on the following rule:

In other words, if expectation of Xm2 for configuration m is not satisfied, E[Xm2]>70, the value to the decision maker for configuration m, f(xm1, E[Xm2])=0; otherwise, f(xm1, E[Xm2])=xm1. As shown in Figure A1, the decision maker should choose Configuration 1: both configurations satisfy the constraint on the mean of Xm2 and so we choose the larger xm1 associated with Configuration 1 and the decision maker receives a value of, f(xm1, E[Xm2])=f(8, 70)=8. Note that although both options fail to satisfy the constraint on X2 half of the time, there is no penalty because E[Xm2]⩽70 for both.

Figure A1
figure 4

Best alternative when performance is defined based on f(xm1, E[Xm2]).

Figure A2 shows the adjusted decision tree when the decision maker receives perfect information concerning the value of Xm2, for Configuration 1, x12. Now, when she learns that x12=90 she chooses Configuration 2 for a value of 3, but when she learns that x12=50 she chooses Configuration 2 for a value of 8. The expected value is 0.5(3)+0.5(8)=5.5. Learning the precise value of x12 results in a loss of 5.5−8=−2.5<0 in expected value when compared with the case with no information. This negative EVPI is commonly referred to as Blau’s Dilemma (Blau, 1974).

Figure A2
figure 5

Best alternative when performance is defined based on f(xm1, E[Xm2]) with Perfect Information about x12.

The remedy for Blau’s dilemma is simple: evaluate the configurations based on the actual realization of Xm2, xm2, rather than E[Xm2]. This change incorporates the fact that, in this example, half of the time each alternative fails to satisfy the constraint, effectively including the constraint as a penalty in the objective function as discussed in Section 3 and above. In other words, we introduce g(xm1, xm2)=xm1 if xm2⩽70; g(xm1, xm2)=0, otherwise. Decisions are then made based on E[g(xm1, xm2)] rather than f(xm1, E[Xm2]); an expected utility function is one of many ways to implement the proper scoring procedure.

As shown in Figure A3, the decision maker whose preferences satisfy this specification still prefers Configuration 1 to Configuration 2, but the expected value of what is received is half of that in Figure EC1 (4 versus 8), due to the infeasibility of x12 half of the time (note that the potential infeasibility of x22 also reduces the value of Configuration 2 by half even though it is not selected).

Figure A3
figure 6

Best alternative when performance is defined based on E[g(xm1, xm2)].

Now when the decision maker is presented with perfect information about the realization of x12 in Figure A4, she chooses Configuration 2 when x12=90 and Configuration 1 when x12=50, for an expected value of 0.5(3)+0.5(8)=4.75. However, note that EVPI=4.75−4=0.75>0=EV and Blau’s dilemma has been resolved. The key feature of evaluations that feature non-negative EVPI is that they reflect the entire distribution of attribute performance and allow tradeoffs or penalties for outcomes that fail to achieve the desired performance levels. A utility function is just one way to capture the proper scoring of the alternatives.

Figure A4
figure 7

Best alternative when performance is defined based on E[g(xm1, xm2)] with Perfect Information about x12.

The previous example provides evidence of the problems associated with putting a constraint on the mean of a performance measure rather than allowing the distribution of performance to determine the value to the decision maker. Some approaches advocate a form of chance constrained programming (CCP) that puts a constraint on the probability that some condition occurs. This formulation also suffers from a negative value of perfect information. Suppose we are choosing between two alternative configurations, Configuration 1 and Configuration 2, based on two criteria that both follow normal distributions: Calls (Xm1) and Accidents (Xm2) for configuration m. Further we are interested in the following CCP formulation which maximizes calls as long as the probability of more than 2.00 accidents is less the 5%:

Further, assume Cfg 1 and Cfg 2 have the following characteristics.

illustration

figure b

Since Pr(X12⩾2.00)⩽0.05 and Pr(X22⩾2.00)⩽0.05, both Configuration 1 and Configuration 2 meet the constraint and we choose Configuration 1, E[X11]= 20>E[X21]= 18. We will refer to this as the expected value of the optimization problem without any additional information, or EV=20.

If we receive perfect information about X12 then we would choose Configuration 1 when X12⩽2.00, Pr(X12⩾2.00)=0.00, which happens with probability 0.9772; with probability 0.0228 we learn that X12>2.00, Pr(X12⩾2.00)=1.00, and select Configuration 2. Hence the expected value with perfect information about X12, EVPI(X12)=0.0228 E[X21]+0.9772 E[X11]=0.0228(18)+0.9772(20)=19.9545. Now we have EVPI(X12)−EV=19.9545−20=−0.0455<0 and the symptom of Blau’s dilemma surfaces again: we are worse off with the information.

In contrast, we could argue that when the constraint on Xm2 is satisfied (not satisfied), u(Xm2)=1 (0), or u(Xm2)=0 × Pr(Xm2>2.00)+1 × Pr(Xm2⩽2.00)=Pr(Xm2⩽2.00) and assume that u(Xm1)=E[Xm1]/20 without loss of generality. Further, let u(Xm1, Xm2)=0.5 u(Xm1)+0.5 u(Xm2).

If we maximize u(Xm1, Xm2) with no additional information

illustration

figure a

We would choose Configuration 1: u(X11, X12)=0.9886=EV>u(X21, X22)=0.9386

If we are able to gather perfect information about X12, then (as shown in Figure A5)

Figure A5
figure 8

Best alternative when performance is defined based on E[u(Xm1, Xm2)] with Perfect Information about X12.

EVPI(X12)=0.9986 and EVPI(X12)−EV=0.9986−0.9886=0.0010>0. We are better off with the information and Blau’s dilemma has been resolved.

In both of these examples the root cause of the problem is that the constrained variable is not part of the objective function as we discuss in detail in Section 2. Featuring both criteria in the objective function allows for a tradeoff between them. For example, if one wants to focus exclusively on satisfying the constraint on Xm2 then w2=1. In this extreme case, the alternatives would be sorted in increasing order of satisfying the constraint. The probability thresholds for the CCP constraints play an analogous role to the weights in an MAU model. These thresholds would have to be assessed and it is not clear how one would do that while there is a long literature on weight assessment in muliattribute utility theory.

Further, consider an extreme case such as Pr(Xm2⩾2.00)⩽0.000001. It is likely that there will no systems that satisfy this constraint and, similar to the setting in Section 2, the decision maker will have to increase the right hand side until one or more systems become feasible. Again, it is not clear that these are really hard constraints at all. In fact Hogan et al (1981) remind us statements like Xm2⩾2.00 are goals and Pr(Xm2⩾2.00)⩽0.000001 is a probability of attaining that goal. Further, note that when using multiattribute utility theory, even if no systems satisfy the desired target of Xm2⩽2.00 the decision maker gets feedback on the performance of every systems in addition to the other benefits discussed in Section 2.

Aside from being a reasonable property, positive EVPI is important when making decisions about what kind of information to gather. For example, suppose we are considering doing some additional testing on factors that could affect our estimates of the distribution of the number of accidents, Xm2, e.g. the effect of adding additional escorts for vessels. Or perhaps we are considering allocating scarce CPU cycles to generate additional replications to reduce the standard error of our performance estimates of Xm2. Given the CSO framework we could conclude that there is no need to gather any more information for X m2 ; in fact we are better off without the additional information. Put another way, a negative EVPI is really a symptom of a more fundamental problem: a lack of tradeoffs in the objective function. Further, given that MAU is no more computationally expensive that other approaches when analyzing the same problem, it is unclear what benefits accrue from the use of other approaches. At first glance one might assume that there are more required assessments with MAU, however, as we argue here and in the paper, when you think carefully about other approaches there are other unknown parameters that must be estimated. For example, with CCP you must estimate the probability threshold for each goal and with constraint based optimization, what are the constraint levels in CSO? If we use a ‘soft’ constraint we’re back to assessing a utility function, etc.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Merrick, J., Morrice, D. & Butler, J. Using multiattribute utility theory to avoid bad outcomes by focusing on the best systems in ranking and selection. J Simulation 9, 238–248 (2015). https://doi.org/10.1057/jos.2014.34

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1057/jos.2014.34

Keywords

Navigation