Payoffs, alternatives, and expected monetary values are terms associated with
A] virtual reality
B] Product Lifecycle Management
C] Quality Function Deployment
D] decision trees
E] make-or-buy analysis
Certainty Equivalent.The minimum guaranteed amount one is willing to accept to avoid the risk associated with a gamble.
Coefficient of Realism [α].A number from 0 to 1 such that when α is close to 1, the decision criterion is optimistic, and when α is close to zero, the decision criterion is pessimistic.
Conditional Value, or Payoff.A consequence or payoff, normally expressed in a monetary value, which occurs as a result of a particular alternative and outcome.
Decision Alternative.A course of action or a strategy that can be chosen by a decision maker.
Decision Making Under Certainty.A decision-making environment in which the future outcomes are known.
Decision Making Under Risk.A decision-making environment in which several outcomes can occur as a result of a decision or alternative. Probabilities of the outcomes are known.
Decision Making Under Uncertainty.A decision-making environment in which several outcomes can occur. Probabilities of these outcomes, however, are not known.
Decision Table.A table in which decision alternatives are listed down the rows and outcomes are listed across the columns. The body of the table contains the payoffs.
Efficiency of Sample Information.A ratio of the expected value of sample information and the expected value of perfect information.
Equally Likely.A decision criterion that places an equal weight on all outcomes. Also known as Laplace.
Expected Monetary Value [EMV].The average or expected monetary outcome of a decision if it can be repeated many times. This is determined by multiplying the monetary outcomes by their respective probabilities. The results are then added to arrive at the EMV.
Expected Opportunity Loss [EOL].The average or expected regret of a decision.
Expected Value of Perfect Information [EVPI].The average or expected value of information if it is completely accurate.
Expected Value with Perfect Information [EVwPI].The average or expected value of the decision if the decision maker knew what would happen ahead of time.
Expected Value of Sample Information [EVSI].The average or expected value of imperfect or survey information.
Maximax.An optimistic decision-making criterion. This is the alternative with the highest possible return.
Maximin.A pessimistic decision-making criterion that maximizes the minimum outcome. It is the best of the worst possible outcomes.
Minimax Regret.A decision criterion that minimizes the maximum opportunity loss.
Opportunity Loss.The amount you would lose by not picking the best alternative. For any outcome, this is the difference between the consequences of any alternative and the best possible alternative. Also called regret.
Outcome.An occurrence over which the decision maker has little or no control. Also known as a state of nature.
Risk Avoider.A person who avoids risk. As the monetary value increases on the utility curve, the utility increases at a decreasing rate. This decision maker gets less utility for a greater risk and higher potential returns.
Risk Neutral.A person who is indifferent toward risk. The utility curve for a risk-neutral person is a straight line.
Risk Premium.The monetary amount that a person is willing to give up in order to avoid the risk associated with a gamble.
Risk Seeker.A person who seeks risk. As the monetary value increases on the utility curve, the utility increases at an increasing rate. This decision maker gets more pleasure for a greater risk and higher potential returns.
Sequential Decisions.Decisions in which the outcome of one decision influences other decisions.
Utility Curve.A graph or curve that illustrates the relationship between utility and monetary values. When this curve has been constructed, utility values from the curve can be used in the decision-making process.
Utility Theory.A theory that allows decision makers to incorporate their risk preference and other factors into the decision making process.
The addition to the criterion of MEMV of three simple statistical measures, variance, volatility, and probability of exceeding a fixed value of the return of a project, sheds significant light on decisions. The probability of exceedance can be used to mandate a priori a fixed probability of monetary return that would be acceptable, and from which the range of parameters of a project that are worthwhile or not to bid for can be assessed. If in a project there exists a chance that it can fail, and because a minimum bid is considered the amount just meeting the cost of the project under optimum conditions, multiples of this bid will of course increase the MEMV correspondingly, but no matter how large these bids are they will not improve the uncertainty about the return of the project.
Inclusion of an extreme scenario will always yield worse than expected returns than if such a case were not considered. However, the full effect of a catastrophic situation is not revealed in the expected return through the MEMV since this does not differ significantly from that of a limited spill case but in the substantial increase in uncertainty. This may lead to overdesign and additional measures for systems backup, monitoring, etc., and is perhaps the reason why in many projects catastrophic scenarios are not considered. To compound the situation, the probability of extreme events cannot be assessed through the frequentist theory of probability, i.e., of the notion that a specific event appearing m times out of a total of N trials has a probability m/N, and, if instead experts' opinions were utilized it has been seen that the human brain has difficulty in assessing the likelihood of extreme events [Kahneman and Tversky, 1979, 2000; Kahneman, 2012].
View chapterPurchase book
Read full chapter
URL: //www.sciencedirect.com/science/article/pii/B9780128095829000025
Costs of Accidents, Costs of Safety, Risk-Based Economic Decision Making
Hans Pasman, in Risk Analysis and Control for Industrial Processes - Gas, Oil and Chemicals, 2015
9.3.7 Decision analysis and decision trees
Decision making has a binary nature—we go for it or not. The primary objective of decision analysis is to identify the decision alternative that maximizes expected utility or expected monetary value with probability of occurrence as the outcome consequence weight factors. As mentioned in the introduction to this section, decision trees are a means to structure decision making taking account of the various aspects or components and motivate gathering the needed information. In computer science much use is made of binary decision trees. Binary refers to a Boolean basis: an aspect must be reckoned with or not, it is true or false, the value is one or zero [as in a truth table]. In fact, it is embodying the “if-then-else” rule. Binary decision trees are very useful in development of digital systems. The tree that branches up to the final decision is a directed acyclic graph.
Here, we are more interested in decision trees that include uncertainty as required for a system approach. Ian Jordaan,27 Memorial University of Newfoundland, described the field. Basic is the distinction of a [binary] choice node, which Jordaan calls a decision fork, followed at each branch by a probability node, or chance fork. In Figure 9.6, a simple example using point values is given of risk-based decision making use of the module PrecisionTree of [email protected] MS Excel-based decision analysis software [easily found on the Internet]. In a process under normal condition a light protective measure is adequate, but one needs a heavy protective measure if a coincidental process condition materializes. From the choice of protection four end states arise, called consequences or utilities because they represent values to the decision maker. Two of these end states can be classified as adequate protection, the other two as under- and over-protection. Hence, basically, the choice depends on how the decision maker perceives the chance or probability that the condition will occur. Here, an increase in occurrence probability of 0.08 to a value of 0.1 will change the preference from light to heavy protection.
Figure 9.6. Top left and right: Shown is an example of a decision tree. As explained in the text, the decision is about choice of a protection system: Light costing €1000 but only adequate for normal process situation, or heavy €10,000. In case of underprotection, damage sustained by the installation is €100,000. Left: For a coincidental hazardous process condition estimated to occur 8% of times or occurrence probability of 0.08, light protection is the best choice based on minimum cost. Right: At occurrence probability of 0.1 or higher, heavy protection makes sense. The calculation to compare monetary value of the two decision alternatives was made by Palisade's Precision Tree. Bottom: The same calculation made with a Bayesian net by means of GeNIe v.2.0 of Decision Systems Laboratory of the University of Pittsburgh [see Section 7.5].
An additional possibility is collecting more information about circumstances influencing the emergence of the coincidence. A value of information calculation can be cost-effective to lower uncertainty and reduce the risk of decision under uncertainty. Developing the knowledge by, for example, testing, requires funding, but the value of this information must be balanced against the gain in knowledge and reduction in uncertainty. The larger the uncertainty, the higher the value of the information and at lower cost to obtain than when the uncertainty is at lower levels. This calculation will form a pre-decision node. In case testing would be very costly or not possible, improved information could be gained through estimation by employing so-called pre-posterior analysis. This is by simulating a posterior distribution by taking the prior and estimate the conditional probabilities of what you would observe in a test [this is a kind of contingency analysis]. Another value of information is the value of control to reduce uncertainty of outcomes. Prior to finalizing a decision, both approaches can be simply added. The bad state probability may be strongly reduced. For this simple case, overview can be kept easily, but evaluating for example a complete bow-tie will be different. Apart from enabling to cope with complexity, the result of the calculation with the software offers clarity in team communication and later review.
By combining PrecisionTree with @Risk, or by calculation in Bayesian nets, uncertainties in the probabilities of occurrence and of consequences can be included. The Precision Tree software allows the tree to be converted to the appearance of an influence diagram. The GeNIe Bayesian Net can do this too, while it is at the same time more versatile and able to calculate a result using distributions of all relevant information.
View chapterPurchase book
Read full chapter
URL: //www.sciencedirect.com/science/article/pii/B9780128000571000092
Economic Analysis
Tarek Ahmed, D. Nathan Meehan, in Advanced Reservoir Management and Engineering [Second Edition], 2012
7.7.3 Decision Trees and Utility Theory
Decision trees are a useful way to describe alternative scenarios and select the decision that maximizes the NPV or whatever the decision maker is trying to optimize. In a subsequent section, we will see that “utils” can express the relative desirability of various outcomes. In decision theory, the most desired outcome is based on the goals and preferences of the decision maker. The reservoir engineer can use decision trees to describe complex scenarios with multiple decisions and multiple probabilities. This discussion can be considered only a brief introduction. In constructing a decision tree, we use rectangles to represent decision nodes and circles to represent probability nodes. Two or more decisions can be associated with each decision node, and multiple nodes can be associated with a probability node. A probability node representing betting $1000 on number “30” at a roulette wheel11 in Las Vegas is shown in Figure 7.11
Figure 7.11. Probability node.
The single bet on number 30 can easily be evaluated as to its expected value as follows:
EV=−1000$+[138]×$36,000+[3738]×0=−$52.63
In other words, a single bet of $1000 on number 30 [or any other number] has a negative expected value of $52.63. Similar analyses will show negative expectations for each of the gambling games explaining the fabulous hotels and inexpensive “all you can eat” buffets in Las Vegas. But is it crazy to play roulette or make other decisions selecting lower expected values than other alternatives? No, the decider may have a different use for $35,000 than $1000. Maybe he owes a debt that is immediately due and has a major negative result if he is unable to generate $35,000 right away. This particular preference for risk is actually unusual; most people have less utility for expected outcomes that have large negative impacts. This analysis does not mean that every player will lose money playing roulette. It is a relatively straightforward exercise to model a roulette wheel with various strategies in which a significant fraction of the players win.12 It is the aggregate EMV of all players over the long run that is negative.
Suppose someone gives you the chance to play a game in which a fair coin is flipped. In the case of heads,13 you receive $2 and for tails you get nothing. You will no doubt be happy to play this game as it has an expected monetary value [EMV] of $1. How much would you be willing to sell your ticket for? It is unlikely anyone will pay you much more than $1, and if you sell it for much less you are “giving away” EMV. Now consider another game. In this game you have to buy a ticket. In this game a heads pays $3 and a tail pays $1. How much would you be willing to pay for this ticket? The EMV of this game is $2, and if you pay any less than that you are [on an expected value basis] gaining money. Would you pay more than $1? If you paid $1, the second game becomes equivalent to the first with the net result of a head being $3−1=$2 and the result of a tail would be $1−1=$0. Is there a difference in how much you are willing to sell your ticket for in the first game and what you are willing to pay for it in the second game? Decision makers often make decisions on other than an expected value basis based on how much investment exposure is necessary.
Let us consider another set of decisions. In the first option, you pay €1000 by investing in a very small percentage [0.1%] of a drilling well that you anticipate has a 50% chance of success [or a coin flip for heads if you prefer]. In the case of a discovery you win a series of cash flows with an NPV of €4000, while a dry hole pays nothing. The EMV is 0.5×€4000−1000=1000. Are you interested in this investment? If you believe these numbers and have €1000 to invest, it is an obvious decision to participate in the project. Now let us look at the 100% working interest position. In this case, you need to invest €1,000,000 and have a 50% chance of €4,000,000. Assume that your net worth is just enough that you could come up with the money by mortgaging your house, cashing in your retirement, and borrowing all of the money that you can; it is unlikely that you would accept such an investment opportunity. A single investment or a series of investments that has the potential to bankrupt an investor is known as “gambler’s ruin.” Your utility for a positive €1,000,000 is considerably less than 1000 times greater than it is for €1000. By analyzing your responses to a series of similarly constructed alternatives, an individual with game theory expertise could construct your “indifference curve.” Your personal utility and indifference curves and those of the decision maker are not as important as are the utility functions of the corporation. For our purposes, we will assume that the corporation has a unit slope linear utility function and makes its decisions entirely on EMV. Exceptions to this would only occur for massive investments.
In the drill vs. farmout example, we had a decision tree, see Fig. 7.12.
Figure 7-12. Decision tree.
There were only two decisions: drill and farmout. The probability nodes were only dry hole or discovery. The analysis of a decision tree proceeds from right to left as the EMV is calculated for each probability node. The expected value of each probability node is replaced with its expected value, and the highest EMV decision node is selected. There can be multiple probabilities at each probability, and the probability node can be replaced by Monte Carlo simulations. In fact, the entire decision tree can be replaced by Monte Carlo simulations with a distribution of decisions being made and the corresponding variability in results conveyed to decision makers.
View chapterPurchase book
Read full chapter
URL: //www.sciencedirect.com/science/article/pii/B9780123855480000075
Decisions in Engineering Design
Kuang-Hua Chang, in Design Theory and Methods Using CAD/CAE, 2015
2.3.3 Decision Under Risk
In this section, we revisit the decision under risk, in which the probabilities that affect the outcomes usually are assumed to be known. We introduce more rigorous treatment to the decision-making model, in which we revisit the car-buying example for illustration.
To facilitate our discussion, we assign N and U to the options of new and used car, respectively; and W and F to the events of car works well without major repairs and car fails due to major problems, respectively. The buyer also consulted with the dealership, and the historical data suggest that there is an 80% chance that a new car does not encounter major problems in 10 years. The historical data also suggest that the probability is 50% for a used car. At the time, this was the best possible estimate that the buyer was able to attain. Next, we construct a decision tree to aid the buyer in making a decision. The payoff table of Table 2.4 is expanded by incorporating the probabilities and expected values. In Table 2.6, the notation P[W|N] stands for the probability of new car [N] that works [W] well in 10 years. Similarly, P[F|U] stands for the probability of used car [U] that fails [F] in 10 years [i.e., encountering major problems], and so on.
Table 2.6. Cost of Different Options and Events [Payoff Table]
States of Nature [Events]Courses of ActionNew Car [N]Used Car [U]ProbabilityPayoffExpected ValueProbabilityPayoffExpected ValueWithout major problems [W]P[W|N] = 0.8$25,000$20,000P[W|U] = 0.5$15,000$7500With major problems [F]P[F|N] = 0.2$35,000$7000P[F|U] = 0.5$30,000$15,000$27,000$22,500
Assuming these historical data are reliable, the payoff table and the probability estimates can be combined to arrive at the expected payoff of individual decisions. The expected payoff is also called the expected monetary value [EMV] in decision theory. The calculations are summarized as follows:
[2.3]E[ai]=∑jP[ϕj]v[ai,ϕj]
where E[ai] is the EMV of event ai.
Hence, for the event of buying a new car, the expected payoff is
E[N]=∑jP[ϕj]v[N,ϕj]=P[W|N]v[N,W]+P[F|N]v[N,F]=0.8[$25,000]+0.2[$35,000]=$27,000
Similarly, for the event of buying a used car, the expected payoff is E[U] = $22,500, as shown in Table 2.6. Therefore, based on the expected payoff, buying a used car presents a better option.
The car-buying example can be represented in a decision tree similar to that of Section 2.2.2. The decision tree for the car-buying example is shown in Figure 2.4.
FIGURE 2.4. Decision tree for a car buyer: [a] starting tree and [b] solution tree.
As discussed in Section 2.2.2, the general approach to solving a decision tree is to move backward through the tree [from right to left] until we reach the originating decision node. We select a payoff node and move left to trace the branch to encounter the next node. If the next node is a chance node, we calculate the expected value [E] of all nodes connected immediately to the right of the encountered node by using Eq. 2.3. In this example, we calculate the expected values for the options of buying a new car E[N] and buying a used car E[U], respectively.
We enter these values to their respective chance nodes, and then move left to encounter a decision node—in this case, the originating decision node, as shown in Figure 2.4b. At the decision node, we select the branch that leads to the best value. In this case, $22,500 is the best value, representing a lesser overall cost. At this point, we reach a decision of buying a used car.
One of the difficulties in using the decision tree method is coming up with the probabilities of the uncertain event occurring. In the car-buying example, how certain is the event that the probability of a used car without major problem is P[W|U] = 50%? In this case, we do not have to agonize too much over the accuracy of the estimate because we can easily test to see if the decision to buy a used car is sensitive to the probability estimated. We first let the probability be P[W|U] = x. Then the probability of a used car having major problems is P[F|U] = 1 − x. The expected value of buying a used car is
E[U]=x[$15k]+[1−x][$30k]=$30k−$15kx
If we equate the two expected values of used and new car E[U] = E[N] = $27k, we have x = 0.2. In other words, as long as the probability of a used car without major problems is greater than 20%, buying a used car is still a good choice.
View chapterPurchase book
Read full chapter
URL: //www.sciencedirect.com/science/article/pii/B9780123985125000025
Units, acronyms, and glossary
Stephen A. Rackley, in Carbon Capture and Storage [Second Edition], 2017
25.2 CCS-related acronyms
A
AABW
Antarctic bottom water
ACFCActivated carbon fiber cloth
ACSAgricultural carbon sequestration
AERAdsorption enhanced reforming
AGRAcid gas removal
AGSAcid gas storage
ALARPAs low as reasonably practicable
ALWAccelerated limestone weathering
AORArea of review
ARAssurance review
AR4/5Fourth/Fifth Assessment Report [IPCC 2007/2014]
ARDAfforestation, reforestation, and deforestation
ARWAmine reclaimer waste
ASCMAdsorption-selective carbon membranes
ASUAir separation unit
ATTotal alkalinity
A-USCAdvanced ultrasupercritical
AVOAmplitude versus offset
AZEPAdvanced zero-emission power plant
B
BAT
Best available technology
BCABenefit–cost analysis; belowground carbon allocation
BCMBiologically controlled mineralization
BECCSBiomass energy with carbon capture and storage [or sequestration]
BFBBubbling fluidized bed
BFLMBulk flow liquid membrane
Biomass integrated gasification combined cycle
BIMBiologically induced mineralization
BLGCCBlack-liquor gasification combined cycle
BOSBasic oxygen steelmaking
C
CA
Competent authority
CAAAClean Air Act Amendments
CAFCost to avoid a fatality
CAPChilled ammonia process
CAPMCapital asset pricing model
CATCarbon abatement technologies
CBACost–benefit analysis
CBLCement bond log
CCCombined cycle
CCGSCarbon capture and geological storage
CCGTCombined cycle gas turbine
CCSCarbon capture and storage [or sequestration]
CCSTNCanadian Carbon Capture and Storage Technology Network
CCTClean coal technology
CCUCarbon capture and utilization
CDCLCoal direct chemical looping
CDMClean Development Mechanism
CDMCClimate Decision Making Center
CERsCertified emission reductions
CFCertification framework
CFBCCirculating fluidized bed combustion
CFDComputational fluid dynamics
CGSCarbon geological storage
CLCChemical looping combustion
CLGChemical looping gasification
CLMContained liquid membrane
CLOUChemical looping with oxygen uncoupling
CLRChemical looping reforming
CMICarbon Mitigation Initiative
CMMVCharacterization, modeling, monitoring, and verification
CMSMCarbon molecular sieve membranes
[CO2]CO2 concentration
CPGCO2 plume geothermal
CSEMControlled-source electromagnetic method
CSEGRCarbon storage with enhanced gas recovery
CSHCalcium silicate hydrate [cement gels]
CSLFCarbon Sequestration Leadership Forum
CS-SSGSCarbon storage in sub-seabed geological structures
CTIClimate Technology Initiative
CTDConductivity temperature depth
CVIChemical vapor infiltration
D
DCF
Discounted cash flow
DEADiethanolamine
DFNDiscrete fracture network
DGDecision gate
DGPSDifferential global positioning systems
DICDissolved inorganic carbon
DInSARDifferential interferometric synthetic aperture radar
DMEPGDimethyl ethers of polyethylene glycol
DNVDet Norske Veritas
DoEUS Department of Energy
DRADeterministic risk analysis
DSFDeep saline formation
DTSDistributed temperature sensor
E
ECBM
Enhanced coal-bed methane
ECCPEuropean Climate Change Programme
EFEPExternal features, events, processes
EGREnhanced gas recovery
EGSEngineered or enhanced geothermal system
EIAEnvironmental impact assessment; Energy Information Agency [US DOE]
EISEnvironmental impact statement
EMVExpected monetary value
EOREnhanced oil recovery
EOSEquation of state
EPA[US] Environmental Protection Agency
EPRIElectric Power Research Institute
EPSExtracellular polymeric substances [biofilm]
EQSEnvironmental quality standards
ERTElectrical resistance tomography
ESAElectrical swing adsorption
ESPElectrostatic precipitator
ETS[EU] Emissions Trading Scheme
F
FACE
Free air carbon dioxide enrichment
FARFirst Assessment Report [IPCC 1990]
FBCFluidized bed combustion
FCCCFramework Convention on Climate Change
FEEDFront-end engineering design
FEPFeatures, events, and processes
FESEMField emission scanning electron microscopy
FGDFlue gas desulfurization
FIDFinal investment decision
FOCEFree ocean carbon dioxide enrichment
G
GCEP
Global Climate and Energy Project
GCSGeological carbon storage
GFBCCGasification fluidized-bed combined cycle
GHGGreenhouse gas
GISGeographical information system
GSGeological storage
Gt-CO2Gigaton CO2 [109 metric tonnes=1012 kg]
GTLGas to liquids
GWGigawatt
GWPGlobal warming potential
H
H
Enthalpy
HATHumid air turbine
HAZOPHazard and operability
Hydrate-based gas separation
HDSHydrodesulfurization
HFCLMHollow-fiber contained liquid membrane
HFMCHollow-fiber membrane contactor
HHVHigher heating value
HNLCHigh nutrient, low chlorophyll
HPHigh pressure
HRSGHeat-recovery steam generator
HSEHealth, safety, and environment [also SHE or HE]
I
IAPWS
International Association for the Properties of Water and Steam
IEAInternational Energy Agency
IGCCIntegrated gasification combined cycle
IGFCIntegrated gasification fuel cells
IGHATIntegrated gasification humid air turbine
ILIonic liquid
ILMIonic liquid membrane, Immobilized liquid membrane
InSARInterferometric synthetic aperture radar
IPCCUN Intergovernmental Panel on Climate Change
IRCCIntegrated reforming combined cycle
IRRInternal rate of return
ITMIon transport membrane
IUPACInternational Union of Pure and Applied Chemistry
K
kPa
Kilopascal
kWKilowatt
L
LCA
Life cycle analysis
LEERTLong electrode electrical resistance tomography
LHVLower heating value
LIDARLight detection and ranging
LNGLiquefied natural gas
LMLiquid membrane
LPLow pressure
LULUCFLand use, land use change and forestry
M
MAOM
Mineral associated organic matter
MCFCMolten carbonate fuel cell
MCLMaximum contaminant level
MCMMixed conducting membrane
MEAMonoethanolamine
MECCMixed electron carbonate conductor
MECSMicroencapsulated carbon sorbents
MFCMicrobial fuel cell
MGAMembrane gas absorption
MGEMicrobial growth efficiency
MICMicrobially influenced corrosion
MIC[C]PMicrobially induced calcite [or calcium carbonate] precipitation
MIECMixed ionic electronic conductors
MMMMixed matrix membranes
MMVMeasurement [or Modeling], monitoring, and verification
MOCCMixed oxide carbonate conductor
MOFMetal organic framework
MOMMicrobial organic matter
MPaMegapascal
MSCMolecular sieve carbon
MSWMunicipal solid waste
Mt-CO2Megaton CO2 [106 metric tonnes=109 kg]
MVARMonitoring, verification, accounting, and reporting
MWMolecular weight
MWeMegawatts electric power
MWthMegawatts thermal power
MWIMunicipal waste incineration
N
NADW
North Atlantic deep water
NBPNormal boiling point [at 1 bar]
NGCCNatural gas combined cycle
NGLNatural gas liquids
NOAANational Oceanic & Atmospheric Administration [US Department of Commerce]
NOELNo observed effects limit
NOxMono-nitrogen oxides [NO, NO2]
NPVNet present value
NSPSNew source performance standards
O
OIF
Ocean iron fertilization
OMAOcean macroalgal afforestation
ONSOrdered nanoporous silica
OTMOxygen transport membrane
P
P
Pressure
PAPerformance assessment
PcCritical pressure
PCPulverized coal
PCCPulverized coal combustion, Post-combustion capture
PCFBCPressurized circulating fluidized bed combustion
PCSFPost-closure stewardship fund
PFPulverized fuel
PFBCPressurized fluidized bed combustion
PFTPerfluorocarbon tracer
PICParticulate inorganic carbon
PIDProcess influence diagram
PISCPost-injection site care
PIRPost-implementation review
PLONORPose little or no risk
POCParticulate organic carbon
POMPartial oxidation of methane; Particulate organic matter
POXPartial oxidation
ppbparts per billion [10−9]
PPCCPressurized pulverized coal combustion
ppmparts per million [10−6]
pptparts per trillion [10−12]
PRAProbabilistic risk analysis
PSAPressure swing adsorption
PSHAProbabilistic seismic hazard analysis
PUPorosity unit [1 PU=1% porosity]
Present value
PVTPressure, volume, temperature
Q
QRA
Quantitative risk assessment
R
RAM
Risk assessment matrix
RBCARisk-based corrective action
RCPReference concentration pathway
RD3Research, development, demonstration, and deployment
RDFRefuse derived fuel
RFARegulatory framework assessment [Alberta]
RORRate of return
RPRecommended practice
RTILRoom temperature ionic liquid
S
S
Entropy
SAPOSilicoaluminophosphate
SARSynthetic aperture radar; Second Assessment Report [IPCC 1996]
SAUStorage assessment unit
SCSupercritical
scCO2Supercritical CO2
SCCStress corrosion cracking
SCPCCSupercritical pulverized coal combustion
SCRSelective catalytic reduction
SDMSurface deformation monitoring
SERSorption-enhanced reaction; sorption-enhanced reforming
SE-SMRSorption-enhanced steam methane reforming
SEWGSSorption-enhanced water–gas shift
SICSoil inorganic carbon
SILMSupported ionic liquid membrane
SLMSupported liquid membrane
SMBSimulated moving bed
SMBCSoil microbial biomass carbon
SMRSteam methane reforming
SNCRSelective non-catalytic reduction
SOCSoil organic carbon
SOFCSolid oxide fuel cell
SOMSoil organic matter
SOxOxides of sulfur [SO, SO2, SO3]
SPCCSolar-enhanced post-combustion capture
SPESociety of Petroleum Engineers
SPSSwitchable polarity solvents
SRBSulfate reducing bacteria
SRESSpecial Report on Emissions Scenarios [IPCC]
SSGSSub-seabed geological storage [or structures]
STIGSteam injected gas turbine
STLSubmerged turret loading
STPStandard temperature and pressure; Social time preference
SWAGSimultaneous water and gas [injection]
STPStandard temperature and pressure [IUPAC; 0°C, 100 kPa]
T
T
Temperature
TALKTotal alkalinity
TARThird Assessment Report [IPCC 2001]
TASRTechnically available storage resource
TBCATotal belowground carbon allocation
TcCritical temperature
TCTotal carbon content
TCO2Total CO2 content
TDSTotal dissolved solids
TEELTemporary emergency exposure limit
THCThermohaline circulation
THMCThermal hydraulic mechanical chemical [coupled modeling]
TICTotal inorganic carbon
TOCTotal organic carbon
TORTransfer of responsibility; Terms of reference
TQTop quartile; Technical qualification
TRLTechnology readiness level
TSATemperature swing adsorption
TSILTask-specific ionic liquid
U
UCG
Underground coal gasification
UICUnderground injection control
UNFCCCUnited Nations Framework Convention on Climate Change
USCUltrasupercritical
USDWUnderground sources of drinking water
V
V
Volume
V&VValidation and verification
VEFVulnerability evaluation framework
VOCVolatile organic compounds
VOIValue of information
VPSAVacuum pressure swing adsorption
VSPVertical seismic profile
W
WAG
Water-alternate-gas
WGSRWatergas shift reaction
Z
ZECA
Zero-Emission Coal Alliance
ZETZero-emissions technologies
ZEIGCCZero-emissions integrated gasification combined cycle
View chapterPurchase book
Read full chapter
URL: //www.sciencedirect.com/science/article/pii/B9780128120415000258
A review on closed-loop field development and management
Abouzar Mirzaei-Paiaman, ... Denis J. Schiozer, in Journal of Petroleum Science and Engineering, 2021
7.2 Objective functions
Based on the type of optimization [i.e., nominal vs. robust], the objective function can accordingly be either deterministic or probabilistic. While deterministic objective functions are evaluated for a single model only, probabilistic objective functions are evaluated across multiple models. Deterministic objective functions can be financial or production/injection indicators such as NPV, revenue, recovery factor, cumulative fluid production [oil, gas or water], cumulative fluid injection [water, gas, solvent or CO2], displacement efficiency at water breakthrough [Sudaryanto and Yortsos, 2011], sweep efficiency through equalizing the arrival times of water/gas front at all producers [Alhuthali et al., 2007, 2008, 2009; Elfeel et al., 2018], and water breakthrough time [Bagherinezhad et al., 2017]. Probabilistic objective functions are usually described as expected indicators such as EMV, expected recovery factor, expected cumulative fluid production [oil, gas or water], and expected cumulative fluid injection [water, gas, solvent or CO2]; sometimes combined with a risk measure.
Those previously described closed-loop studies that have used nominal optimization, RM nominal optimization or ensemble nominal optimization, represent the cases of using deterministic objective functions [with discounted or undiscounted NPV as a widely used objective function]. Furthermore, those working with RM robust optimization or robust optimization have used probabilistic objective functions [with expected discounted or undiscounted NPV as a widely used objective function]. Alhuthali et al. [2009] and TAMU in Peters et al. [2010] used expected arrival time of water fronts as the objective function in their water-flooding RM robust optimization problems.
Furthermore, regardless of deterministic or probabilistic optimization, the optimization problem could be either single or multi-objective, depending on the number of objective functions considered. Single-objective optimization aims to maximize or minimize a certain objective function [measures of production, injection, economic, risk, etc]. Another form of single-objective optimization also exists in which optimization is performed by lumping several objective functions into a single general balanced objective function, each objective function having its own weight [Marler and Arora, 2004]. Nevertheless, the difficulty is finding the suitable weighting factor corresponding to each objective function. As the weighting factors strongly govern the characteristics of the optimal solution, a vast number of trial and error runs with different weighting factors may be required to obtain a satisfactory solution [van Essen et al., 2011].
The practical optimization problems should normally consider multiple, possibly competitive and conflicting, objectives [Yasari et al., 2013; Moradi and Rasaei, 2017]. The multi-objective [or multi-criterion] optimization overcomes the difficulty of the single-objective optimization to address objectives with differing data types, to accommodate multiple objectives, and to handle the possible conflicts between objectives [Isebor and Durlofsky, 2014; Hutahaean et al., 2019]. For instance, in a water flood project, one may be interested in maximizing oil recovery while minimizing water injection, or maximizing produced oil while minimizing produced water.
In simultaneous multi-objective optimization, several objective functions are optimized simultaneously. Usually, the final optimal solution set [Pareto front] provides different solutions for decision-makers to select the production strategy by trade-off between objectives [Bagherinezhad et al., 2017; Hutahaean et al., 2019]. Yasari et al. [2013] performed a multi-objective robust optimization to optimize the different components of NPV under economical and geological uncertainty with the aim of omitting the relevancy of the optimization problem to the prices. Liu and Reynolds [2015, 2016], Isebor and Durlofsky [2014], Yasari and Pisvaie [2015] and Hutahaean et al. [2019] documented cases where the optimization objectives were to maximize the expected NPV while minimizing its associated uncertainty [standard deviation] over a set of models. Liu and Reynolds [2016] studied an optimization case where the objective was to maximize life-cycle NPV and to maximize the short-term NPV of production. In a work by Bagherinezhad et al. [2017], a procedure was applied for reservoir development optimization subject to maximization of the cumulative oil production and minimization of water front velocity [or respectively maximization of water breakthrough time]. Hasan et al. [2013] documented a case where short-term and life-cycle objective functions were optimized simultaneously.
Although production optimization studies normally focus on a life-cycle window, in practice short-term objectives usually dictate the course of the production strategy, especially in view of geological and economic uncertainties [van Essen et al., 2011; Chen et al., 2012]. Therefore, short-term objectives should also be incorporated into the life-cycle optimization problem [Pinto et al., 2015]. Following Jansen et al. [2009], who showed that a life-cycle performance could be optimized while maintaining freedom to perform short-term production optimization, van Essen et al. [2011], Chen et al. [2012] and Fonseca et al. [2014] utilized hierarchical optimization processes where maximization of the life-cycle NPV served as the primary objective and maximization of the short-term operational performance was the secondary objective [short-term in the context of reservoir engineering, in contrast to production engineering]. In their approach, optimality of the primary objective function constrains the secondary optimization problem. In other words, optimization of the second objection function is constrained by the requirement that the primary objective function must remain close to its optimal value.
To the best of our knowledge, all of the previous closed-loop studies have been based on optimization of a single-objective function.