ﻻ يوجد ملخص باللغة العربية
Least worst regret (and sometimes minimax) analysis are often used for decision making whenever it is difficult, or inappropriate, to attach probabilities to possible future scenarios. We show that, for each of these two approaches and subject only to the convexity of the cost functions involved, it is always the case that there exist two extreme scenarios whose costs determine the outcome of the analysis in the sense we make clear. The results of either analysis are therefore particularly sensitive to the cost functions associated with these two scenarios, while being largely unaffected by those associated with the remainder. Great care is therefore required in applications to identify these scenarios and to consider their reasonableness. We also consider the relationship between the outcome of a least worst regret and a Bayesian analysis, particularly in the case where the regret functions associated with the scenarios largely differ from each other by shifts in their arguments, as is the case in many applications. We study in detail the problem of determining an appropriate level of electricity capacity procurement in Great Britain, where decisions must be made several years in advance, in spite of considerable uncertainty as to which of a number of future scenarios may occur, and where least worst regret analysis is currently used as the basis of decision making.
We study the design of autonomous agents that are capable of deceiving outside observers about their intentions while carrying out tasks in stochastic, complex environments. By modeling the agents behavior as a Markov decision process, we consider a
We propose a new approach for solving a class of discrete decision making problems under uncertainty with positive cost. This issue concerns multiple and diverse fields such as engineering, economics, artificial intelligence, cognitive science and ma
Thompson sampling and other Bayesian sequential decision-making algorithms are among the most popular approaches to tackle explore/exploit trade-offs in (contextual) bandits. The choice of prior in these algorithms offers flexibility to encode domain
We propose SLTD (`Sequential Learning-to-Defer) a framework for learning-to-defer pre-emptively to an expert in sequential decision-making settings. SLTD measures the likelihood of improving value of deferring now versus later based on the underlying
Methods to find counterfactual explanations have predominantly focused on one step decision making processes. In this work, we initiate the development of methods to find counterfactual explanations for decision making processes in which multiple, de