ترغب بنشر مسار تعليمي؟ اضغط هنا

Apportionment with Parity Constraints

69   0   0.0 ( 0 )
 نشر من قبل Victor Verdugo
 تاريخ النشر 2021
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

In the classic apportionment problem the goal is to decide how many seats of a parliament should be allocated to each party as a result of an election. The divisor methods provide a way of solving this problem by defining a notion of proportionality guided by some rounding rule. Motivated by recent challenges in the context of electoral apportionment, we consider the question of how to allocate the seats of a parliament under parity constraints between candidate types (e.g. equal number of men and women elected) while at the same time satisfying party proportionality. We consider two different approaches for this problem. The first mechanism, that follows a greedy approach, corresponds to a recent mechanism used in the Chilean Constitutional Convention 2021 election. We analyze this mechanism from a theoretical point of view. The second mechanism follows the idea of biproportionality introduced by Balinski and Demange [Math. Program. 1989, Math. Oper. Res. 1989]. In contrast with the classic biproportional method by Balinski and Demange, this mechanism is ruled by two levels of proportionality: Proportionality is satisfied at the level of parties by means of a divisor method, and then biproportionality is used to decide the number of candidates allocated to each type and party. We provide a theoretical analysis of this mechanism, making progress on the theoretical understanding of methods with two levels of proportionality. A typical benchmark used in the context of two-dimensional apportionment is the fair share (a.k.a matrix scaling), which corresponds to an ideal fractional biproportional solution. We provide lower bounds on the distance between these two types of solutions, and we explore their consequences in the context of two-dimensional apportionment.

قيم البحث

اقرأ أيضاً

We consider the implications of low-energy precision tests of parity violation on t-channel mediator models explaining the top AFB excess measured by CDF and D0. Flavor-violating u-t or d-t couplings of new scalar or vector mediators generate at one- loop an anomalous contribution to the nuclear weak charge. As a result, atomic parity violation constraints disfavor at >3 sigma t-channel models that give rise to a greater than 20% AFB at the parton level for M_tt > 450 GeV while not producing too large a top cross-section. Even stronger constraints are expected through future measurements of the proton weak charge by the Q-Weak experiment.
Ensemble Kalman methods constitute an increasingly important tool in both state and parameter estimation problems. Their popularity stems from the derivative-free nature of the methodology which may be readily applied when computer code is available for the underlying state-space dynamics (for state estimation) or for the parameter-to-observable map (for parameter estimation). There are many applications in which it is desirable to enforce prior information in the form of equality or inequality constraints on the state or parameter. This paper establishes a general framework for doing so, describing a widely applicable methodology, a theory which justifies the methodology, and a set of numerical experiments exemplifying it.
In this paper, we study a retailer price optimization problem which includes the practical constraints: maximum number of price changes and minimum amount of price change (if a change is recommended). We provide a closed-form formula for the Euclidea n projection onto the feasible set defined by these two constraints, based on which a simple gradient projection algorithm is proposed to solve the price optimization problem. We study the convergence and solution quality of the proposed algorithm. We extend the base model to include upper/lower bounds on the individual product prices and solve it with some adjustments to the gradient projection algorithm. Numerical results are reported to demonstrate the performance of the proposed algorithm.
The early sections of this paper present an analysis of a Markov decision model that is known as the multi-armed bandit under the assumption that the utility function of the decision maker is either linear or exponential. The analysis includes effici ent procedures for computing the expected utility associated with the use of a priority policy and for identifying a priority policy that is optimal. The methodology in these sections is novel, building on the use of elementary row operations. In the later sections of this paper, the analysis is adapted to accommodate constraints that link the bandits.
In this paper, we address a variant of the marketing mix optimization (MMO) problem which is commonly encountered in many industries, e.g., retail and consumer packaged goods (CPG) industries. This problem requires the spend for each marketing activi ty, if adjusted, be changed by a non-negligible degree (minimum change) and also the total number of activities with spend change be limited (maximum number of changes). With these two additional practical requirements, the original resource allocation problem is formulated as a mixed integer nonlinear program (MINLP). Given the size of a realistic problem in the industrial setting, the state-of-the-art integer programming solvers may not be able to solve the problem to optimality in a straightforward way within a reasonable amount of time. Hence, we propose a systematic reformulation to ease the computational burden. Computational tests show significant improvements in the solution process.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا