Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
      More Filters
      Clear All
      More Filters
      Source
    • Language
1,590 result(s) for "stochastic linear optimization"
Sort by:
Robust Stochastic Optimization Made Easy with RSOME
We present a new distributionally robust optimization model called robust stochastic optimization (RSO), which unifies both scenario-tree-based stochastic linear optimization and distributionally robust optimization in a practicable framework that can be solved using the state-of-the-art commercial optimization solvers. We also develop a new algebraic modeling package, Robust Stochastic Optimization Made Easy (RSOME), to facilitate the implementation of RSO models. The model of uncertainty incorporates both discrete and continuous random variables, typically assumed in scenario-tree-based stochastic linear optimization and distributionally robust optimization, respectively. To address the nonanticipativity of recourse decisions, we introduce the event-wise recourse adaptations, which integrate the scenario-tree adaptation originating from stochastic linear optimization and the affine adaptation popularized in distributionally robust optimization. Our proposed event-wise ambiguity set is rich enough to capture traditional statistic-based ambiguity sets with convex generalized moments, mixture distribution, φ-divergence, Wasserstein (Kantorovich-Rubinstein) metric, and also inspire machine-learning-based ones using techniques such as K-means clustering and classification and regression trees. Several interesting RSO models, including optimizing over the Hurwicz criterion and two-stage problems over Wasserstein ambiguity sets, are provided. This paper was accepted by David Simchi-Levi, optimization.
Stochastic linear optimization under partial uncertainty and incomplete information using the notion of probability multimeasure
We consider a scalar stochastic linear optimization problem subject to linear constraints. We introduce the notion of deterministic equivalent formulation when the underlying probability space is equipped with a probability multimeasure. The initial problem is then transformed into a set-valued optimization problem with linear constraints. We also provide a method for estimating the expected value with respect to a probability multimeasure and prove extensions of the classical strong law of large numbers, the Glivenko-Cantelli theorem, and the central limit theorem to this setting. The notion of sampling with respect to a probability multimeasure and the definition of cumulative distribution multifunction are also discussed. Finally, we show some properties of the deterministic equivalent problem.
A scalable parallel implementation of the Cluster Benders Decomposition algorithm
Benders Decomposition (BD) is a method used to solve stochastic linear problems via scenario analysis. Cluster BD (CBD) is one of its smart improvements that speed up the execution time, taking advantage of tighter feasible cuts found by grouping scenarios into clusters. In this paper, we propose a new design for CBD, one which takes into account the role played by optimal cuts in the solution. Besides, we propose a new parallel scheme for CBD to deal with large-scale two-stage stochastic linear problems. Moreover, we characterise the problems for which our proposal performs best. The results obtained show computational gains from our proposal compared with the plain use of CPLEX, serial BD, parallel BD, serial CBD and parallel CBD.
Robust Optimization
Robust optimization is still a relatively new approach to optimization problems affected by uncertainty, but it has already proved so useful in real applications that it is difficult to tackle such problems today without considering this powerful methodology. Written by the principal developers of robust optimization, and describing the main achievements of a decade of research, this is the first book to provide a comprehensive and up-to-date account of the subject. Robust optimization is designed to meet some major challenges associated with uncertainty-affected optimization problems: to operate under lack of full information on the nature of uncertainty; to model the problem in a form that can be solved efficiently; and to provide guarantees about the performance of the solution. The book starts with a relatively simple treatment of uncertain linear programming, proceeding with a deep analysis of the interconnections between the construction of appropriate uncertainty sets and the classical chance constraints (probabilistic) approach. It then develops the robust optimization theory for uncertain conic quadratic and semidefinite optimization problems and dynamic (multistage) problems. The theory is supported by numerous examples and computational illustrations. An essential book for anyone working on optimization and decision making under uncertainty,Robust Optimizationalso makes an ideal graduate textbook on the subject.
Momentum and stochastic momentum for stochastic gradient, Newton, proximal point and subspace descent methods
In this paper we study several classes of stochastic optimization algorithms enriched with heavy ball momentum. Among the methods studied are: stochastic gradient descent, stochastic Newton, stochastic proximal point and stochastic dual subspace ascent. This is the first time momentum variants of several of these methods are studied. We choose to perform our analysis in a setting in which all of the above methods are equivalent: convex quadratic problems. We prove global non-asymptotic linear convergence rates for all methods and various measures of success, including primal function values, primal iterates, and dual function values. We also show that the primal iterates converge at an accelerated linear rate in a somewhat weaker sense. This is the first time a linear rate is shown for the stochastic heavy ball method (i.e., stochastic gradient descent method with momentum). Under somewhat weaker conditions, we establish a sublinear convergence rate for Cesàro averages of primal iterates. Moreover, we propose a novel concept, which we call stochastic momentum, aimed at decreasing the cost of performing the momentum step. We prove linear convergence of several stochastic methods with stochastic momentum, and show that in some sparse data regimes and for sufficiently small momentum parameters, these methods enjoy better overall complexity than methods with deterministic momentum. Finally, we perform extensive numerical testing on artificial and real datasets, including data coming from average consensus problems.
Stochastic multi-objective optimization: a survey on non-scalarizing methods
Currently, stochastic optimization on the one hand and multi-objective optimization on the other hand are rich and well-established special fields of Operations Research. Much less developed, however, is their intersection: the analysis of decision problems involving multiple objectives and stochastically represented uncertainty simultaneously. This is amazing, since in economic and managerial applications, the features of multiple decision criteria and uncertainty are very frequently co-occurring. Part of the existing quantitative approaches to deal with problems of this class apply scalarization techniques in order to reduce a given stochastic multi-objective problem to a stochastic single-objective one. The present article gives an overview over a second strand of the recent literature, namely methods that preserve the multi-objective nature of the problem during the computational analysis. We survey publications assuming a risk-neutral decision maker, but also articles addressing the situation where the decision maker is risk-averse. In the second case, modern risk measures play a prominent role, and generalizations of stochastic orders from the univariate to the multivariate case have recently turned out as a promising methodological tool. Modeling questions as well as issues of computational solution are discussed.
A bi-objective robust optimization model for a blood collection and testing problem: an accelerated stochastic Benders decomposition
Blood transfusion services are vital components of healthcare systems all over the world. In this paper, a generalized network optimization model is developed for a complex blood supply chain in accordance with Iranian blood transfusion organization (IBTO) structure. This structure consist of four types facilities. Blood collection centers, blood collection and processing centers, mobile teams and blood transfusion center have various duties in IBTO structure. The major contribution is to develop a novel hybrid approach based on stochastic programming, ε-constraint and robust optimization (HSERO) to simultaneously model two types of uncertainties by including stochastic scenarios for total blood donations and polyhedral uncertainty sets for demands. An accelerated stochastic Benders decomposition algorithm is proposed to solve the problem modeled in this paper. To speed up the convergence of the solution algorithm, valid inequalities are introduced to get better quality lower bounds. In addition, a Pareto-optimal cut generation scheme is used to strengthen the Benders optimality cuts. Numerical illustrations are given to verify the mathematical formulation and also to show the benefits of using the HSERO approach. At the end, the performance improvements achieved by the valid inequalities and the Pareto-optimal cuts are demonstrated in a real world application.
Adaptive Distributionally Robust Optimization
We develop a modular and tractable framework for solving an adaptive distributionally robust linear optimization problem, where we minimize the worst-case expected cost over an ambiguity set of probability distributions. The adaptive distributionally robust optimization framework caters for dynamic decision making, where decisions adapt to the uncertain outcomes as they unfold in stages. For tractability considerations, we focus on a class of second-order conic (SOC) representable ambiguity set, though our results can easily be extended to more general conic representations. We show that the adaptive distributionally robust linear optimization problem can be formulated as a classical robust optimization problem. To obtain a tractable formulation, we approximate the adaptive distributionally robust optimization problem using linear decision rule (LDR) techniques. More interestingly, by incorporating the primary and auxiliary random variables of the lifted ambiguity set in the LDR approximation, we can significantly improve the solutions, and for a class of adaptive distributionally robust optimization problems, exact solutions can also be obtained. Using the new LDR approximation, we can transform the distributionally adaptive robust optimization problem to a classical robust optimization problem with an SOC representable uncertainty set. Finally, to demonstrate the potential for solving management decision problems, we develop an algebraic modeling package and illustrate how it can be used to facilitate modeling and obtain high-quality solutions for medical appointment scheduling and inventory management problems. The electronic companion is available at https://doi.org/10.1287/mnsc.2017.2952 . This paper was accepted by Noah Gans, optimization.
Theory and Applications of Robust Optimization
In this paper we survey the primary research, both theoretical and applied, in the area of robust optimization (RO). Our focus is on the computational attractiveness of RO approaches, as well as the modeling power and broad applicability of the methodology. In addition to surveying prominent theoretical results of RO, we also present some recent results linking RO to adaptable models for multistage decision-making problems. Finally, we highlight applications of RO across a wide spectrum of domains, including finance, statistics, learning, and various areas of engineering.
Distributionally Robust Optimization and Its Tractable Approximations
In this paper we focus on a linear optimization problem with uncertainties, having expectations in the objective and in the set of constraints. We present a modular framework to obtain an approximate solution to the problem that is distributionally robust and more flexible than the standard technique of using linear rules. Our framework begins by first affinely extending the set of primitive uncertainties to generate new linear decision rules of larger dimensions and is therefore more flexible. Next, we develop new piecewise-linear decision rules that allow a more flexible reformulation of the original problem. The reformulated problem will generally contain terms with expectations on the positive parts of the recourse variables. Finally, we convert the uncertain linear program into a deterministic convex program by constructing distributionally robust bounds on these expectations. These bounds are constructed by first using different pieces of information on the distribution of the underlying uncertainties to develop separate bounds and next integrating them into a combined bound that is better than each of the individual bounds.