Home | << 1 2 >> |
![]() |
Coudert, D., Luedtke, J., Moreno, E., & Priftis, K. (2018). Computing and maximizing the exact reliability of wireless backhaul networks. In Electronic Notes in Discrete Mathematics (Vol. 64, pp. 85–94). |
Ramirez-Pico, C., & Moreno, E. (2021). Generalized Adaptive Partition-based Method for Two-Stage Stochastic Linear Programs with Fixed Recourse. Math. Program., to appear.
Abstract: We present a method to solve two-stage stochastic linear programming problems with fixed recourse when the uncertainty space can have either discrete or continuous distributions. Given a partition of the uncertainty space, the method is addressed to solve a discrete problem with one scenario for each element of the partition (subregions of the uncertainty space). Fixing first-stage variables, we formulate a second-stage subproblem for each element, and exploiting information from the dual of these problems, we provide conditions that the partition must satisfy to obtain an optimal solution. These conditions provide guidance on how to refine the partition, iteratively approaching an optimal solution. The results from computational experiments show how the method automatically refines the partition of the uncertainty space in the regions of interest for the problem. Our algorithm is a generalization of the adaptive partition-based method presented by Song & Luedtke for discrete distributions, extending its applicability to more general cases.
|
Letelier, O. R., Espinoza, D., Goycoolea, M., Moreno, E., & Munoz, G. (2020). Production Scheduling for Strategic Open Pit Mine Planning: A Mixed-Integer Programming Approach. Oper. Res., 68(5), 1425–1444.
Abstract: Given a discretized representation of an ore body known as a block model, the open pit mining production scheduling problem that we consider consists of defining which blocks to extract, when to extract them, and how or whether to process them, in such a way as to comply with operational constraints and maximize net present value. Although it has been established that this problem can be modeled with mixed-integer programming, the number of blocks used to represent real-world mines (millions) has made solving large instances nearly impossible in practice. In this article, we introduce a new methodology for tackling this problem and conduct computational tests using real problem sets ranging in size from 20,000 to 5,000,000 blocks and spanning 20 to 50 time periods. We consider both direct block scheduling and bench-phase scheduling problems, with capacity, blending, and minimum production constraints. Using new preprocessing and cutting planes techniques, we are able to reduce the linear programming relaxation value by up to 33%, depending on the instance. Then, using new heuristics, we are able to compute feasible solutions with an average gap of 1.52% relative to the previously computed bound. Moreover, after four hours of running a customized branch-and-bound algorithm on the problems with larger gaps, we are able to further reduce the average from 1.52% to 0.71%.
|
Canessa, G., Moreno, E., & Pagnoncelli, B. K. (2021). The risk-averse ultimate pit problem. Optim. Eng., Early access, 24 pp.
Abstract: In this work, we consider a risk-averse ultimate pit problem where the grade of the mineral is uncertain. We derive conditions under which we can generate a set of nested pits by varying the risk level instead of using revenue factors. We propose two properties that we believe are desirable for the problem: risk nestedness, which means the pits generated for different risk aversion levels should be contained in one another, and additive consistency, which states that preferences in terms of order of extraction should not change if independent sectors of the mine are added as precedences. We show that only an entropic risk measure satisfies these properties and propose a two-stage stochastic programming formulation of the problem, including an efficient approximation scheme to solve it. We illustrate our approach in a small self-constructed example, and apply our approximation scheme to a real-world section of the Andina mine, in Chile.
Keywords: Ultimate pit; Mining; Risk-averse optimization; Integer programming
|
Barrera, J., Carrasco, R. A., & Moreno, E. (2020). Real-time fleet management decision support system with security constraints. TOP, 28(3), 728–748.
Abstract: Intelligent transportation, and in particular, fleet management, has been a forefront concern for a plethora of industries. This statement is especially true for the production of commodities, where transportation represents a central element for operational continuity. Additionally, in many industries, and in particular those with hazardous environments, fleet control must satisfy a wide range of security restrictions to ensure that risks are kept at bay and accidents are minimum. Furthermore, in these environments, any decision support tool must cope with noisy and incomplete data and give recommendations every few minutes. In this work, a fast and efficient decision support tool is presented to help fleet managers oversee and control ore trucks, in a mining setting. The main objective of this system is to help managers avoid interactions between ore trucks and personnel buses, one of the most critical security constraints in our case study, keeping a minimum security distance between the two at all times. Furthermore, additional algorithms are developed and implemented, so that this approach can work with real-life noisy GPS data. Through the use of historical data, the performance of this decision support system is studied, validating that it works under the real-life conditions presented by the company. The experimental results show that the proposed approach improved truck and road utilization significantly while allowing the fleet manager to control the security distance required by their procedures.
|
Rezakhah, M., Moreno, E., & Newman, A. (2020). Practical performance of an open pit mine scheduling model considering blending and stockpiling. Comput. Oper. Res., 115, 12 pp.
Abstract: Open pit mine production scheduling (OPMPS) is a decision problem which seeks to maximize net present value (NPV) by determining the extraction time of each block of ore and/or waste in a deposit and the destination to which this block is sent, e.g., a processing plant or waste dump. Spatial precedence constraints are imposed, as are resource capacities. Stockpiles can be used to maintain low-grade ore for future processing, to store extracted material until processing capacity is available, and/or to blend material based on single or multiple block characteristics (i.e., metal grade and/or contaminant). We adapt an existing integer-linear program to an operational polymetallic (gold and copper) open pit mine, in which the stockpile is used to blend materials based on multiple block characteristics, and call it ((P) over cap (la)). We observe that the linear programming relaxation of our objective function is unimodal for different grade combinations (metals and contaminants) in the stockpile, which allows us to search systematically for an optimal grade combination while exploiting the linear structure of our optimization model. We compare the schedule of ((P) over cap (la)) with that produced by (P-ns) which does not consider stockpiling, and with ((P) over tilde (la)), which controls only the metal content in the stockpile and ignores the contaminant level at the mill and in the stockpile. Our proposed solution technique provides schedules for large instances in a few seconds up to a few minutes with significantly different stockpiling and material flow strategies depending on the model. We show that our model improves the NPV of the project while satisfying operational constraints. (C) 2019 Elsevier Ltd. All rights reserved.
|
Barrera, J., Moreno, E., & Varas, S. (2020). A decomposition algorithm for computing income taxes with pass-through entities and its application to the Chilean case. Ann. Oper. Res., 286(1-2), 545–557.
Abstract: Income tax systems with “pass-through” entities transfer a firm's income to shareholders, which are taxed individually. In 2014, a Chilean tax reform introduced this type of entity and changed to an accrual basis that distributes incomes (but not losses) to shareholders. A crucial step for the Chilean taxation authority is to compute the final income of each individual given the complex network of corporations and companies, usually including cycles between them. In this paper, we show the mathematical conceptualization and the solution to the problem, proving that there is only one way to distribute income to taxpayers. Using the theory of absorbing Markov chains, we define a mathematical model for computing the taxable income of each taxpayer, and we propose a decomposition algorithm for this problem. This approach allows us to compute the solution accurately and to efficiently use computational resources. Finally, we present some characteristics of Chilean taxpayers' network and the computational results of the algorithm using this network.
Keywords: Income taxes; Markov processes; Networks; Algorithms
|
Rivera Letelier, O., Espinoza, D., Goycoolea, M., Moreno, E., & Munoz, G. (2020). Production scheduling for strategic open pit mine planning: A mixed integer programming approach. Oper. Res., 68(5), 1425–1444.
Abstract: Given a discretized representation of an ore body known as a block model, the open pit mining production scheduling problem that we consider consists of defining which blocks to extract, when to extract them, and how or whether to process them, in such a way as to comply with operational constraints and maximize net present value. Although it has been established that this problem can be modeled with mixed-integer programming, the number of blocks used to represent real-world mines (millions) has made solving large instances nearly impossible in practice. In this article, we introduce a new methodology for tackling this problem and conduct computational tests using real problem sets ranging in size from 20,000 to 5,000,000 blocks and spanning 20 to 50 time periods. We consider both direct block scheduling and bench-phase scheduling problems, with capacity, blending, and minimum production constraints. Using new preprocessing and cutting planes techniques, we are able to reduce the linear programming relaxation value by up to 33\%, depending on the instance. Then, using new heuristics, we are able to compute feasible solutions with an average gap of 1.52% relative to the previously computed bound. Moreover, after four hours of running a customized branch-and-bound algorithm on the problems with larger gaps, we are able to further reduce the average from 1.52% to 0.71%
|
Matus, O., Barrera, J., Moreno, E., & Rubino, G. (2019). On the Marshall-Olkin Copula Model for Network Reliability Under Dependent Failures. IEEE Trans. Reliab., 68(2), 451–461.
Abstract: The Marshall-Olkin (MO) copulamodel has emerged as the standard tool for capturing dependence between components in failure analysis in reliability. In this model, shocks arise at exponential random times, that affect one or several components inducing a natural correlation in the failure process. However, because the number of parameter of the model grows exponentially with the number of components, MO suffers of the “curse of dimensionality.” MO models are usually intended to be applied to design a network before its construction; therefore, it is natural to assume that only partial information about failure behavior can be gathered, mostly from similar existing networks. To construct such an MO model, we propose an optimization approach to define the shock's parameters in the MO copula, in order to match marginal failures probabilities and correlations between these failures. To deal with the exponential number of parameters of this problem, we use a column-generation technique. We also discuss additional criteria that can be incorporated to obtain a suitable model. Our computational experiments show that the resulting MO model produces a close estimation of the network reliability, especially when the correlation between component failures is significant.
|
Munoz, G., Espinoza, D., Goycoolea, M., Moreno, E., Queyranne, M., & Rivera Letelier, O. (2018). A study of the Bienstock-Zuckerberg algorithm: applications in mining and resource constrained project scheduling. Comput. Optim. Appl., 69(2), 501–534.
Abstract: We study a Lagrangian decomposition algorithm recently proposed by Dan Bienstock and Mark Zuckerberg for solving the LP relaxation of a class of open pit mine project scheduling problems. In this study we show that the Bienstock-Zuckerberg (BZ) algorithm can be used to solve LP relaxations corresponding to a much broader class of scheduling problems, including the well-known Resource Constrained Project Scheduling Problem (RCPSP), and multi-modal variants of the RCPSP that consider batch processing of jobs. We present a new, intuitive proof of correctness for the BZ algorithm that works by casting the BZ algorithm as a column generation algorithm. This analysis allows us to draw parallels with the well-known Dantzig-Wolfe decomposition (DW) algorithm. We discuss practical computational techniques for speeding up the performance of the BZ and DW algorithms on project scheduling problems. Finally, we present computational experiments independently testing the effectiveness of the BZ and DW algorithms on different sets of publicly available test instances. Our computational experiments confirm that the BZ algorithm significantly outperforms the DW algorithm for the problems considered. Our computational experiments also show that the proposed speed-up techniques can have a significant impact on the solve time. We provide some insights on what might be explaining this significant difference in performance.
Keywords: Column generation; Dantzig-Wolfe; Optimization; RCPSP
|
Ljubic, I., & Moreno, E. (2018). Outer approximation and submodular cuts for maximum capture facility location problems with random utilities. Eur. J. Oper. Res., 266(1), 46–56.
Abstract: We consider a family of competitive facility location problems in which a “newcomer” company enters the market and has to decide where to locate a set of new facilities so as to maximize its market share. The multinomial logit model is used to estimate the captured customer demand. We propose a first branch-and-cut approach for this family of difficult mixed-integer non-linear problems. Our approach combines two types of cutting planes that exploit particular properties of the objective function: the first one are the outer-approximation cuts and the second one are the submodular cuts. The approach is computationally evaluated on three datasets from the recent literature. The obtained results show that our new branch-and-cut drastically outperforms state-of-the-art exact approaches, both in terms of the computing times, and in terms of the number of instances solved to optimality. (C) 2017 Elsevier B.V. All rights reserved.
|
Moreno, E., Rezakhah, M., Newman, A., & Ferreira, F. (2017). Linear models for stockpiling in open-pit mine production scheduling problems. Eur. J. Oper. Res., 260(1), 212–221.
Abstract: The open pit mine production scheduling (OPMPS) problem seeks to determine when, if ever, to extract each notional, three-dimensional block of ore and/or waste in a deposit and what to do with each, e.g., send it to a particular processing plant or to the waste dump. This scheduling model maximizes net present value subject to spatial precedence constraints, and resource capacities. Certain mines use stockpiles for blending different grades of extracted material, storing excess until processing capacity is available, or keeping low-grade ore for possible future processing. Common models assume that material in these stockpiles, or “buckets,” is theoretically immediately mixed and becomes homogeneous. We consider stockpiles as part of our open pit mine scheduling strategy, propose multiple models to solve the OPMPS problem, and compare the solution quality and tractability of these linear-integer and nonlinear-integer models. Numerical experiments show that our proposed models are tractable, and correspond to instances which can be solved in a few seconds up to a few minutes in contrast to previous nonlinear models that fail to solve. (C) 2016 Elsevier B.V. All rights reserved.
|
Moreno, E., Beghelli, A., & Cugini, F. (2017). Traffic engineering in segment routing networks. Comput. Netw., 114, 23–31.
Abstract: Segment routing (SR) has been recently proposed as an alternative traffic engineering (TE) technology enabling relevant simplifications in control plane operations. In the literature, preliminary investigations on SR have focused on label encoding algorithms and experimental assessments, without carefully addressing some key aspects of SR in terms of the overall network TE performance. In this study, ILP models and heuristics are proposed and successfully utilized to assess the TE performance of SR-based packet networks. Results show that the default SR behavior of exploiting equal cost multiple paths (ECMP) may lead to several drawbacks, including higher network resource utilization with respect to cases where ECMP is avoided. Moreover, results show that, by properly performing segment list computations, it is possible to achieve very effective TE solutions by just using a very limited number of stacked labels, thus successfully exploiting the benefits of the SR technology. (C) 2017 Elsevier B.V. All rights reserved.
Keywords: Segment routing; Integer Linear Programming; Heuristic
|
Tarifeno-Gajardo, M., Beghelli, A., & Moreno, E. (2016). Availability-Driven Optimal Design of Shared Path Protection in WDM Networks. Networks, 68(3), 224–237.
Abstract: Availability, defined as the fraction of time a network service is operative, is a key network service parameter. Dedicated protection increases availability but also the cost. Shared protection instead decreases the cost, but also the availability. In this article, we formulate and solve an integer linear programming (ILP) model for the problem of minimizing the backup resources required by a shared-protected static optical network whilst guaranteeing an availability target per connection. The main research challenge is dealing with the nonlinear expression for the availability constraint. Taking the working/backup routes and the availability requirements as input data, the ILP model identifies the set of connections sharing backup resources in any given network link. We also propose a greedy heuristic to solve large instances in much shorter time than the ILP model with low levels of relative error (2.49% average error in the instances studied) and modify the ILP model to evaluate the impact of wavelength conversion. Results show that considering availability requirements can lead up to 56.4% higher backup resource requirements than not considering them at all, highlighting the importance of availability requirements in budget estimation. (C) 2016 Wiley Periodicals, Inc.
|
Pineda, C., Cortes, C. E., Jara-Moroni, P., & Moreno, E. (2016). Integrated traffic-transit stochastic equilibrium model with park-and-ride facilitiesd. Transp. Res. Pt. C-Emerg. Technol., 71, 86–107.
Abstract: We propose an Integrated Stochastic Equilibrium model that considers both private automobile traffic and transit networks to incorporate the interactions between these two modes in terms of travel time and generalized costs. In addition, in the general version of the model, travelers are allowed to switch from personal vehicles to mass transit at specific locations in a park-and-ride scheme. The assignment for traffic equilibrium is based on the Markovian Traffic Equilibrium model of Baillon and Cominetti (2008), whereas the equilibrium of the transit network is represented by the Stochastic Transit Equilibrium model of Cortes et al. (2013). Stochastic travel decisions are made at the node level, thereby avoiding the enumeration of routes or strategies and incorporating various perception and uncertainty issues. We propose a Method-of-Successive-Averages algorithm to calculate an Integrated Stochastic Equilibrium and conduct numerical experiments to highlight the effect of stochasticity on equilibrium flows and travel times. Our experiments show that higher stochasticity implies greater dispersion of equilibrium flows and longer expected travel times. Results on a real network with mode combination and park and ride facilities provide insights regarding the use of park and ride in terms of number and location, potential modal share of the combined mode option under different circumstances, and travel time impact due to the implementation of such park and ride facilities in a real setting. (C) 2016 Elsevier Ltd. All rights reserved.
|
Barrera, J., Homem-De-Mello, T., Moreno, E., Pagnoncelli, B. K., & Canessa, G. (2016). Chance-constrained problems and rare events: an importance sampling approach. Math. Program., 157(1), 153–189.
Abstract: We study chance-constrained problems in which the constraints involve the probability of a rare event. We discuss the relevance of such problems and show that the existing sampling-based algorithms cannot be applied directly in this case, since they require an impractical number of samples to yield reasonable solutions. We argue that importance sampling (IS) techniques, combined with a Sample Average Approximation (SAA) approach, can be effectively used in such situations, provided that variance can be reduced uniformly with respect to the decision variables. We give sufficient conditions to obtain such uniform variance reduction, and prove asymptotic convergence of the combined SAA-IS approach. As it often happens with IS techniques, the practical performance of the proposed approach relies on exploiting the structure of the problem under study; in our case, we work with a telecommunications problem with Bernoulli input distributions, and show how variance can be reduced uniformly over a suitable approximation of the feasibility set by choosing proper parameters for the IS distributions. Although some of the results are specific to this problem, we are able to draw general insights that can be useful for other classes of problems. We present numerical results to illustrate our findings.
|
Freire, A. S., Moreno, E., & Yushimito, W. F. (2016). A branch-and-bound algorithm for the maximum capture problem with random utilities. Eur. J. Oper. Res., 252(1), 204–212.
Abstract: The MAXIMUM CAPTURE PROBLEM WITH RANDOM UTILITIES seeks to locate new facilities in a competitive market such that the captured demand of users is maximized, assuming that each individual chooses among all available facilities according to the well-know a random utility model namely the multinomial logit. The problem is complex mostly due to its integer nonlinear objective function. Currently, the most efficient approaches deal with this complexity by either using a nonlinear programing solver or reformulating the problem into a Mixed-Integer Linear Programing (MILP) model. In this paper, we show how the best MILP reformulation available in the literature can be strengthened by using tighter coefficients in some inequalities. We also introduce a new branch-and-bound algorithm based on a greedy approach for solving a relaxation of the original problem. Extensive computational experiments are presented, bench marking the proposed approach with other linear and non-linear relaxations of the problem. The computational experiments show that our proposed algorithm is competitive with all other methods as there is no method which outperforms the others in all instances. We also show a large-scale real instance of the problem, which comes from an application in park-and-ride facility location, where our proposed branch-and-bound algorithm was the most effective method for solving this type of problem. (C) 2015 Elsevier B.V. All rights reserved.
|
Espinoza, D., Goycoolea, M., & Moreno, E. (2015). The precedence constrained knapsack problem: Separating maximally violated inequalities. Discret Appl. Math., 194, 65–80.
Abstract: We consider the problem of separating maximally violated inequalities for the precedence constrained knapsack problem. Though we consider maximally violated constraints in a very general way, special emphasis is placed on induced cover inequalities and induced clique inequalities. Our contributions include a new partial characterization of maximally violated inequalities, a new safe shrinking technique, and new insights on strengthening and lifting. This work follows on the work of Boyd (1993), Park and Park (1997), van de Leensel et al. (1999) and Boland et al. (2011). Computational experiments show that our new techniques and insights can be used to significantly improve the performance of cutting plane algorithms for this problem. (C) 2015 Elsevier B.V. All rights reserved.
|
Barrera, J., Cancela, H., & Moreno, E. (2015). Topological optimization of reliable networks under dependent failures. Oper. Res. Lett., 43(2), 132–136.
Abstract: We address the design problem of a reliable network. Previous work assumes that link failures are independent. We discuss the impact of dropping this assumption. We show that under a common-cause failure model, dependencies between failures can affect the optimal design. We also provide an integer-programming formulation to solve this problem. Furthermore, we discuss how the dependence between the links that participate in the solution and those that do not can be handled. Other dependency models are discussed as well. (C) 2014 Elsevier B.V. All rights reserved.
|
Lagos, G., Espinoza, D., Moreno, E., & Vielma, J. P. (2015). Restricted risk measures and robust optimization. Eur. J. Oper. Res., 241(3), 771–782.
Abstract: In this paper we consider characterizations of the robust uncertainty sets associated with coherent and distortion risk measures. In this context we show that if we are willing to enforce the coherent or distortion axioms only on random variables that are affine or linear functions of the vector of random parameters, we may consider some new variants of the uncertainty sets determined by the classical characterizations. We also show that in the finite probability case these variants are simple transformations of the classical sets. Finally we present results of computational experiments that suggest that the risk measures associated with these new uncertainty sets can help mitigate estimation errors of the Conditional Value-at-Risk. (C) 2014 Elsevier B.V. All rights reserved.
|