
Allende, H., Bravo, D., & Canessa, E. (2010). Robust design in multivariate systems using genetic algorithms. Qual. Quant., 44(2), 315–332.
Abstract: This paper presents a methodology based oil genetic algorithms, which finds feasible and reasonably adequate Solutions to problems of robust design in multivariate systems. We use a genetic algorithm to determine the appropriate control factor levels for simultaneously optimizing all of the responses of the system, considering the noise factors which affect it. The algorithm is guided by a desirability function which works with only one fitness function although the system May have many responses. We validated the methodology using data obtained from a real system and also from a process simulator, considering univariate and multivariate systems. In all cases, the methodology delivered feasible solutions, which accomplished the goals of robust design: obtain responses very close to the target values of each of them, and with minimum variability. Regarding the adjustment of the mean of each response to the target value, the algorithm performed very well. However, only in some of the multivariate cases, the algorithm was able to significantly reduce the variability of the responses.



AlvarezMiranda, E., & Pereira, J. (2017). Designing and constructing networks under uncertainty in the construction stage: Definition and exact algorithmic approach. Comput. Oper. Res., 81, 178–191.
Abstract: The present work proposes a novel Network Optimization problem whose core is to combine both network design and network construction scheduling under uncertainty into a single twostage robust optimization model. The firststage decisions correspond to those of a classical network design problem, while the secondstage decisions correspond to those of a network construction scheduling problem (NCS) under uncertainty. The resulting problem, which we will refer to as the TwoStage Robust Network Design and Construction Problem (2SRNDC), aims at providing a modeling framework in which the design decision not only depends on the design costs (e.g., distances) but also on the corresponding construction plan (e.g., time to provide service to costumers). We provide motivations, mixed integer programming formulations, and an exact algorithm for the 2SRNDC. Experimental results on a large set of instances show the effectiveness of the model in providing robust solutions, and the capability of the proposed algorithm to provide good solutions in reasonable running times. (C) 2017 Elsevier Ltd. All rights reserved.



Araujo, J., Ducoffe, G., Nisse, N., & Suchan, K. (2018). On interval number in cycle convexity. Discret. Math. Theor. Comput. Sci., 20(1), 35 pp.
Abstract: Recently, Araujo et al. [Manuscript in preparation, 2017] introduced the notion of Cycle Convexity of graphs. In their seminal work, they studied the graph convexity parameter called hull number for this new graph convexity they proposed, and they presented some of its applications in Knot theory. Roughly, the tunnel number of a knot embedded in a plane is upper bounded by the hull number of a corresponding planar 4regular graph in cycle convexity. In this paper, we go further in the study of this new graph convexity and we study the interval number of a graph in cycle convexity. This parameter is, alongside the hull number, one of the most studied parameters in the literature about graph convexities. Precisely, given a graph G, its interval number in cycle convexity, denoted by in(cc)(G), is the minimum cardinality of a set S subset of V (G) such that every vertex w is an element of E V (G) \ S has two distinct neighbors u, v is an element of S such that u and v lie in same connected component of G[S], i.e. the subgraph of G induced by the vertices in S. In this work, first we provide bounds on in(cc) (G) and its relations to other graph convexity parameters, and explore its behaviour on grids. Then, we present some hardness results by showing that deciding whetherin(cc) (G) <= k is NPcomplete, even if G is a split graph or a boundeddegree planar graph, and that the problem is W[2]hard in bipartite graphs when k is the parameter. As a consequence, we obtain that in(cc) (G) cannot be approximated up to a constant factor in the classes of split graphs and bipartite graphs (unless P = NP). On the positive side, we present polynomialtime algorithms to compute in(cc) (G) for outerplanar graphs, cobipartite graphs and interval graphs. We also present fixedparameter tractable (FPT) algorithms to compute it for (q, q – 4)graphs when q is the parameter and for general graphs G when parameterized either by the treewidth or the neighborhood diversity of G. Some of our hardness results and positive results are not known to hold for related graph convexities and domination problems. We hope that the design of our new reductions and polynomialtime algorithms can be helpful in order to advance in the study of related graph problems.



Barrera, J., Moreno, E., & Varas, S. (2020). A decomposition algorithm for computing income taxes with passthrough entities and its application to the Chilean case. Ann. Oper. Res., 286(12), 545–557.
Abstract: Income tax systems with “passthrough” entities transfer a firm's income to shareholders, which are taxed individually. In 2014, a Chilean tax reform introduced this type of entity and changed to an accrual basis that distributes incomes (but not losses) to shareholders. A crucial step for the Chilean taxation authority is to compute the final income of each individual given the complex network of corporations and companies, usually including cycles between them. In this paper, we show the mathematical conceptualization and the solution to the problem, proving that there is only one way to distribute income to taxpayers. Using the theory of absorbing Markov chains, we define a mathematical model for computing the taxable income of each taxpayer, and we propose a decomposition algorithm for this problem. This approach allows us to compute the solution accurately and to efficiently use computational resources. Finally, we present some characteristics of Chilean taxpayers' network and the computational results of the algorithm using this network.



BorquezParedes, D., Beghelli, A., Leiva, A., & Murrugarra, R. (2018). Does fragmentation avoidance improve the performance of dynamic spectrum allocation in elastic optical networks? Photonic Netw. Commun., 35(3), 287–299.
Abstract: Most spectrum allocation algorithms in elastic optical networks apply a greedy approach: A new connection is allocated as long as there are enough spectrum slots to accommodate it. Recently, a different approach was proposed. Named DeadlockAvoidance (DA), it only establishes a new connection if the portion of spectrum left after allocating it is zero (fulllink utilization) or is big enough to accommodate future requests. Otherwise, the connection request is blocked as a way to avoid fragmentation. The performance of DA has been evaluated in a singlelink scenario, where its performance is not affected by the slot continuity constraint. In this paper, we evaluate for the first time the blocking performance and fragmentation level of DA in a fully dynamic network scenario with different bitrates and number of slots for a single link, a 4node bus and a mesh topology. The performance was evaluated by simulation, and a lower bound was also derived using a continuous Markov chain model. Results are obtained for DA and three greedy algorithms: First Fit, Exact Fit and FirstLast Fit. Results show that DA significantly decreases fragmentation, and thus, it exhibits a much lower blocking due to fragmentation than the greedy algorithms. However, this decrease is compensated by a new type of blocking due to the selective acceptance of connections. As a result, the extra computational complexity of DA does not compensate a gain in performance.



Canessa, E., & Chaigneau, S. (2015). Calibrating AgentBased Models Using a Genetic Algorithm. Stud. Inform. Control, 24(1), 79–90.
Abstract: We present a Genetic Algorithm (GA)based tool that calibrates Agentbased Models (ABMs). The GA searches through a userdefined set of input parameters of an ABM, delivering values for those parameters so that the output time series of an ABM may match the real system's time series to certain precision. Once that set of possible values has been available, then a domain expert can select among them, the ones that better make sense from a practical point of view and match the explanation of the phenomenon under study. In developing the GA, we have had three main goals in mind. First, the GA should be easily used by nonexpert computer users and allow the seamless integration of the GA with different ABMs. Secondly, the GA should achieve a relatively short convergence time, so that it may be practical to apply it to many situations, even if the corresponding ABMs exhibit complex dynamics. Thirdly, the GA should use a few data points of the real system's time series and even so, achieve a sufficiently good match with the ABM's time series to attaining relational equivalence between the real system under study and the ABM that models it. That feature is important since social science longitudinal studies commonly use few data points. The results show that all of those goals have been accomplished.



Canessa, E., & Chaigneau, S. (2017). Response surface methodology for estimating missing values in a pareto genetic algorithm used in parameter design. Ing. Invest., 37(2), 89–98.
Abstract: We present an improved Pareto Genetic Algorithm (PGA), which finds solutions to problems of robust design in multiresponse systems with 4 responses and as many as 10 control and 5 noise factors. Because some response values might not have been obtained in the robust design experiment and are needed in the search process, the PGA uses Response Surface Methodology (RSM) to estimate them. Not only the PGA delivered solutions that adequately adjusted the response means to their target values, and with low variability, but also found more Pareto efficient solutions than a previous version of the PGA. This improvement makes it easier to find solutions that meet the tradeoff among variance reduction, mean adjustment and economic considerations. Furthermore, RSM allows estimating outputs' means and variances in highly nonlinear systems, making the new PGA appropriate for such systems.



Canessa, E., Droop, C., & Allende, H. (2012). An improved genetic algorithm for robust design in multivariate systems. Qual. Quant., 46(2), 665–678.
Abstract: In a previous article, we presented a genetic algorithm (GA), which finds solutions to problems of robust design in multivariate systems. Based on that GA, we developed a new GA that uses a new desirability function, based on the aggregation of the observed variance of the responses and the squared deviation between the mean of each response and its corresponding target value. Additionally, we also changed the crossover operator from a onepoint to a uniform one. We used three different case studies to evaluate the performance of the new GA and also to compare it with the original one. The first case study involved using data from a univariate real system, and the other two employed data obtained from multivariate process simulators. In each of the case studies, the new GA delivered good solutions, which simultaneously adjusted the mean of each response to its corresponding target value. This performance was similar to the one of the original GA. Regarding variability reduction, the new GA worked much better than the original one. In all the case studies, the new GA delivered solutions that simultaneously decreased the standard deviation of each response to almost the minimum possible value. Thus, we conclude that the new GA performs better than the original one, especially regarding variance reduction, which was the main problem exhibited by the original GA.



Canessa, E., Vera, S., & Allende, H. (2012). A new method for estimating missing values for a genetic algorithm used in robust design. Eng. Optimiz., 44(7), 787–800.
Abstract: This article presents an improved genetic algorithm (GA), which finds solutions to problems of robust design in multivariate systems with many control and noise factors. Since some values of responses of the system might not have been obtained from the robust design experiment, but may be needed in the search process, the GA uses response surface methodology (RSM) to estimate those values. In all test cases, the GA delivered solutions that adequately adjusted the mean of the responses to their corresponding target values and with low variability. The GA found more solutions than the previous versions of the GA, which makes it easier to find a solution that may meet the tradeoff among variance reduction, mean adjustment and economic considerations. Moreover, RSM is a good method for estimating the mean and variance of the outputs of highly nonlinear systems, which makes the new GA appropriate for optimizing such systems.



Carrasco, R. A., Iyengar, G., & Stein, C. (2018). Resource cost aware scheduling. Eur. J. Oper. Res., 269(2), 621–632.
Abstract: We are interested in the scheduling problem where there are several different resources that determine the speed at which a job runs and we pay depending on the amount of each resource that we use. This work is an extension of the resource dependent job processing time problem and the energy aware scheduling problems. We develop a new constant factor approximation algorithm for resource cost aware scheduling problems: the objective is to minimize the sum of the total cost of resources and the total weighted completion time in the one machine nonpreemptive setting, allowing for arbitrary precedence constraints and release dates. Our algorithm handles general jobdependent resource cost functions. We also analyze the practical performance of our algorithms, showing that it is significantly superior to the theoretical bounds and in fact it is very close to optimal. The analysis is done using simulations and real instances, which are left publicly available for future benchmarks. We also present additional heuristic improvements and we study their performance in other settings. (C) 2018 Elsevier B.V. All rights reserved.



Celis, P., de la Cruz, R., Fuentes, C., & Gomez, H. W. (2021). Survival and Reliability Analysis with an EpsilonPositive Family of Distributions with Applications. Symmetry, 13(5), 908.
Abstract: We introduce a new class of distributions called the epsilonpositive family, which can be viewed as generalization of the distributions with positive support. The construction of the epsilonpositive family is motivated by the ideas behind the generation of skew distributions using symmetric kernels. This new class of distributions has as special cases the exponential, Weibull, lognormal, loglogistic and gamma distributions, and it provides an alternative for analyzing reliability and survival data. An interesting feature of the epsilonpositive family is that it can viewed as a finite scale mixture of positive distributions, facilitating the derivation and implementation of EMtype algorithms to obtain maximum likelihood estimates (MLE) with (un)censored data. We illustrate the flexibility of this family to analyze censored and uncensored data using two real examples. One of them was previously discussed in the literature; the second one consists of a new application to model recidivism data of a group of inmates released from the Chilean prisons during 2007. The results show that this new family of distributions has a better performance fitting the data than some common alternatives such as the exponential distribution.



Dumett, M. A., & Cominetti, R. (2018). On The Stability Of An Adaptive Learning Dynamics In Traffic Games. J. Dyn. Games, 5(4), 265–282.
Abstract: This paper investigates the dynamic stability of an adaptive learning procedure in a traffic game. Using the RouthHurwitz criterion we study the stability of the rest points of the corresponding mean field dynamics. In the special case with two routes and two players we provide a full description of the number and nature of these rest points as well as the global asymptotic behavior of the dynamics. Depending on the parameters of the model, we find that there are either one, two or three equilibria and we show that in all cases the mean field trajectories converge towards a rest point for almost all initial conditions.



Fernandez, C., Valle, C., Saravia, F., & Allende, H. (2012). Behavior analysis of neural network ensemble algorithm on a virtual machine cluster. Neural Comput. Appl., 21(3), 535–542.
Abstract: Ensemble learning has gained considerable attention in different learning tasks including regression, classification, and clustering problems. One of the drawbacks of the ensemble is the high computational cost of training stages. Resampling local negative correlation (RLNC) is a technique that combines two wellknown methods to generate ensemble diversityresampling and error negative correlationand a finegrain parallel approach that allows us to achieve a satisfactory balance between accuracy and efficiency. In this paper, we introduce a structure of the virtual machine aimed to test diverse selection strategies of parameters in neural ensemble designs, such as RLNC. We assess the parallel performance of this approach on a virtual machine cluster based on the full virtualization paradigm, using speedup and efficiency as performance metrics, for different numbers of processors and training data sizes.



Feuilloley, L., Fraigniaud, P., Montealegre, P., Rapaport, I., Remila, E., & Todinca, I. (2021). Compact Distributed Certification of Planar Graphs. Algorithmica, Early Access.
Abstract: Naor M., Parter M., Yogev E.: (The power of distributed verifiers in interactive proofs. In: 31st ACMSIAM symposium on discrete algorithms (SODA), pp 1096115, 2020. https://doi.org/10.1137/1.9781611975994.67) have recently demonstrated the existence of a distributed interactive proof for planarity (i.e., for certifying that a network is planar), using a sophisticated generic technique for constructing distributed IP protocols based on sequential IP protocols. The interactive proof for planarity is based on a distributed certification of the correct execution of any given sequential lineartime algorithm for planarity testing. It involves three interactions between the prover and the randomized distributed verifier (i.e., it is a dMAM protocol), and uses small certificates, on O(log n) bits in nnode networks. We show that a single interaction with the prover suffices, and randomization is unecessary, by providing an explicit description of a prooflabeling scheme for planarity, still using certificates on just O(log n) bits. We also show that there are no prooflabeling schemesin fact, even no locally checkable proofsfor planarity using certificates on o(log n) bits.



Goles, E., & Montealegre, P. (2020). The complexity of the asynchronous prediction of the majority automata. Inf. Comput., 274(SI).
Abstract: We consider the asynchronous prediction problem for some automaton as the one consisting in determining, given an initial configuration, if there exists a nonzero probability that some selected site changes its state, when the network is updated picking one site at a time uniformly at random. We show that for the majority automaton, the asynchronous prediction problem is in NC in the twodimensional lattice with von Neumann neighborhood. Later, we show that in three or more dimensions the problem is NPComplete.



Goles, E., Maldonado, D., Montealegre, P., & Ollinger, N. (2020). On the complexity of the stability problem of binary freezing totalistic cellular automata. Inf. Comput., 274, 21 pp.
Abstract: In this paper we study the family of twostate Totalistic Freezing Cellular Automata (TFCA) defined over the triangular and square grids with von Neumann neighborhoods. We say that a Cellular Automaton is Freezing and Totalistic if the active cells remain unchanged, and the new value of an inactive cell depends only on the sum of its active neighbors. We classify all the Cellular Automata in the class of TFCA, grouping them in five different classes: the Trivial rules, Turing Universal rules, Algebraic rules, Topological rules and Fractal Growing rules. At the same time, we study in this family the STABILITY problem, consisting in deciding whether an inactive cell becomes active, given an initial configuration. We exploit the properties of the automata in each group to show that: For Algebraic and Topological Rules the STABILITY problem is in NC. For Turing Universal rules the STABILITY problem is PComplete. (C) 2020 Elsevier Inc. All rights reserved.



Henriquez, P. A., & Ruz, G. A. (2018). A noniterative method for pruning hidden neurons in neural networks with random weights. Appl. Soft. Comput., 70, 1109–1121.
Abstract: Neural networks with random weights have the advantage of fast computational time in both training and testing. However, one of the main challenges of single layer feedforward neural networks is the selection of the optimal number of neurons in the hidden layer, since few/many neurons lead to problems of underfitting/overfitting. Adapting Garson's algorithm, this paper introduces a new efficient and fast noniterative algorithm for the selection of neurons in the hidden layer for randomization based neural networks. The proposed approach is divided into three steps: (1) train the network with h hidden neurons, (2) apply Garson's algorithm to the matrix of the hidden layer, and (3) perform pruning reducing hidden layer neurons based on the harmonic mean. Our experiments in regression and classification problems confirmed that the combination of the pruning technique with these types of neural networks improved their predictive performance in terms of mean square error and accuracy. Additionally, we tested our proposed pruning method with neural networks trained under sequential learning algorithms, where Random Vector Functional Link obtained, in general, the best predictive performance compared to online sequential versions of extreme learning machines and single hidden layer neural network with random weights. (C) 2018 Elsevier B.V. All rights reserved.



Jenkins, J. S., Diaz, M. R., Kurtovic, N. T., Espinoza, N., Vines, J. I., Rojas, P. A. P., et al. (2020). An ultrahot Neptune in the Neptune desert. Nat. Astron., 4(12), 1148–1157.
Abstract: About 1 out of 200 Sunlike stars has a planet with an orbital period shorter than one day: an ultrashortperiod planet(1,2). All of the previously known ultrashortperiod planets are either hot Jupiters, with sizes above 10 Earth radii (Rcircle plus), or apparently rocky planets smaller than 2 Rcircle plus. Such lack of planets of intermediate size (the `hot Neptune desert') has been interpreted as the inability of lowmass planets to retain any hydrogen/ helium (H/He) envelope in the face of strong stellar irradiation. Here we report the discovery of an ultrashortperiod planet with a radius of 4.6 Rcircle plus and a mass of 29 Mcircle plus, firmly in the hot Neptune desert. Data from the Transiting Exoplanet Survey Satellite(3) revealed transits of the bright Sunlike star LTT 9779 every 0.79 days. The planet's mean density is similar to that of Neptune, and according to thermal evolution models, it has a H/Herich envelope constituting 9.0(2.9)(+2.7) % of the total mass. With an equilibrium temperature around 2,000 K, it is unclear how this `ultrahot Neptune' managed to retain such an envelope. Followup observations of the planet's atmosphere to better understand its origin and physical nature will be facilitated by the star's brightness (Vmag = 9.8).



Lagos, T., Armstrong, M., HomemdeMello, T., Lagos, G., & Saure, D. (2021). A framework for adaptive openpit mining planning under geological uncertainty. Optim. Eng., Early Access, 36 pp.
Abstract: Mine planning optimization aims at maximizing the profit obtained from extracting valuable ore. Beyond its theoretical complexitythe openpit mining problem with capacity constraints reduces to a knapsack problem with precedence constraints, which is NPhardpractical instances of the problem usually involve a large to very large number of decision variables, typically of the order of millions for large mines. Additionally, any comprehensive approach to mine planning ought to consider the underlying geostatistical uncertainty as only limited information obtained from drill hole samples of the mineral is initially available. In this regard, as blocks are extracted sequentially, information about the ore grades of blocks yet to be extracted changes based on the blocks that have already been mined. Thus, the problem lies in the class of multiperiod large scale stochastic optimization problems with decisiondependent information uncertainty. Such problems are exceedingly hard to solve, so approximations are required. This paper presents an adaptive optimization scheme for multiperiod production scheduling in openpit mining under geological uncertainty that allows us to solve practical instances of the problem. Our approach is based on a rollinghorizon adaptive optimization framework that learns from new information that becomes available as blocks are mined. By considering the evolution of geostatistical uncertainty, the proposed optimization framework produces an operational policy that reduces the risk of the production schedule. Our numerical tests with mines of moderate sizes show that our rolling horizon adaptive policy gives consistently better results than a nonadaptive stochastic optimization formulation, for a range of realistic problem instances.



Leiva, V., Saulo, H., Leao, J., & Marchant, C. (2014). A family of autoregressive conditional duration models applied to financial data. Comput. Stat. Data Anal., 79, 175–191.
Abstract: The BirnbaumSaunders distribution is receiving considerable attention due to its good properties. One of its extensions is the class of scalemixture BirnbaumSaunders (SBS) distributions, which shares its good properties, but it also has further properties. The autoregressive conditional duration models are the primary family used for analyzing highfrequency financial data. We propose a methodology based on SBS autoregressive conditional duration models, which includes insample inference, goodnessoffit and outofsample forecast techniques. We carry out a Monte Carlo study to evaluate its performance and assess its practical usefulness with realworld data of financial transactions from the New York stock exchange. (C) 2014 Elsevier B.V. All rights reserved.

