15h30 - 15h55
Combining Neural Networks and Marginal Probabilities Derived from Constraint Programming Models to Perform Structured Prediction Tasks
Despite their many successes on various challenging tasks, Deep Neural Networks still struggle on structured prediction problems, which feature multiple discrete outputs that are interdependent, i.e. between which constraints should hold. Constraint Programming (CP) is all about structure: it enjoys a long and successful history of identifying often-recurring combinatorial structures and of designing sophisticated algorithms to extract information from such structures. Of particular interest here is the relative frequency of a given variable-value assignment in a combinatorial structure over a set of variables. We revisit two recent papers from the CP community that propose ways to combine Machine Learning (ML) and Constraint Programming during training or inference. We propose instead to use a CP-BP solver, which generalizes CP by propagating the above relative frequencies in order to approximate the marginal probabilities over each of the variables. We show that it offers more natural opportunities for combinations with ML and provide empirical evidence that it also boosts performance.
15h55 - 16h20
Dynamic routing and wavelength assignment using graph neural networks and deep reinforcement learning
We propose a novel method based on Deep Reinforcement Learning for the routing and wavelength assignment problem in all-optical wavelength-decision-multiplexing networks. We consider dynamic incoming requests, in which their arrival and holding times are not known in advance. The objective is to devise a strategy that minimizes the number of rejected packages due to the lack of resources in the long term. We employ Graph Neural Networks to capture promising information from a graph-structured input. A deep reinforcement learning agent is then trained to find the optimal strategy. The proposed algorithm selects a route and a wavelength simultaneously for each incoming traffic connection as they arrive. We show that this algorithm outperforms the baselines to solve the dynamic RWA problem.
16h20 - 16h45
A learning approach for constraint customization in optimization models
This study focuses on embedding linear side constraints in some known classical mixed integer linear programs (MILP). Industrial customers use optimization software packages to optimize their planning and decision making. However, Normally, in real applications, they customize the optimal solutions obtained by softwares based on some implicit internal operational rules. To access to a one-step decision support system and to avoid relying on experts’ knowledge of each particular domain, we propose a framework to find such rules from the historical data implemented by customers. We use some machine learning methods like clustering and decision tree to infer the rules and embed them as constraints in the optimization models
16h45 - 17h10
Reinforcement Learning for Ride-Hailing: Dynamic Vehicle Fleet Management
We study the use of reinforcement learning for ride-hailing with centralized control. The problem consists in managing a fleet of vehicles that have to be assigned to arising demands and relocated dynamically in order to maximize overall profit. Traditionally tackled by heuristics and mathematical optimization, this problem can be addressed by using machine learning to forecast demand ; reinforcement learning allows to learn from previous assignment and relocation decisions to improve the fleet behavior. Results are presented for the pyhailing ride-hailing simulator.