3260 papers • 126 benchmarks • 313 datasets
Combinatorial Optimization is a category of problems which requires optimizing a function over a combination of discrete objects and the solutions are constrained. Examples include finding shortest paths in a graph, maximizing value in the Knapsack problem and finding boolean settings that satisfy a set of constraints. Many of these problems are NP-Hard, which means that no polynomial time solution can be developed for them. Instead, we can only produce approximations in polynomial time that are guaranteed to be some factor worse than the true optimal solution. Source: Recent Advances in Neural Program Synthesis
(Image credit: Papersgraph)
These leaderboards are used to track progress in combinatorial-optimization-8
No benchmarks available.
Use these libraries to find combinatorial-optimization-8 models and implementations
No subtasks available.
A new neural architecture to learn the conditional probability of an output sequence with elements that are discrete tokens corresponding to positions in an input sequence using a recently proposed mechanism of neural attention, called Ptr-Nets, which improves over sequence-to-sequence with input attention, but also allows it to generalize to variable size output dictionaries.
This work introduces Autodidactic Iteration: a novel reinforcement learning algorithm that is able to teach itself how to solve the Rubik's Cube with no human assistance.
A framework to tackle combinatorial optimization problems using neural networks and reinforcement learning, and Neural Combinatorial Optimization achieves close to optimal results on 2D Euclidean graphs with up to 100 nodes.
This paper proposes a unique combination of reinforcement learning and graph embedding that behaves like a meta-algorithm that incrementally constructs a solution, and the action is determined by the output of agraph embedding network capturing the current state of the solution.
A new graph convolutional neural network model is proposed for learning branch-and-bound variable selection policies, which leverages the natural variable-constraint bipartite graph representation of mixed-integer linear programs.
This work proposes an end-to-end trainable architecture for deep graph matching that contains unmodified combinatorial solvers and highlights the conceptual advantages of incorporating solvers into deep learning architectures.
Memory Augmented Policy Optimization is presented, a simple and novel way to leverage a memory buffer of promising trajectories to reduce the variance of policy gradient estimate and improves the sample efficiency and robustness of Policy gradient, especially on tasks with sparse rewards.
This work presents an end-to-end framework for solving the Vehicle Routing Problem (VRP) using reinforcement learning, and demonstrates how this approach can handle problems with split delivery and explore the effect of such deliveries on the solution quality.
This paper introduces a probabilistic perspective on CO layers, which lends itself naturally to approximate differentiation and the construction of structured losses, and presents InferOpt.jl, an open-source Julia package that allows turning any CO oracle with a linear objective into a differentiable layer, and defines adequate losses to train pipelines containing such layers.
Adding a benchmark result helps the community track progress.