3260 papers • 126 benchmarks • 313 datasets
Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters. Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks
(Image credit: Papersgraph)
These leaderboards are used to track progress in network-pruning-9
Use these libraries to find network-pruning-9 models and implementations
No subtasks available.
This work finds that dense, randomly-initialized, feed-forward networks contain subnetworks ("winning tickets") that - when trained in isolation - reach test accuracy comparable to the original network in a similar number of iterations, and articulate the "lottery ticket hypothesis".
This work presents an acceleration method for CNNs, where it is shown that even simple filter pruning techniques can reduce inference costs for VGG-16 and ResNet-110 by up to 38% on CIFAR10 while regaining close to the original accuracy by retraining the networks.
This work presents a new approach that prunes a given network once at initialization prior to training, and introduces a saliency criterion based on connection sensitivity that identifies structurally important connections in the network for the given task.
A new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks (dubbed as ManiDP) is proposed, which shows better performance in terms of both accuracy and computational cost compared to the state-of-the-art methods.
This paper is able to add three fine-grained classification tasks to a single ImageNet-trained VGG-16 network and achieve accuracies close to those of separately trained networks for each task.
Neural architecture search is proposed to apply neural architecture search to search directly for a network with flexible channel and layer sizes to break the structure limitation of the pruned networks.
DNW provides an effective mechanism for discovering sparse subnetworks of predefined architectures in a single training run and is regarded as unifying core aspects of the neural architecture search problem with sparse neural network learning.
This work proposes to make pruning techniques aware of the robust training objective and let the training objective guide the search for which connections to prune, and demonstrates the success of this approach across CIFAR-10, SVHN, and ImageNet dataset with four robust training techniques.
Adding a benchmark result helps the community track progress.