3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in neural-network-compression-15
Use these libraries to find neural-network-compression-15 models and implementations
No subtasks available.
This paper shows that competitive compression rates can be achieved by using a version of “soft weight-sharing” (Nowlan & Hinton, 1992) and achieves both quantization and pruning in one simple (re-)training procedure, exposing the relation between compression and the minimum description length (MDL) principle.
This work proposes outlier channel splitting (OCS), which duplicates channels containing outliers, then halves the channel values, and shows that OCS can outperform state-of-the-art clipping techniques with only minor overhead.
A new simple and efficient iterative approach, which alternates low-rank factorization with a smart rank selection and fine-tuning, which improves the compression rate while maintaining the accuracy for a variety of tasks.
A novel framework for training efficient deep neural networks by exploiting generative adversarial networks (GANs) is proposed, where the pre-trained teacher networks are regarded as a fixed discriminator and the generator is utilized for derivating training samples which can obtain the maximum response on the discriminator.
THE AUTHORS' enables mixed-precision quantization without any access to the training or validation data, and it can finish the entire quantization process in less than 30s, which is very low computational overhead.
This paper tries to reduce the number of parameters of CNNs by learning a basis of the filters in convolutional layers, and validate the proposed solution for multiple CNN architectures on image classification and image super-resolution benchmarks.
A novel neural representation for videos (NeRV) which encodes videos in neural networks taking frame index as input, which can be used as a proxy for video compression, and achieve comparable performance to traditional frame-based video compression approaches.
A novel scheme for lossy weight encoding co-designed with weight simplification techniques that can compress weights by up to 496x without loss of model accuracy, resulting in up to a 1.51x improvement over the state-of-the-art.
This paper sets new state-of-the-art in neural network compression, as it strictly dominates previous approaches in a Pareto sense: on the benchmarks LeNet-5/MNIST and VGG-16/CIFAR-10, the approach yields the best test performance for a fixed memory budget, and vice versa, it achieves the highest compression rates for aFixed test performance.
Adding a benchmark result helps the community track progress.