3260 papers • 126 benchmarks • 313 datasets
Medical Image Classification is a task in medical image analysis that involves classifying medical images, such as X-rays, MRI scans, and CT scans, into different categories based on the type of image or the presence of specific structures or diseases. The goal is to use computer algorithms to automatically identify and classify medical images based on their content, which can help in diagnosis, treatment planning, and disease monitoring.
(Image credit: Papersgraph)
These leaderboards are used to track progress in medical-image-classification-5
Use these libraries to find medical-image-classification-5 models and implementations
This work presents a residual learning framework to ease the training of networks that are substantially deeper than those used previously, and provides comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth.
The Dense Convolutional Network (DenseNet), which connects each layer to every other layer in a feed-forward fashion, and has several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.
A regulator module is proposed as a memory mechanism to extract complementary features of the intermediate layers, which are further fed to the ResNet, and named the new regulated network as regulated residual network (RegNet).
This paper proposes a novel building block for CNNs, namely Res2Net, by constructing hierarchical residual-like connections within one single residual block that represents multi-scale features at a granular level and increases the range of receptive fields for each network layer.
A new scaling method is proposed that uniformly scales all dimensions of depth/width/resolution using a simple yet highly effective compound coefficient and is demonstrated the effectiveness of this method on scaling up MobileNets and ResNet.
This paper re-evaluate the performance of the vanilla ResNet-50 when trained with a procedure that integrates novel optimization and data-augmentation advances, and shares competitive training settings and pre-trained models in the timm open-source library, with the hope that they will serve as better baselines for future work.
A Recurrent Convolutional Neural Network (RCNN) based on U-Net as well as a Recurrent Residual convolutional neural Network (RRCNN), which are named RU-Net and R2U-Net respectively are proposed, which show superior performance on segmentation tasks compared to equivalent models including U-nets and residual U- net.
This work proposes a new margin-based min-max surrogate loss function for the AUC score that is more robust than the commonly used AUC square loss, while enjoying the same advantage in terms of large-scale stochastic optimization.
This work introduces Dual Attention Vision Transformers (DaViT), a simple yet effective vision transformer architecture that is able to capture global context while maintaining computational efficiency and achieves state-of-the-art performance on four different tasks with efficient computations.
This work proposes an alternative unsupervised strategy to learn medical visual representations directly from the naturally occurring pairing of images and textual data, and shows that this method leads to image representations that considerably outperform strong baselines in most settings.
Adding a benchmark result helps the community track progress.