Repository navigation
maxvit
- Website
- Wikipedia
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
PyTorch Volume Models for 3D data
[ECCV 2022] unofficial pytorch implementation of the paper "MaxViT: Multi-Axis Vision Transformer"
This is a warehouse for MaxViT-Pytorch-model, can be used to train your image-datasets for vision tasks.
MTMAUNet: Multi-Task Multi-axis Attention UNet
Comprehensive Performance Analysis of Three Pretrained Transformer Models (ViT, Swin, and MaxViT) on ImageNet and Fine-tuned on the NIH Chest X-rays Dataset for Classifying 14 Chest Radiograph Pathologies
Facial Expression Recognition using the KDEF dataset.