We fine-tune 7 models including ViTs, DINO, CLIP, ConvNeXt, ResNet, on
$ 20.50 · 4.5 (648) · In stock
A Broad Study of Pre-training for Domain Generalization and Adaptation
2301.02240] Skip-Attention: Improving Vision Transformers by Paying Less Attention
2301.02240] Skip-Attention: Improving Vision Transformers by Paying Less Attention
A Broad Study of Pre-training for Domain Generalization and Adaptation
Review — ConvNeXt: A ConvNet for the 2020s, by Sik-Ho Tsang
Vision Transformer (ViT)
pytorch-image-models/README.md at main · huggingface/pytorch-image-models · GitHub
Ruoqi Shen's research works
Remote Sensing, Free Full-Text
PDF] ConvNeXt V2: Co-designing and Scaling ConvNets with Masked Autoencoders
PDF) Adversarial Attacks on Foundational Vision Models
GitHub - rwightman/timm: PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, EfficientNetV2, NFNet, Vision Transformer, MixNet, MobileNet-V3/V2, RegNet, DPN, CSPNet, and more
GitHub - huggingface/pytorch-image-models: PyTorch image models, scripts, pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more