GitXplorerGitXplorer
f

deit

public
4009 stars
549 forks
20 issues

Commits

List of commits on branch main.
Verified
7e160fe43f0252d17191b71cbb5826254114ea5b

Delete .circleci directory

bbigfootjon committed 6 months ago
Verified
263a3fcafc2bf17885a4af62e6030552f346dc71

Merge pull request #229 from bhheo/main

TTouvronHugo committed a year ago
Unverified
0db467b09c9643fc46bcdd2e5de64b651a0d6525

cosub bugfix

bbhheo committed a year ago
Verified
a6557706425be559de61b635d135ba553a7e8059

Merge pull request #228 from facebookresearch/dependabot/pip/torch-1.13.1

TTouvronHugo committed a year ago
Verified
dfeffae0e4b7f8d0577a1581a4dcdd3795011550

Bump torch from 1.7.0 to 1.13.1

ddependabot[bot] committed a year ago
Verified
a8ee80084c5b29d7d6c2cda76d7b8b74306c9365

Merge pull request #223 from fabfish/main

TTouvronHugo committed a year ago

README

The README file for this repository.

Data-Efficient architectures and training for Image classification

This repository contains PyTorch evaluation code, training code and pretrained models for the following papers:

DeiT Data-Efficient Image Transformers, ICML 2021 [bib]
@InProceedings{pmlr-v139-touvron21a,
  title =     {Training data-efficient image transformers & distillation through attention},
  author =    {Touvron, Hugo and Cord, Matthieu and Douze, Matthijs and Massa, Francisco and Sablayrolles, Alexandre and Jegou, Herve},
  booktitle = {International Conference on Machine Learning},
  pages =     {10347--10357},
  year =      {2021},
  volume =    {139},
  month =     {July}
}
CaiT (Going deeper with Image Transformers), ICCV 2021 [bib]
@InProceedings{Touvron_2021_ICCV,
    author    = {Touvron, Hugo and Cord, Matthieu and Sablayrolles, Alexandre and Synnaeve, Gabriel and J\'egou, Herv\'e},
    title     = {Going Deeper With Image Transformers},
    booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
    month     = {October},
    year      = {2021},
    pages     = {32-42}
}
ResMLP (ResMLP: Feedforward networks for image classification with data-efficient training), TPAMI 2022 [bib]
@article{touvron2021resmlp,
  title={ResMLP: Feedforward networks for image classification with data-efficient training},
  author={Hugo Touvron and Piotr Bojanowski and Mathilde Caron and Matthieu Cord and Alaaeldin El-Nouby and Edouard Grave and Gautier Izacard and Armand Joulin and Gabriel Synnaeve and Jakob Verbeek and Herv'e J'egou},
  journal={arXiv preprint arXiv:2105.03404},
  year={2021},
}
PatchConvnet (Augmenting Convolutional networks with attention-based aggregation) [bib]
@article{touvron2021patchconvnet,
  title={Augmenting Convolutional networks with attention-based aggregation},
  author={Hugo Touvron and Matthieu Cord and Alaaeldin El-Nouby and Piotr Bojanowski and Armand Joulin and Gabriel Synnaeve and Jakob Verbeek and Herve Jegou},
  journal={arXiv preprint arXiv:2112.13692},
  year={2021},
}
3Things (Three things everyone should know about Vision Transformers), ECCV 2022 [bib]
@article{Touvron2022ThreeTE,
  title={Three things everyone should know about Vision Transformers},
  author={Hugo Touvron and Matthieu Cord and Alaaeldin El-Nouby and Jakob Verbeek and Herve Jegou},
  journal={arXiv preprint arXiv:2203.09795},
  year={2022},
}
DeiT III (DeiT III: Revenge of the ViT), ECCV 2022 [bib]
@article{Touvron2022DeiTIR,
  title={DeiT III: Revenge of the ViT},
  author={Hugo Touvron and Matthieu Cord and Herve Jegou},
  journal={arXiv preprint arXiv:2204.07118},
  year={2022},
}
Cosub (Co-training 2L Submodels for Visual Recognition), CVPR 2023 [bib]
@article{Touvron2022Cotraining2S,
  title={Co-training 2L Submodels for Visual Recognition},
  author={Hugo Touvron and Matthieu Cord and Maxime Oquab and Piotr Bojanowski and Jakob Verbeek and Herv'e J'egou},
  journal={arXiv preprint arXiv:2212.04884},
  year={2022},
}
If you find this repository useful, please consider giving a star ⭐ and cite the relevant papers.

License

This repository is released under the Apache 2.0 license as found in the LICENSE file.

Contributing

We actively welcome your pull requests! Please see CONTRIBUTING.md and CODE_OF_CONDUCT.md for more info.