YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Deep Incubation
This repository contains the pre-trained models for Deep Incubation.
Title:ββDeep Incubation: Training Large Models by Divide-and-Conquering
Authors: Zanlin Ni, Yulin Wang, Jiangwei Yu, Haojun Jiang, Yue Cao, Gao Huang (Corresponding Author)
Institute: Tsinghua University and Beijing Academy of Artificial Intelligence (BAAI)
Publish: arXiv preprint (arXiv 2212.04129)
Contact: nzl22 at mails dot tsinghua dot edu dot cn
Models
model | image size | #param. | top-1 acc. | checkpoint |
---|---|---|---|---|
ViT-B | 224x224 | 87M | 82.4% | π€ HF link |
ViT-B | 384x384 | 87M | 84.2% | π€ HF link |
ViT-L | 224x224 | 304M | 83.9% | π€ HF link |
ViT-L | 384x384 | 304M | 85.3% | π€ HF link |
ViT-H | 224x224 | 632M | 84.3% | π€ HF link |
ViT-H | 392x392 | 632M | 85.6% | π€ HF link |
Data Preparation
- The ImageNet dataset should be prepared as follows:
data
βββ train
β βββ folder 1 (class 1)
β βββ folder 2 (class 1)
β βββ ...
βββ val
β βββ folder 1 (class 1)
β βββ folder 2 (class 1)
β βββ ...
Citation
If you find our work helpful, please starπ this repo and citeπ our paper. Thanks for your support!
@article{Ni2022Incub,
title={Deep Incubation: Training Large Models by Divide-and-Conquering},
author={Ni, Zanlin and Wang, Yulin and Yu, Jiangwei and Jiang, Haojun and Cao, Yue and Huang, Gao},
journal={arXiv preprint arXiv:2212.04129},
year={2022}
}
Acknowledgements
Our implementation is mainly based on deit. We thank to their clean codebase.
Contact
If you have any questions or concerns, please send mail to [email protected].