Name | DL | Torrents | Total Size | Deep Learning [edit] | 50 | 963.86GB | 475 | 0 |
inception-21k.tar.gz | 125.14MB |
Type: Dataset
Tags:
Bibtex:
Tags:
Bibtex:
@article{, title= {MXNet pre-trained model Full ImageNet Network inception-21k.tar.gz}, keywords= {}, journal= {}, author= {dmlc}, year= {}, url= {https://github.com/dmlc/mxnet-model-gallery/blob/master/imagenet-21k-inception.md}, license= {}, abstract= {# Full ImageNet Network This model is a pretrained model on full imagenet dataset [1] with 14,197,087 images in 21,841 classes. The model is trained by only random crop and mirror augmentation. The network is based on Inception-BN network [2], and added more capacity. This network runs roughly 2 times slower than standard Inception-BN Network. We trained this network on a machine with 4 GeForce GTX 980 GPU. Each round costs 23 hours, the released model is the 9 round. Train Top-1 Accuracy over 21,841 classes: 37.19% Single image prediction memory requirement: 15MB ILVRC2012 Validation Performance: | | Over 1,000 classes | Over 21,841 classes | | ------ | ------------------ | ------------------- | | Top-1 | 68.3% | 41.9% | | Top-5 | 89.0% | 69.6% | | Top=20 | 96.0% | 83.6% | Note: Directly use 21k prediction may lose diversity in output. You may choose a subset from 21k to make perdiction more reasonable. The compressed file contains: - ```Inception-symbol.json```: symbolic network - ```Inception-0009.params```: network parameter - ```synset.txt```: prediction label/text mapping There is no mean image file for this model. We use ```mean_r=117```, ```mean_g=117``` and ```mean_b=117``` to noramlize the image. ##### Reference: [1] Deng, Jia, et al. "Imagenet: A large-scale hierarchical image database." *Computer Vision and Pattern Recognition*, 2009. CVPR 2009. IEEE Conference on. IEEE, 2009. [2] Ioffe, Sergey, and Christian Szegedy. "Batch normalization: Accelerating deep network training by reducing internal covariate shift." *arXiv preprint arXiv:1502.03167* (2015).}, superseded= {}, terms= {} }