Stay organized with collections
Save and categorize content based on your preferences.
Description:
ILSVRC 2012, commonly known as 'ImageNet' is an image dataset organized
according to the WordNet hierarchy. Each meaningful concept in WordNet, possibly
described by multiple words or word phrases, is called a "synonym set" or
"synset". There are more than 100,000 synsets in WordNet, majority of them are
nouns (80,000+). In ImageNet, we aim to provide on average 1000 images to
illustrate each synset. Images of each concept are quality-controlled and
human-annotated. In its completion, we hope ImageNet will offer tens of millions
of cleanly sorted images for most of the concepts in the WordNet hierarchy.
The test split contains 100K images but no labels because no labels have been
publicly released. We provide support for the test split from 2012 with the
minor patch released on October 10, 2019. In order to manually download this
data, a user must perform the following operations:
Download the October 10, 2019 patch. There is a Google Drive link to the
patch provided on the same page.
Combine the two tar-balls, manually overwriting any images in the original
archive with images from the patch. According to the instructions on
image-net.org, this procedure overwrites just a few images.
The resulting tar-ball may then be processed by TFDS.
To assess the accuracy of a model on the ImageNet test split, one must run
inference on all images in the split, export those results to a text file that
must be uploaded to the ImageNet evaluation server. The maintainers of the
ImageNet evaluation server permits a single user to submit up to 2 submissions
per week in order to prevent overfitting.
To evaluate the accuracy on the test split, one must first create an account at
image-net.org. This account must be approved by the site administrator. After
the account is created, one can submit the results to the test server at
https://image-net.org/challenges/LSVRC/eval_server.php The submission consists
of several ASCII text files corresponding to multiple tasks. The task of
interest is "Classification submission (top-5 cls error)". A sample of an
exported text file looks like the following:
The export format is described in full in "readme.txt" within the 2013
development kit available here:
https://image-net.org/data/ILSVRC/2013/ILSVRC2013_devkit.tgz Please see the
section entitled "3.3 CLS-LOC submission format". Briefly, the format of the
text file is 100,000 lines corresponding to each image in the test split. Each
line of integers correspond to the rank-ordered, top 5 predictions for each test
image. The integers are 1-indexed corresponding to the line number in the
corresponding labels file. See labels.txt.
2.0.1: Encoding fix. No changes from user point of view.
3.0.0: Fix colorization on ~12 images (CMYK -> RGB). Fix format for
consistency (convert the single png image to Jpeg). Faster generation
reading directly from the archive.
Manual download instructions: This dataset requires you to
download the source data manually into download_config.manual_dir
(defaults to ~/tensorflow_datasets/downloads/manual/):
manual_dir should contain two files: ILSVRC2012_img_train.tar and
ILSVRC2012_img_val.tar.
You need to register on https://image-net.org/download-images in order
to get the link to download the dataset.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2024-06-01 UTC."],[],[],null,["# imagenet2012\n\n\u003cbr /\u003e\n\n| **Warning:** Manual download required. See instructions below.\n\n- **Description**:\n\nILSVRC 2012, commonly known as 'ImageNet' is an image dataset organized\naccording to the WordNet hierarchy. Each meaningful concept in WordNet, possibly\ndescribed by multiple words or word phrases, is called a \"synonym set\" or\n\"synset\". There are more than 100,000 synsets in WordNet, majority of them are\nnouns (80,000+). In ImageNet, we aim to provide on average 1000 images to\nillustrate each synset. Images of each concept are quality-controlled and\nhuman-annotated. In its completion, we hope ImageNet will offer tens of millions\nof cleanly sorted images for most of the concepts in the WordNet hierarchy.\n\nThe test split contains 100K images but no labels because no labels have been\npublicly released. We provide support for the test split from 2012 with the\nminor patch released on October 10, 2019. In order to manually download this\ndata, a user must perform the following operations:\n\n1. Download the 2012 test split available [here](https://image-net.org/challenges/LSVRC/2012/2012-downloads.php#Images).\n2. Download the October 10, 2019 patch. There is a Google Drive link to the patch provided on the same page.\n3. Combine the two tar-balls, manually overwriting any images in the original archive with images from the patch. According to the instructions on image-net.org, this procedure overwrites just a few images.\n\nThe resulting tar-ball may then be processed by TFDS.\n\nTo assess the accuracy of a model on the ImageNet test split, one must run\ninference on all images in the split, export those results to a text file that\nmust be uploaded to the ImageNet evaluation server. The maintainers of the\nImageNet evaluation server permits a single user to submit up to 2 submissions\nper week in order to prevent overfitting.\n\nTo evaluate the accuracy on the test split, one must first create an account at\nimage-net.org. This account must be approved by the site administrator. After\nthe account is created, one can submit the results to the test server at\n\u003chttps://image-net.org/challenges/LSVRC/eval_server.php\u003e The submission consists\nof several ASCII text files corresponding to multiple tasks. The task of\ninterest is \"Classification submission (top-5 cls error)\". A sample of an\nexported text file looks like the following: \n\n 771 778 794 387 650\n 363 691 764 923 427\n 737 369 430 531 124\n 755 930 755 59 168\n\nThe export format is described in full in \"readme.txt\" within the 2013\ndevelopment kit available here:\n\u003chttps://image-net.org/data/ILSVRC/2013/ILSVRC2013_devkit.tgz\u003e Please see the\nsection entitled \"3.3 CLS-LOC submission format\". Briefly, the format of the\ntext file is 100,000 lines corresponding to each image in the test split. Each\nline of integers correspond to the rank-ordered, top 5 predictions for each test\nimage. The integers are 1-indexed corresponding to the line number in the\ncorresponding labels file. See labels.txt.\n\n- **Additional Documentation** :\n [Explore on Papers With Code\n north_east](https://paperswithcode.com/dataset/imagenet)\n\n- **Homepage** : \u003chttps://image-net.org/\u003e\n\n- **Source code** :\n [`tfds.datasets.imagenet2012.Builder`](https://github.com/tensorflow/datasets/tree/master/tensorflow_datasets/datasets/imagenet2012/imagenet2012_dataset_builder.py)\n\n- **Versions**:\n\n - `2.0.0`: Fix validation labels.\n - `2.0.1`: Encoding fix. No changes from user point of view.\n - `3.0.0`: Fix colorization on \\~12 images (CMYK -\\\u003e RGB). Fix format for\n consistency (convert the single png image to Jpeg). Faster generation\n reading directly from the archive.\n\n - `4.0.0`: (unpublished)\n\n - `5.0.0`: New split API (\u003chttps://tensorflow.org/datasets/splits\u003e)\n\n - **`5.1.0`** (default): Added test split.\n\n- **Download size** : `Unknown size`\n\n- **Dataset size** : `155.84 GiB`\n\n- **Manual download instructions** : This dataset requires you to\n download the source data manually into `download_config.manual_dir`\n (defaults to `~/tensorflow_datasets/downloads/manual/`): \n\n manual_dir should contain two files: ILSVRC2012_img_train.tar and\n ILSVRC2012_img_val.tar.\n You need to register on \u003chttps://image-net.org/download-images\u003e in order\n to get the link to download the dataset.\n\n- **Auto-cached**\n ([documentation](https://www.tensorflow.org/datasets/performances#auto-caching)):\n No\n\n- **Splits**:\n\n| Split | Examples |\n|----------------|-----------|\n| `'test'` | 100,000 |\n| `'train'` | 1,281,167 |\n| `'validation'` | 50,000 |\n\n- **Feature structure**:\n\n FeaturesDict({\n 'file_name': Text(shape=(), dtype=string),\n 'image': Image(shape=(None, None, 3), dtype=uint8),\n 'label': ClassLabel(shape=(), dtype=int64, num_classes=1000),\n })\n\n- **Feature documentation**:\n\n| Feature | Class | Shape | Dtype | Description |\n|-----------|--------------|-----------------|--------|-------------|\n| | FeaturesDict | | | |\n| file_name | Text | | string | |\n| image | Image | (None, None, 3) | uint8 | |\n| label | ClassLabel | | int64 | |\n\n- **Supervised keys** (See\n [`as_supervised` doc](https://www.tensorflow.org/datasets/api_docs/python/tfds/load#args)):\n `('image', 'label')`\n\n- **Figure**\n ([tfds.show_examples](https://www.tensorflow.org/datasets/api_docs/python/tfds/visualization/show_examples)):\n\n- **Examples** ([tfds.as_dataframe](https://www.tensorflow.org/datasets/api_docs/python/tfds/as_dataframe)):\n\nDisplay examples... \n\n- **Citation**:\n\n @article{ILSVRC15,\n Author = {Olga Russakovsky and Jia Deng and Hao Su and Jonathan Krause and Sanjeev Satheesh and Sean Ma and Zhiheng Huang and Andrej Karpathy and Aditya Khosla and Michael Bernstein and Alexander C. Berg and Li Fei-Fei},\n Title = { {ImageNet Large Scale Visual Recognition Challenge} },\n Year = {2015},\n journal = {International Journal of Computer Vision (IJCV)},\n doi = {10.1007/s11263-015-0816-y},\n volume={115},\n number={3},\n pages={211-252}\n }"]]