clarifai 9.7.0__py3-none-any.whl → 9.7.2__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- clarifai/auth/__init__.py +6 -0
- clarifai/auth/helper.py +35 -36
- clarifai/auth/register.py +23 -0
- clarifai/{client → auth}/stub.py +10 -10
- clarifai/client/__init__.py +1 -4
- clarifai/client/app.py +483 -0
- clarifai/client/auth/__init__.py +4 -0
- clarifai/client/{abc.py → auth/abc.py} +2 -2
- clarifai/client/auth/helper.py +377 -0
- clarifai/client/auth/register.py +23 -0
- {clarifai_utils/client → clarifai/client/auth}/stub.py +10 -10
- clarifai/client/base.py +112 -0
- clarifai/client/dataset.py +290 -0
- clarifai/client/input.py +730 -0
- clarifai/client/lister.py +41 -0
- clarifai/client/model.py +218 -0
- clarifai/client/module.py +82 -0
- clarifai/client/user.py +125 -0
- clarifai/client/workflow.py +194 -0
- clarifai/datasets/upload/base.py +66 -0
- clarifai/datasets/upload/examples/README.md +31 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/dataset.py +42 -0
- clarifai/datasets/upload/examples/image_classification/food-101/dataset.py +39 -0
- clarifai/datasets/upload/examples/text_classification/imdb_dataset/dataset.py +37 -0
- clarifai/{data_upload/datasets → datasets/upload}/features.py +4 -12
- clarifai/datasets/upload/image.py +156 -0
- clarifai/datasets/upload/loaders/README.md +49 -0
- clarifai/{data_upload/datasets/zoo → datasets/upload/loaders}/coco_captions.py +24 -21
- {clarifai_utils/data_upload/datasets/zoo → clarifai/datasets/upload/loaders}/coco_detection.py +46 -42
- clarifai/datasets/upload/loaders/coco_segmentation.py +166 -0
- clarifai/{data_upload/datasets/zoo → datasets/upload/loaders}/imagenet_classification.py +22 -12
- clarifai/{data_upload/datasets/zoo → datasets/upload/loaders}/xview_detection.py +44 -53
- clarifai/datasets/upload/text.py +50 -0
- clarifai/datasets/upload/utils.py +62 -0
- clarifai/errors.py +90 -0
- clarifai/urls/helper.py +16 -17
- clarifai/utils/logging.py +40 -0
- clarifai/utils/misc.py +33 -0
- clarifai/versions.py +6 -0
- {clarifai-9.7.0.dist-info → clarifai-9.7.2.dist-info}/LICENSE +1 -1
- clarifai-9.7.2.dist-info/METADATA +179 -0
- clarifai-9.7.2.dist-info/RECORD +350 -0
- clarifai_utils/auth/__init__.py +6 -0
- clarifai_utils/auth/helper.py +35 -36
- clarifai_utils/auth/register.py +23 -0
- clarifai_utils/auth/stub.py +127 -0
- clarifai_utils/client/__init__.py +1 -4
- clarifai_utils/client/app.py +483 -0
- clarifai_utils/client/auth/__init__.py +4 -0
- clarifai_utils/client/{abc.py → auth/abc.py} +2 -2
- clarifai_utils/client/auth/helper.py +377 -0
- clarifai_utils/client/auth/register.py +23 -0
- clarifai_utils/client/auth/stub.py +127 -0
- clarifai_utils/client/base.py +112 -0
- clarifai_utils/client/dataset.py +290 -0
- clarifai_utils/client/input.py +730 -0
- clarifai_utils/client/lister.py +41 -0
- clarifai_utils/client/model.py +218 -0
- clarifai_utils/client/module.py +82 -0
- clarifai_utils/client/user.py +125 -0
- clarifai_utils/client/workflow.py +194 -0
- clarifai_utils/datasets/upload/base.py +66 -0
- clarifai_utils/datasets/upload/examples/README.md +31 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/dataset.py +42 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/dataset.py +39 -0
- clarifai_utils/datasets/upload/examples/text_classification/imdb_dataset/dataset.py +37 -0
- clarifai_utils/{data_upload/datasets → datasets/upload}/features.py +4 -12
- clarifai_utils/datasets/upload/image.py +156 -0
- clarifai_utils/datasets/upload/loaders/README.md +49 -0
- clarifai_utils/{data_upload/datasets/zoo → datasets/upload/loaders}/coco_captions.py +24 -21
- {clarifai/data_upload/datasets/zoo → clarifai_utils/datasets/upload/loaders}/coco_detection.py +46 -42
- clarifai_utils/datasets/upload/loaders/coco_segmentation.py +166 -0
- clarifai_utils/{data_upload/datasets/zoo → datasets/upload/loaders}/imagenet_classification.py +22 -12
- clarifai_utils/{data_upload/datasets/zoo → datasets/upload/loaders}/xview_detection.py +44 -53
- clarifai_utils/datasets/upload/text.py +50 -0
- clarifai_utils/datasets/upload/utils.py +62 -0
- clarifai_utils/errors.py +90 -0
- clarifai_utils/urls/helper.py +16 -17
- clarifai_utils/utils/logging.py +40 -0
- clarifai_utils/utils/misc.py +33 -0
- clarifai_utils/versions.py +6 -0
- clarifai/data_upload/README.md +0 -63
- clarifai/data_upload/convert_csv.py +0 -182
- clarifai/data_upload/datasets/base.py +0 -87
- clarifai/data_upload/datasets/image.py +0 -253
- clarifai/data_upload/datasets/text.py +0 -60
- clarifai/data_upload/datasets/zoo/README.md +0 -55
- clarifai/data_upload/datasets/zoo/coco_segmentation.py +0 -160
- clarifai/data_upload/examples/README.md +0 -5
- clarifai/data_upload/examples/image_classification/cifar10/dataset.py +0 -40
- clarifai/data_upload/examples/image_classification/food-101/dataset.py +0 -39
- clarifai/data_upload/examples/image_classification/food-101/images/beignets/1036242.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/beignets/1114182.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/beignets/2012944.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/beignets/2464389.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/beignets/478632.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/hamburger/1061270.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/hamburger/1202261.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/hamburger/1381751.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/hamburger/3289634.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/hamburger/862025.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/102197.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/2749372.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/2938268.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/3590861.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/746716.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/ramen/2955110.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/ramen/3208966.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/ramen/3270629.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/ramen/3424562.jpg +0 -0
- clarifai/data_upload/examples/image_classification/food-101/images/ramen/544680.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/annotations/2007_000464.xml +0 -39
- clarifai/data_upload/examples/image_detection/voc/annotations/2008_000853.xml +0 -28
- clarifai/data_upload/examples/image_detection/voc/annotations/2008_003182.xml +0 -54
- clarifai/data_upload/examples/image_detection/voc/annotations/2008_008526.xml +0 -67
- clarifai/data_upload/examples/image_detection/voc/annotations/2009_004315.xml +0 -28
- clarifai/data_upload/examples/image_detection/voc/annotations/2009_004382.xml +0 -28
- clarifai/data_upload/examples/image_detection/voc/annotations/2011_000430.xml +0 -28
- clarifai/data_upload/examples/image_detection/voc/annotations/2011_001610.xml +0 -46
- clarifai/data_upload/examples/image_detection/voc/annotations/2011_006412.xml +0 -99
- clarifai/data_upload/examples/image_detection/voc/annotations/2012_000690.xml +0 -43
- clarifai/data_upload/examples/image_detection/voc/dataset.py +0 -76
- clarifai/data_upload/examples/image_detection/voc/images/2007_000464.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2008_000853.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2008_003182.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2008_008526.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2009_004315.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2009_004382.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2011_000430.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2011_001610.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2011_006412.jpg +0 -0
- clarifai/data_upload/examples/image_detection/voc/images/2012_000690.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/annotations/instances_val2017_subset.json +0 -5342
- clarifai/data_upload/examples/image_segmentation/coco/dataset.py +0 -107
- clarifai/data_upload/examples/image_segmentation/coco/images/000000074646.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000086956.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000166563.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000176857.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000182202.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000193245.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000384850.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000409630.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000424349.jpg +0 -0
- clarifai/data_upload/examples/image_segmentation/coco/images/000000573008.jpg +0 -0
- clarifai/data_upload/examples/text_classification/imdb_dataset/dataset.py +0 -40
- clarifai/data_upload/examples.py +0 -17
- clarifai/data_upload/upload.py +0 -356
- clarifai/dataset_export/dataset_export_inputs.py +0 -205
- clarifai/listing/concepts.py +0 -37
- clarifai/listing/datasets.py +0 -37
- clarifai/listing/inputs.py +0 -111
- clarifai/listing/installed_module_versions.py +0 -40
- clarifai/listing/lister.py +0 -200
- clarifai/listing/models.py +0 -46
- clarifai/listing/module_versions.py +0 -42
- clarifai/listing/modules.py +0 -36
- clarifai/runners/base.py +0 -140
- clarifai/runners/example.py +0 -36
- clarifai-9.7.0.dist-info/METADATA +0 -99
- clarifai-9.7.0.dist-info/RECORD +0 -456
- clarifai_utils/data_upload/README.md +0 -63
- clarifai_utils/data_upload/convert_csv.py +0 -182
- clarifai_utils/data_upload/datasets/base.py +0 -87
- clarifai_utils/data_upload/datasets/image.py +0 -253
- clarifai_utils/data_upload/datasets/text.py +0 -60
- clarifai_utils/data_upload/datasets/zoo/README.md +0 -55
- clarifai_utils/data_upload/datasets/zoo/coco_segmentation.py +0 -160
- clarifai_utils/data_upload/examples/README.md +0 -5
- clarifai_utils/data_upload/examples/image_classification/cifar10/dataset.py +0 -40
- clarifai_utils/data_upload/examples/image_classification/food-101/dataset.py +0 -39
- clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/1036242.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/1114182.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/2012944.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/2464389.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/478632.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1061270.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1202261.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1381751.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/3289634.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/862025.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/102197.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/2749372.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/2938268.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/3590861.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/746716.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/2955110.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/3208966.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/3270629.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/3424562.jpg +0 -0
- clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/544680.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/__init__.py +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/__init__.py +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2007_000464.xml +0 -39
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2008_000853.xml +0 -28
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2008_003182.xml +0 -54
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2008_008526.xml +0 -67
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2009_004315.xml +0 -28
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2009_004382.xml +0 -28
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2011_000430.xml +0 -28
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2011_001610.xml +0 -46
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2011_006412.xml +0 -99
- clarifai_utils/data_upload/examples/image_detection/voc/annotations/2012_000690.xml +0 -43
- clarifai_utils/data_upload/examples/image_detection/voc/dataset.py +0 -76
- clarifai_utils/data_upload/examples/image_detection/voc/images/2007_000464.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2008_000853.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2008_003182.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2008_008526.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2009_004315.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2009_004382.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2011_000430.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2011_001610.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2011_006412.jpg +0 -0
- clarifai_utils/data_upload/examples/image_detection/voc/images/2012_000690.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/__init__.py +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/__init__.py +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/annotations/instances_val2017_subset.json +0 -5342
- clarifai_utils/data_upload/examples/image_segmentation/coco/dataset.py +0 -107
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000074646.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000086956.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000166563.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000176857.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000182202.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000193245.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000384850.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000409630.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000424349.jpg +0 -0
- clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000573008.jpg +0 -0
- clarifai_utils/data_upload/examples/text_classification/__init__.py +0 -0
- clarifai_utils/data_upload/examples/text_classification/imdb_dataset/__init__.py +0 -0
- clarifai_utils/data_upload/examples/text_classification/imdb_dataset/dataset.py +0 -40
- clarifai_utils/data_upload/examples.py +0 -17
- clarifai_utils/data_upload/upload.py +0 -356
- clarifai_utils/dataset_export/dataset_export_inputs.py +0 -205
- clarifai_utils/listing/__init__.py +0 -0
- clarifai_utils/listing/concepts.py +0 -37
- clarifai_utils/listing/datasets.py +0 -37
- clarifai_utils/listing/inputs.py +0 -111
- clarifai_utils/listing/installed_module_versions.py +0 -40
- clarifai_utils/listing/lister.py +0 -200
- clarifai_utils/listing/models.py +0 -46
- clarifai_utils/listing/module_versions.py +0 -42
- clarifai_utils/listing/modules.py +0 -36
- clarifai_utils/runners/__init__.py +0 -0
- clarifai_utils/runners/base.py +0 -140
- clarifai_utils/runners/example.py +0 -36
- /clarifai/{data_upload/__init__.py → cli.py} +0 -0
- /clarifai/{data_upload/datasets → datasets}/__init__.py +0 -0
- /clarifai/{data_upload/datasets/zoo → datasets/upload}/__init__.py +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/__init__.py +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/__init__.py +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/cifar_small_test.csv +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/cifar_small_train.csv +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_700.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_701.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_702.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_703.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_704.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_705.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_706.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_707.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_708.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_709.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/__init__.py +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/1420783.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/3287885.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/3617075.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/38052.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/39147.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/139558.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/1636096.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/2480925.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/3385808.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/3647386.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/1826869.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/2243245.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/259212.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/2842688.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/3035414.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/1545393.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/2427642.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/3520891.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/377566.jpg +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/503504.jpg +0 -0
- /clarifai/{data_upload/examples/image_detection → datasets/upload/examples/text_classification}/__init__.py +0 -0
- /clarifai/{data_upload/examples/image_detection/voc → datasets/upload/examples/text_classification/imdb_dataset}/__init__.py +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/text_classification/imdb_dataset/test.csv +0 -0
- /clarifai/{data_upload → datasets/upload}/examples/text_classification/imdb_dataset/train.csv +0 -0
- /clarifai/{data_upload/examples/image_segmentation → datasets/upload/loaders}/__init__.py +0 -0
- /clarifai/{data_upload/examples/image_segmentation/coco → utils}/__init__.py +0 -0
- {clarifai-9.7.0.dist-info → clarifai-9.7.2.dist-info}/WHEEL +0 -0
- {clarifai-9.7.0.dist-info → clarifai-9.7.2.dist-info}/entry_points.txt +0 -0
- {clarifai-9.7.0.dist-info → clarifai-9.7.2.dist-info}/top_level.txt +0 -0
- /clarifai/data_upload/examples/text_classification/__init__.py → /clarifai_utils/cli.py +0 -0
- {clarifai/data_upload/examples/text_classification/imdb_dataset → clarifai_utils/datasets}/__init__.py +0 -0
- {clarifai/listing → clarifai_utils/datasets/upload}/__init__.py +0 -0
- {clarifai/runners → clarifai_utils/datasets/upload/examples/image_classification}/__init__.py +0 -0
- /clarifai_utils/{data_upload → datasets/upload/examples/image_classification/cifar10}/__init__.py +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/cifar_small_test.csv +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/cifar_small_train.csv +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_700.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_701.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_702.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_703.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_704.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_705.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_706.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_707.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_708.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/cifar10/images/test_batch_709.jpg +0 -0
- /clarifai_utils/{data_upload/datasets → datasets/upload/examples/image_classification/food-101}/__init__.py +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/1420783.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/3287885.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/3617075.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/38052.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/beignets/39147.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/139558.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/1636096.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/2480925.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/3385808.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/hamburger/3647386.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/1826869.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/2243245.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/259212.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/2842688.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/prime_rib/3035414.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/1545393.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/2427642.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/3520891.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/377566.jpg +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/image_classification/food-101/images/ramen/503504.jpg +0 -0
- /clarifai_utils/{data_upload/datasets/zoo → datasets/upload/examples/text_classification}/__init__.py +0 -0
- /clarifai_utils/{data_upload/examples/image_classification → datasets/upload/examples/text_classification/imdb_dataset}/__init__.py +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/text_classification/imdb_dataset/test.csv +0 -0
- /clarifai_utils/{data_upload → datasets/upload}/examples/text_classification/imdb_dataset/train.csv +0 -0
- /clarifai_utils/{data_upload/examples/image_classification/cifar10 → datasets/upload/loaders}/__init__.py +0 -0
- /clarifai_utils/{data_upload/examples/image_classification/food-101 → utils}/__init__.py +0 -0
|
@@ -1,160 +0,0 @@
|
|
|
1
|
-
#! COCO 2017 Image Segmentation dataset
|
|
2
|
-
|
|
3
|
-
import gc
|
|
4
|
-
import os
|
|
5
|
-
import zipfile
|
|
6
|
-
from functools import reduce
|
|
7
|
-
from glob import glob
|
|
8
|
-
|
|
9
|
-
import cv2
|
|
10
|
-
import numpy as np
|
|
11
|
-
import requests
|
|
12
|
-
from pycocotools import mask as maskUtils
|
|
13
|
-
from pycocotools.coco import COCO
|
|
14
|
-
from tqdm import tqdm
|
|
15
|
-
|
|
16
|
-
from ..features import VisualSegmentationFeatures
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
class COCOSegmentationDataset:
|
|
20
|
-
"""COCO 2017 Image Segmentation Dataset."""
|
|
21
|
-
|
|
22
|
-
def __init__(self, split: str = "train"):
|
|
23
|
-
"""
|
|
24
|
-
Initialize coco dataset.
|
|
25
|
-
Args:
|
|
26
|
-
filenames: the coco zip filenames: Dict[str, str] to be downloaded if download=True,
|
|
27
|
-
data_dir: the local coco dataset directory
|
|
28
|
-
split: "train" or "val"
|
|
29
|
-
"""
|
|
30
|
-
self.filenames = {
|
|
31
|
-
"train": "train2017.zip",
|
|
32
|
-
"val": "val2017.zip",
|
|
33
|
-
"annotations": "annotations_trainval2017.zip"
|
|
34
|
-
}
|
|
35
|
-
self.split = split
|
|
36
|
-
self.url = "http://images.cocodataset.org/zips/" # coco base image-zip url
|
|
37
|
-
self.data_dir = os.path.join(os.curdir, "data") # data storage dir
|
|
38
|
-
self.extracted_coco_dirs = {"train": None, "val": None, "annotations": None}
|
|
39
|
-
|
|
40
|
-
def coco_download(self, save_dir):
|
|
41
|
-
"""Download coco dataset."""
|
|
42
|
-
if not os.path.exists(save_dir):
|
|
43
|
-
os.mkdir(save_dir)
|
|
44
|
-
|
|
45
|
-
#check if train, val and annotation dirs exist
|
|
46
|
-
#so that the coco2017 data isn't downloaded
|
|
47
|
-
for key, filename in self.filenames.items():
|
|
48
|
-
existing_files = glob(f"{save_dir}/{key}*")
|
|
49
|
-
if existing_files:
|
|
50
|
-
print(f"{key} dataset already downloded and extracted")
|
|
51
|
-
continue
|
|
52
|
-
|
|
53
|
-
print("-" * 80)
|
|
54
|
-
print(f"Downloading {filename}")
|
|
55
|
-
print("-" * 80)
|
|
56
|
-
|
|
57
|
-
if "annotations" in filename:
|
|
58
|
-
self.url = "http://images.cocodataset.org/annotations/"
|
|
59
|
-
|
|
60
|
-
response = requests.get(self.url + filename, stream=True)
|
|
61
|
-
response.raise_for_status()
|
|
62
|
-
with open(os.path.join(save_dir, filename), "wb") as _file:
|
|
63
|
-
for chunk in tqdm(response.iter_content(chunk_size=5124000)):
|
|
64
|
-
if chunk:
|
|
65
|
-
_file.write(chunk)
|
|
66
|
-
print("Coco data download complete...")
|
|
67
|
-
|
|
68
|
-
#extract files
|
|
69
|
-
zf = zipfile.ZipFile(os.path.join(save_dir, filename))
|
|
70
|
-
print(f" Extracting {filename} file")
|
|
71
|
-
zf.extractall(path=save_dir)
|
|
72
|
-
# Delete coco zip
|
|
73
|
-
print(f" Deleting {filename}")
|
|
74
|
-
os.remove(path=os.path.join(save_dir, filename))
|
|
75
|
-
|
|
76
|
-
def dataloader(self):
|
|
77
|
-
"""
|
|
78
|
-
Transform coco data into clarifai proto compatible format for upload.
|
|
79
|
-
Returns:
|
|
80
|
-
VisualSegmentationFeatures type generator.
|
|
81
|
-
"""
|
|
82
|
-
if isinstance(self.filenames, dict) and len(self.filenames) == 3:
|
|
83
|
-
self.coco_download(self.data_dir)
|
|
84
|
-
self.extracted_coco_dirs["train"] = [os.path.join(self.data_dir, i) \
|
|
85
|
-
for i in os.listdir(self.data_dir) if "train" in i][0]
|
|
86
|
-
self.extracted_coco_dirs["val"] = [os.path.join(self.data_dir, i) \
|
|
87
|
-
for i in os.listdir(self.data_dir) if "val" in i][0]
|
|
88
|
-
|
|
89
|
-
self.extracted_coco_dirs["annotations"] = [os.path.join(self.data_dir, i) \
|
|
90
|
-
for i in os.listdir(self.data_dir) if "annotations" in i][0]
|
|
91
|
-
else:
|
|
92
|
-
raise Exception(f"`filenames` must be a dict of atleast 3 coco zip file names; \
|
|
93
|
-
train, val and annotations. Found {len(self.filenames)} items instead.")
|
|
94
|
-
|
|
95
|
-
annot_file = glob(self.extracted_coco_dirs["annotations"] + "/" +\
|
|
96
|
-
f"instances_{self.split}*")[0]
|
|
97
|
-
coco = COCO(annot_file)
|
|
98
|
-
categories = coco.loadCats(coco.getCatIds())
|
|
99
|
-
cat_id_map = {category["id"]: category["name"] for category in categories}
|
|
100
|
-
cat_img_ids = {}
|
|
101
|
-
for cat_id in list(cat_id_map.keys()):
|
|
102
|
-
cat_img_ids[cat_id] = coco.getImgIds(catIds=[cat_id])
|
|
103
|
-
|
|
104
|
-
img_ids = []
|
|
105
|
-
for i in list(cat_img_ids.values()):
|
|
106
|
-
img_ids.extend(i)
|
|
107
|
-
|
|
108
|
-
#get annotations for each image id
|
|
109
|
-
for _id in set(img_ids):
|
|
110
|
-
annots = [] # polygons
|
|
111
|
-
class_names = []
|
|
112
|
-
labels = [i for i in list(filter(lambda x: _id in cat_img_ids[x], cat_img_ids))]
|
|
113
|
-
image_path = glob(self.extracted_coco_dirs[self.split]+"/"+\
|
|
114
|
-
f"{str(_id).zfill(12)}*")[0]
|
|
115
|
-
|
|
116
|
-
image_height, image_width = cv2.imread(image_path).shape[:2]
|
|
117
|
-
for cat_id in labels:
|
|
118
|
-
annot_ids = coco.getAnnIds(imgIds=_id, catIds=[cat_id])
|
|
119
|
-
if len(annot_ids) > 0:
|
|
120
|
-
img_annotations = coco.loadAnns(annot_ids)
|
|
121
|
-
for ann in img_annotations:
|
|
122
|
-
# get polygons
|
|
123
|
-
if type(ann['segmentation']) == list:
|
|
124
|
-
for seg in ann['segmentation']:
|
|
125
|
-
poly = np.array(seg).reshape((int(len(seg) / 2), 2))
|
|
126
|
-
poly[:, 0], poly[:, 1] = poly[:, 0] / image_width, poly[:, 1] / image_height
|
|
127
|
-
annots.append(poly.tolist()) #[[x=col, y=row],...]
|
|
128
|
-
class_names.append(cat_id_map[cat_id])
|
|
129
|
-
else: # seg: {"counts":[...]}
|
|
130
|
-
if type(ann['segmentation']['counts']) == list:
|
|
131
|
-
rle = maskUtils.frPyObjects([ann['segmentation']], image_height, image_width)
|
|
132
|
-
else:
|
|
133
|
-
rle = ann['segmentation']
|
|
134
|
-
mask = maskUtils.decode(rle) #binary mask
|
|
135
|
-
#convert mask to polygons and add to annots
|
|
136
|
-
contours, _ = cv2.findContours(mask, cv2.RETR_TREE, cv2.CHAIN_APPROX_SIMPLE)
|
|
137
|
-
polygons = []
|
|
138
|
-
for cont in contours:
|
|
139
|
-
if cont.size >= 6:
|
|
140
|
-
polygons.append(cont.astype(float).flatten().tolist())
|
|
141
|
-
# store polygons in (x,y) pairs
|
|
142
|
-
polygons_flattened = reduce(lambda x, y: x + y, polygons)
|
|
143
|
-
del polygons
|
|
144
|
-
del contours
|
|
145
|
-
del mask
|
|
146
|
-
gc.collect()
|
|
147
|
-
|
|
148
|
-
polygons = np.array(polygons_flattened).reshape((int(len(polygons_flattened) / 2),
|
|
149
|
-
2))
|
|
150
|
-
polygons[:, 0] = polygons[:, 0] / image_width
|
|
151
|
-
polygons[:, 1] = polygons[:, 1] / image_height
|
|
152
|
-
|
|
153
|
-
annots.append(polygons.tolist()) #[[x=col, y=row],...,[x=col, y=row]]
|
|
154
|
-
class_names.append(cat_id_map[cat_id])
|
|
155
|
-
else: # if no annotations for given image_id-cat_id pair
|
|
156
|
-
continue
|
|
157
|
-
assert len(class_names) == len(annots), f"Num classes must match num annotations\
|
|
158
|
-
for a single image. Found {len(class_names)} classes and {len(annots)} polygons."
|
|
159
|
-
|
|
160
|
-
yield VisualSegmentationFeatures(image_path, class_names, annots, id=_id)
|
|
@@ -1,5 +0,0 @@
|
|
|
1
|
-
## Data upload from packages
|
|
2
|
-
|
|
3
|
-
Examples of how to upload your local directory datasets into clarifai app datasets using data_upload utils `from_module` feature.
|
|
4
|
-
|
|
5
|
-
The `examples.py` script imports the data upload module and executes the upload of the IMDB_reviews text_classification and [food-101 image classification](https://data.vision.ee.ethz.ch/cvl/datasets_extra/food-101/) datasets by passing the package paths as a parameter in UploadConfig().
|
|
@@ -1,40 +0,0 @@
|
|
|
1
|
-
#! Cifar10 Dataset
|
|
2
|
-
|
|
3
|
-
import csv
|
|
4
|
-
import os
|
|
5
|
-
|
|
6
|
-
from clarifai.data_upload.datasets.features import VisualClassificationFeatures
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
class Cifar10Dataset:
|
|
10
|
-
"""Cifar10 Dataset."""
|
|
11
|
-
|
|
12
|
-
def __init__(self, split: str = "train"):
|
|
13
|
-
"""
|
|
14
|
-
Initialize dataset params.
|
|
15
|
-
Args:
|
|
16
|
-
data_dir: the local dataset directory.
|
|
17
|
-
split: "train" or "test"
|
|
18
|
-
"""
|
|
19
|
-
self.split = split
|
|
20
|
-
self.data_dirs = {
|
|
21
|
-
"train": os.path.join(os.path.dirname(__file__), "cifar_small_train.csv"),
|
|
22
|
-
"test": os.path.join(os.path.dirname(__file__), "cifar_small_test.csv")
|
|
23
|
-
}
|
|
24
|
-
|
|
25
|
-
def dataloader(self):
|
|
26
|
-
"""
|
|
27
|
-
Transform text data into clarifai proto compatible
|
|
28
|
-
format for upload.
|
|
29
|
-
Returns:
|
|
30
|
-
TextFeatures type generator.
|
|
31
|
-
"""
|
|
32
|
-
## Your preprocessing code here
|
|
33
|
-
with open(self.data_dirs[self.split]) as _file:
|
|
34
|
-
reader = csv.reader(_file)
|
|
35
|
-
next(reader, None) # skip header
|
|
36
|
-
for review in reader:
|
|
37
|
-
yield VisualClassificationFeatures(
|
|
38
|
-
image_path='examples/image_classification/cifar10/' + review[0],
|
|
39
|
-
label=review[1],
|
|
40
|
-
id=None)
|
|
@@ -1,39 +0,0 @@
|
|
|
1
|
-
#! Food-101 image classification dataset
|
|
2
|
-
|
|
3
|
-
import os
|
|
4
|
-
|
|
5
|
-
from clarifai.data_upload.datasets.features import VisualClassificationFeatures
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
class Food101Dataset:
|
|
9
|
-
"""Food-101 Image Classification Dataset.
|
|
10
|
-
url: https://data.vision.ee.ethz.ch/cvl/datasets_extra/food-101/
|
|
11
|
-
"""
|
|
12
|
-
|
|
13
|
-
def __init__(self, split: str = "train"):
|
|
14
|
-
"""
|
|
15
|
-
Initialize dataset params.
|
|
16
|
-
Args:
|
|
17
|
-
data_dir: the local dataset directory.
|
|
18
|
-
split: "train" or "test"
|
|
19
|
-
"""
|
|
20
|
-
self.split = split
|
|
21
|
-
self.image_dir = {"train": os.path.join(os.path.dirname(__file__), "images")}
|
|
22
|
-
|
|
23
|
-
def dataloader(self):
|
|
24
|
-
"""
|
|
25
|
-
Transform food-101 dataset into clarifai proto compatible
|
|
26
|
-
format for upload.
|
|
27
|
-
Returns:
|
|
28
|
-
VisualClassificationFeatures type generator.
|
|
29
|
-
"""
|
|
30
|
-
## Your preprocessing code here
|
|
31
|
-
class_names = os.listdir(self.image_dir[self.split])
|
|
32
|
-
for class_name in class_names:
|
|
33
|
-
for image in os.listdir(os.path.join(self.image_dir[self.split], class_name)):
|
|
34
|
-
image_path = os.path.join(self.image_dir[self.split], class_name, image)
|
|
35
|
-
yield VisualClassificationFeatures(
|
|
36
|
-
image_path=image_path,
|
|
37
|
-
label=class_name,
|
|
38
|
-
id=None # or image_id
|
|
39
|
-
)
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/1036242.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/1114182.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/2012944.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/2464389.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/478632.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1061270.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1202261.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1381751.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/3289634.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/862025.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/102197.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/2749372.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/2938268.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/3590861.jpg
DELETED
|
Binary file
|
clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/746716.jpg
DELETED
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
File without changes
|
|
File without changes
|
|
@@ -1,39 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<folder>VOC2012</folder>
|
|
3
|
-
<filename>2007_000464.jpg</filename>
|
|
4
|
-
<source>
|
|
5
|
-
<database>The VOC2007 Database</database>
|
|
6
|
-
<annotation>PASCAL VOC2007</annotation>
|
|
7
|
-
<image>flickr</image>
|
|
8
|
-
</source>
|
|
9
|
-
<size>
|
|
10
|
-
<width>375</width>
|
|
11
|
-
<height>500</height>
|
|
12
|
-
<depth>3</depth>
|
|
13
|
-
</size>
|
|
14
|
-
<segmented>1</segmented>
|
|
15
|
-
<object>
|
|
16
|
-
<name>cow</name>
|
|
17
|
-
<pose>Left</pose>
|
|
18
|
-
<truncated>0</truncated>
|
|
19
|
-
<difficult>0</difficult>
|
|
20
|
-
<bndbox>
|
|
21
|
-
<xmin>71</xmin>
|
|
22
|
-
<ymin>252</ymin>
|
|
23
|
-
<xmax>216</xmax>
|
|
24
|
-
<ymax>314</ymax>
|
|
25
|
-
</bndbox>
|
|
26
|
-
</object>
|
|
27
|
-
<object>
|
|
28
|
-
<name>cow</name>
|
|
29
|
-
<pose>Left</pose>
|
|
30
|
-
<truncated>0</truncated>
|
|
31
|
-
<difficult>0</difficult>
|
|
32
|
-
<bndbox>
|
|
33
|
-
<xmin>58</xmin>
|
|
34
|
-
<ymin>202</ymin>
|
|
35
|
-
<xmax>241</xmax>
|
|
36
|
-
<ymax>295</ymax>
|
|
37
|
-
</bndbox>
|
|
38
|
-
</object>
|
|
39
|
-
</annotation>
|
|
@@ -1,28 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<folder>VOC2012</folder>
|
|
3
|
-
<filename>2008_000853.jpg</filename>
|
|
4
|
-
<source>
|
|
5
|
-
<database>The VOC2008 Database</database>
|
|
6
|
-
<annotation>PASCAL VOC2008</annotation>
|
|
7
|
-
<image>flickr</image>
|
|
8
|
-
</source>
|
|
9
|
-
<size>
|
|
10
|
-
<width>375</width>
|
|
11
|
-
<height>500</height>
|
|
12
|
-
<depth>3</depth>
|
|
13
|
-
</size>
|
|
14
|
-
<segmented>1</segmented>
|
|
15
|
-
<object>
|
|
16
|
-
<name>cat</name>
|
|
17
|
-
<pose>Frontal</pose>
|
|
18
|
-
<truncated>0</truncated>
|
|
19
|
-
<occluded>1</occluded>
|
|
20
|
-
<bndbox>
|
|
21
|
-
<xmin>37</xmin>
|
|
22
|
-
<ymin>345</ymin>
|
|
23
|
-
<xmax>186</xmax>
|
|
24
|
-
<ymax>417</ymax>
|
|
25
|
-
</bndbox>
|
|
26
|
-
<difficult>0</difficult>
|
|
27
|
-
</object>
|
|
28
|
-
</annotation>
|
|
@@ -1,54 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<folder>VOC2012</folder>
|
|
3
|
-
<filename>2008_003182.jpg</filename>
|
|
4
|
-
<source>
|
|
5
|
-
<database>The VOC2008 Database</database>
|
|
6
|
-
<annotation>PASCAL VOC2008</annotation>
|
|
7
|
-
<image>flickr</image>
|
|
8
|
-
</source>
|
|
9
|
-
<size>
|
|
10
|
-
<width>333</width>
|
|
11
|
-
<height>500</height>
|
|
12
|
-
<depth>3</depth>
|
|
13
|
-
</size>
|
|
14
|
-
<segmented>0</segmented>
|
|
15
|
-
<object>
|
|
16
|
-
<name>person</name>
|
|
17
|
-
<pose>Left</pose>
|
|
18
|
-
<truncated>1</truncated>
|
|
19
|
-
<occluded>0</occluded>
|
|
20
|
-
<bndbox>
|
|
21
|
-
<xmin>181</xmin>
|
|
22
|
-
<ymin>187</ymin>
|
|
23
|
-
<xmax>319</xmax>
|
|
24
|
-
<ymax>500</ymax>
|
|
25
|
-
</bndbox>
|
|
26
|
-
<difficult>0</difficult>
|
|
27
|
-
</object>
|
|
28
|
-
<object>
|
|
29
|
-
<name>horse</name>
|
|
30
|
-
<pose>Unspecified</pose>
|
|
31
|
-
<truncated>1</truncated>
|
|
32
|
-
<occluded>1</occluded>
|
|
33
|
-
<bndbox>
|
|
34
|
-
<xmin>71</xmin>
|
|
35
|
-
<ymin>181</ymin>
|
|
36
|
-
<xmax>333</xmax>
|
|
37
|
-
<ymax>500</ymax>
|
|
38
|
-
</bndbox>
|
|
39
|
-
<difficult>0</difficult>
|
|
40
|
-
</object>
|
|
41
|
-
<object>
|
|
42
|
-
<name>bottle</name>
|
|
43
|
-
<pose>Unspecified</pose>
|
|
44
|
-
<truncated>0</truncated>
|
|
45
|
-
<occluded>1</occluded>
|
|
46
|
-
<bndbox>
|
|
47
|
-
<xmin>211</xmin>
|
|
48
|
-
<ymin>312</ymin>
|
|
49
|
-
<xmax>235</xmax>
|
|
50
|
-
<ymax>358</ymax>
|
|
51
|
-
</bndbox>
|
|
52
|
-
<difficult>0</difficult>
|
|
53
|
-
</object>
|
|
54
|
-
</annotation>
|
|
@@ -1,67 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<folder>VOC2012</folder>
|
|
3
|
-
<filename>2008_008526.jpg</filename>
|
|
4
|
-
<source>
|
|
5
|
-
<database>The VOC2008 Database</database>
|
|
6
|
-
<annotation>PASCAL VOC2008</annotation>
|
|
7
|
-
<image>flickr</image>
|
|
8
|
-
</source>
|
|
9
|
-
<size>
|
|
10
|
-
<width>500</width>
|
|
11
|
-
<height>375</height>
|
|
12
|
-
<depth>3</depth>
|
|
13
|
-
</size>
|
|
14
|
-
<segmented>0</segmented>
|
|
15
|
-
<object>
|
|
16
|
-
<name>sofa</name>
|
|
17
|
-
<pose>Frontal</pose>
|
|
18
|
-
<truncated>1</truncated>
|
|
19
|
-
<occluded>1</occluded>
|
|
20
|
-
<bndbox>
|
|
21
|
-
<xmin>1</xmin>
|
|
22
|
-
<ymin>152</ymin>
|
|
23
|
-
<xmax>500</xmax>
|
|
24
|
-
<ymax>375</ymax>
|
|
25
|
-
</bndbox>
|
|
26
|
-
<difficult>0</difficult>
|
|
27
|
-
</object>
|
|
28
|
-
<object>
|
|
29
|
-
<name>bottle</name>
|
|
30
|
-
<pose>Unspecified</pose>
|
|
31
|
-
<truncated>0</truncated>
|
|
32
|
-
<occluded>1</occluded>
|
|
33
|
-
<bndbox>
|
|
34
|
-
<xmin>162</xmin>
|
|
35
|
-
<ymin>232</ymin>
|
|
36
|
-
<xmax>197</xmax>
|
|
37
|
-
<ymax>338</ymax>
|
|
38
|
-
</bndbox>
|
|
39
|
-
<difficult>0</difficult>
|
|
40
|
-
</object>
|
|
41
|
-
<object>
|
|
42
|
-
<name>person</name>
|
|
43
|
-
<pose>Frontal</pose>
|
|
44
|
-
<truncated>1</truncated>
|
|
45
|
-
<occluded>1</occluded>
|
|
46
|
-
<bndbox>
|
|
47
|
-
<xmin>66</xmin>
|
|
48
|
-
<ymin>138</ymin>
|
|
49
|
-
<xmax>352</xmax>
|
|
50
|
-
<ymax>375</ymax>
|
|
51
|
-
</bndbox>
|
|
52
|
-
<difficult>0</difficult>
|
|
53
|
-
</object>
|
|
54
|
-
<object>
|
|
55
|
-
<name>person</name>
|
|
56
|
-
<pose>Frontal</pose>
|
|
57
|
-
<truncated>1</truncated>
|
|
58
|
-
<occluded>1</occluded>
|
|
59
|
-
<bndbox>
|
|
60
|
-
<xmin>234</xmin>
|
|
61
|
-
<ymin>141</ymin>
|
|
62
|
-
<xmax>500</xmax>
|
|
63
|
-
<ymax>375</ymax>
|
|
64
|
-
</bndbox>
|
|
65
|
-
<difficult>0</difficult>
|
|
66
|
-
</object>
|
|
67
|
-
</annotation>
|
|
@@ -1,28 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<filename>2009_004315.jpg</filename>
|
|
3
|
-
<folder>VOC2012</folder>
|
|
4
|
-
<object>
|
|
5
|
-
<name>bird</name>
|
|
6
|
-
<bndbox>
|
|
7
|
-
<xmax>461</xmax>
|
|
8
|
-
<xmin>219</xmin>
|
|
9
|
-
<ymax>327</ymax>
|
|
10
|
-
<ymin>51</ymin>
|
|
11
|
-
</bndbox>
|
|
12
|
-
<difficult>0</difficult>
|
|
13
|
-
<occluded>0</occluded>
|
|
14
|
-
<pose>Left</pose>
|
|
15
|
-
<truncated>0</truncated>
|
|
16
|
-
</object>
|
|
17
|
-
<segmented>0</segmented>
|
|
18
|
-
<size>
|
|
19
|
-
<depth>3</depth>
|
|
20
|
-
<height>463</height>
|
|
21
|
-
<width>500</width>
|
|
22
|
-
</size>
|
|
23
|
-
<source>
|
|
24
|
-
<annotation>PASCAL VOC2009</annotation>
|
|
25
|
-
<database>The VOC2009 Database</database>
|
|
26
|
-
<image>flickr</image>
|
|
27
|
-
</source>
|
|
28
|
-
</annotation>
|
|
@@ -1,28 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<filename>2009_004382.jpg</filename>
|
|
3
|
-
<folder>VOC2012</folder>
|
|
4
|
-
<object>
|
|
5
|
-
<name>cat</name>
|
|
6
|
-
<bndbox>
|
|
7
|
-
<xmax>264</xmax>
|
|
8
|
-
<xmin>50</xmin>
|
|
9
|
-
<ymax>500</ymax>
|
|
10
|
-
<ymin>1</ymin>
|
|
11
|
-
</bndbox>
|
|
12
|
-
<difficult>0</difficult>
|
|
13
|
-
<occluded>0</occluded>
|
|
14
|
-
<pose>Frontal</pose>
|
|
15
|
-
<truncated>0</truncated>
|
|
16
|
-
</object>
|
|
17
|
-
<segmented>0</segmented>
|
|
18
|
-
<size>
|
|
19
|
-
<depth>3</depth>
|
|
20
|
-
<height>500</height>
|
|
21
|
-
<width>313</width>
|
|
22
|
-
</size>
|
|
23
|
-
<source>
|
|
24
|
-
<annotation>PASCAL VOC2009</annotation>
|
|
25
|
-
<database>The VOC2009 Database</database>
|
|
26
|
-
<image>flickr</image>
|
|
27
|
-
</source>
|
|
28
|
-
</annotation>
|
|
@@ -1,28 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<filename>2011_000430.jpg</filename>
|
|
3
|
-
<folder>VOC2012</folder>
|
|
4
|
-
<object>
|
|
5
|
-
<name>dog</name>
|
|
6
|
-
<bndbox>
|
|
7
|
-
<xmax>500</xmax>
|
|
8
|
-
<xmin>17</xmin>
|
|
9
|
-
<ymax>365</ymax>
|
|
10
|
-
<ymin>33</ymin>
|
|
11
|
-
</bndbox>
|
|
12
|
-
<difficult>0</difficult>
|
|
13
|
-
<occluded>0</occluded>
|
|
14
|
-
<pose>Left</pose>
|
|
15
|
-
<truncated>1</truncated>
|
|
16
|
-
</object>
|
|
17
|
-
<segmented>0</segmented>
|
|
18
|
-
<size>
|
|
19
|
-
<depth>3</depth>
|
|
20
|
-
<height>375</height>
|
|
21
|
-
<width>500</width>
|
|
22
|
-
</size>
|
|
23
|
-
<source>
|
|
24
|
-
<annotation>PASCAL VOC2011</annotation>
|
|
25
|
-
<database>The VOC2011 Database</database>
|
|
26
|
-
<image>flickr</image>
|
|
27
|
-
</source>
|
|
28
|
-
</annotation>
|
|
@@ -1,46 +0,0 @@
|
|
|
1
|
-
<annotation>
|
|
2
|
-
<filename>2011_001610.jpg</filename>
|
|
3
|
-
<folder>VOC2012</folder>
|
|
4
|
-
<object>
|
|
5
|
-
<name>person</name>
|
|
6
|
-
<bndbox>
|
|
7
|
-
<xmax>283</xmax>
|
|
8
|
-
<xmin>113</xmin>
|
|
9
|
-
<ymax>390</ymax>
|
|
10
|
-
<ymin>112</ymin>
|
|
11
|
-
</bndbox>
|
|
12
|
-
<difficult>0</difficult>
|
|
13
|
-
<occluded>1</occluded>
|
|
14
|
-
<pose>Frontal</pose>
|
|
15
|
-
<truncated>1</truncated>
|
|
16
|
-
<part>
|
|
17
|
-
<name>hand</name>
|
|
18
|
-
<bndbox>
|
|
19
|
-
<xmin>128</xmin>
|
|
20
|
-
<ymin>231</ymin>
|
|
21
|
-
<xmax>183</xmax>
|
|
22
|
-
<ymax>283</ymax>
|
|
23
|
-
</bndbox>
|
|
24
|
-
</part>
|
|
25
|
-
<part>
|
|
26
|
-
<name>head</name>
|
|
27
|
-
<bndbox>
|
|
28
|
-
<xmin>123</xmin>
|
|
29
|
-
<ymin>115</ymin>
|
|
30
|
-
<xmax>209</xmax>
|
|
31
|
-
<ymax>226</ymax>
|
|
32
|
-
</bndbox>
|
|
33
|
-
</part>
|
|
34
|
-
</object>
|
|
35
|
-
<segmented>0</segmented>
|
|
36
|
-
<size>
|
|
37
|
-
<depth>3</depth>
|
|
38
|
-
<height>500</height>
|
|
39
|
-
<width>375</width>
|
|
40
|
-
</size>
|
|
41
|
-
<source>
|
|
42
|
-
<annotation>PASCAL VOC2011</annotation>
|
|
43
|
-
<database>The VOC2011 Database</database>
|
|
44
|
-
<image>flickr</image>
|
|
45
|
-
</source>
|
|
46
|
-
</annotation>
|