clarifai 9.10.1__py3-none-any.whl → 9.10.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- clarifai/client/__init__.py +3 -2
- clarifai/client/app.py +39 -23
- clarifai/client/base.py +6 -6
- clarifai/client/dataset.py +113 -55
- clarifai/client/input.py +47 -55
- clarifai/client/model.py +27 -25
- clarifai/client/module.py +13 -11
- clarifai/client/runner.py +5 -3
- clarifai/client/search.py +29 -10
- clarifai/client/user.py +14 -8
- clarifai/client/workflow.py +22 -20
- clarifai/constants/dataset.py +22 -0
- clarifai/datasets/upload/base.py +9 -7
- clarifai/datasets/upload/features.py +3 -3
- clarifai/datasets/upload/image.py +49 -50
- clarifai/datasets/upload/loaders/coco_captions.py +26 -80
- clarifai/datasets/upload/loaders/coco_detection.py +56 -115
- clarifai/datasets/upload/loaders/coco_segmentation.py +69 -137
- clarifai/datasets/upload/loaders/imagenet_classification.py +2 -3
- clarifai/datasets/upload/loaders/xview_detection.py +3 -3
- clarifai/datasets/upload/text.py +16 -16
- clarifai/datasets/upload/utils.py +196 -21
- clarifai/utils/misc.py +21 -0
- clarifai/versions.py +1 -1
- {clarifai-9.10.1.dist-info → clarifai-9.10.3.dist-info}/METADATA +3 -3
- clarifai-9.10.3.dist-info/RECORD +96 -0
- clarifai-9.10.3.dist-info/top_level.txt +1 -0
- clarifai/auth/__init__.py +0 -6
- clarifai/auth/helper.py +0 -367
- clarifai/auth/register.py +0 -23
- clarifai/auth/stub.py +0 -127
- clarifai/datasets/upload/examples/README.md +0 -31
- clarifai/datasets/upload/examples/image_classification/__init__.py +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/__init__.py +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/cifar_small_test.csv +0 -10
- clarifai/datasets/upload/examples/image_classification/cifar10/cifar_small_train.csv +0 -10
- clarifai/datasets/upload/examples/image_classification/cifar10/dataset.py +0 -46
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_700.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_701.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_702.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_703.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_704.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_705.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_706.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_707.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_708.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/cifar10/images/test_batch_709.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/__init__.py +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/dataset.py +0 -39
- clarifai/datasets/upload/examples/image_classification/food-101/images/beignets/1420783.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/beignets/3287885.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/beignets/3617075.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/beignets/38052.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/beignets/39147.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/hamburger/139558.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/hamburger/1636096.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/hamburger/2480925.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/hamburger/3385808.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/hamburger/3647386.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/prime_rib/1826869.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/prime_rib/2243245.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/prime_rib/259212.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/prime_rib/2842688.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/prime_rib/3035414.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/ramen/1545393.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/ramen/2427642.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/ramen/3520891.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/ramen/377566.jpg +0 -0
- clarifai/datasets/upload/examples/image_classification/food-101/images/ramen/503504.jpg +0 -0
- clarifai/datasets/upload/examples/text_classification/__init__.py +0 -0
- clarifai/datasets/upload/examples/text_classification/imdb_dataset/__init__.py +0 -0
- clarifai/datasets/upload/examples/text_classification/imdb_dataset/dataset.py +0 -42
- clarifai/datasets/upload/examples/text_classification/imdb_dataset/test.csv +0 -201
- clarifai/datasets/upload/examples/text_classification/imdb_dataset/train.csv +0 -201
- clarifai/datasets/upload/loaders/README.md +0 -49
- clarifai/models/model_serving/README.md +0 -155
- clarifai/models/model_serving/docs/custom_config.md +0 -33
- clarifai/models/model_serving/docs/dependencies.md +0 -11
- clarifai/models/model_serving/docs/inference_parameters.md +0 -134
- clarifai/models/model_serving/docs/model_types.md +0 -20
- clarifai/models/model_serving/docs/output.md +0 -28
- clarifai/models/model_serving/examples/README.md +0 -7
- clarifai/models/model_serving/examples/image_classification/README.md +0 -9
- clarifai/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/README.md +0 -11
- clarifai/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/config.json +0 -42
- clarifai/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/preprocessor_config.json +0 -15
- clarifai/models/model_serving/examples/image_classification/age_vit/config.pbtxt +0 -23
- clarifai/models/model_serving/examples/image_classification/age_vit/labels.txt +0 -9
- clarifai/models/model_serving/examples/image_classification/age_vit/requirements.txt +0 -7
- clarifai/models/model_serving/examples/text_classification/README.md +0 -9
- clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/README.md +0 -12
- clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/config.json +0 -34
- clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/special_tokens_map.json +0 -1
- clarifai/models/model_serving/examples/text_classification/xlm-roberta/config.pbtxt +0 -21
- clarifai/models/model_serving/examples/text_classification/xlm-roberta/labels.txt +0 -3
- clarifai/models/model_serving/examples/text_classification/xlm-roberta/requirements.txt +0 -7
- clarifai/models/model_serving/examples/text_embedding/README.md +0 -9
- clarifai/models/model_serving/examples/text_to_image/README.md +0 -9
- clarifai/models/model_serving/examples/text_to_image/sd-v1.5/1/__init__.py +0 -0
- clarifai/models/model_serving/examples/text_to_image/sd-v1.5/1/inference.py +0 -52
- clarifai/models/model_serving/examples/text_to_image/sd-v1.5/1/model.py +0 -60
- clarifai/models/model_serving/examples/text_to_image/sd-v1.5/config.pbtxt +0 -22
- clarifai/models/model_serving/examples/text_to_image/sd-v1.5/requirements.txt +0 -6
- clarifai/models/model_serving/examples/text_to_text/README.md +0 -10
- clarifai/models/model_serving/examples/text_to_text/bart-summarize/config.pbtxt +0 -20
- clarifai/models/model_serving/examples/text_to_text/bart-summarize/requirements.txt +0 -4
- clarifai/models/model_serving/examples/visual_detection/README.md +0 -11
- clarifai/models/model_serving/examples/visual_detection/yolov5x/config.pbtxt +0 -36
- clarifai/models/model_serving/examples/visual_detection/yolov5x/labels.txt +0 -80
- clarifai/models/model_serving/examples/visual_detection/yolov5x/requirements.txt +0 -12
- clarifai/models/model_serving/examples/visual_embedding/README.md +0 -9
- clarifai/models/model_serving/examples/visual_embedding/vit-base/config.pbtxt +0 -22
- clarifai/models/model_serving/examples/visual_embedding/vit-base/requirements.txt +0 -5
- clarifai/models/model_serving/examples/visual_segmentation/README.md +0 -9
- clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/config.pbtxt +0 -24
- clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/labels.txt +0 -18
- clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/requirements.txt +0 -5
- clarifai/models/model_serving/model_config/model_types_config/multimodal-embedder.yaml +0 -24
- clarifai/models/model_serving/model_config/model_types_config/text-classifier.yaml +0 -18
- clarifai/models/model_serving/model_config/model_types_config/text-embedder.yaml +0 -18
- clarifai/models/model_serving/model_config/model_types_config/text-to-image.yaml +0 -18
- clarifai/models/model_serving/model_config/model_types_config/text-to-text.yaml +0 -18
- clarifai/models/model_serving/model_config/model_types_config/visual-classifier.yaml +0 -18
- clarifai/models/model_serving/model_config/model_types_config/visual-detector.yaml +0 -28
- clarifai/models/model_serving/model_config/model_types_config/visual-embedder.yaml +0 -18
- clarifai/models/model_serving/model_config/model_types_config/visual-segmenter.yaml +0 -18
- clarifai/modules/README.md +0 -5
- clarifai/modules/style.css +0 -217
- clarifai-9.10.1.dist-info/RECORD +0 -386
- clarifai-9.10.1.dist-info/top_level.txt +0 -2
- clarifai_utils/__init__.py +0 -0
- clarifai_utils/auth/__init__.py +0 -6
- clarifai_utils/auth/helper.py +0 -367
- clarifai_utils/auth/register.py +0 -23
- clarifai_utils/auth/stub.py +0 -127
- clarifai_utils/cli.py +0 -0
- clarifai_utils/client/__init__.py +0 -16
- clarifai_utils/client/app.py +0 -684
- clarifai_utils/client/auth/__init__.py +0 -4
- clarifai_utils/client/auth/helper.py +0 -367
- clarifai_utils/client/auth/register.py +0 -23
- clarifai_utils/client/auth/stub.py +0 -127
- clarifai_utils/client/base.py +0 -131
- clarifai_utils/client/dataset.py +0 -442
- clarifai_utils/client/input.py +0 -892
- clarifai_utils/client/lister.py +0 -54
- clarifai_utils/client/model.py +0 -575
- clarifai_utils/client/module.py +0 -94
- clarifai_utils/client/runner.py +0 -161
- clarifai_utils/client/search.py +0 -239
- clarifai_utils/client/user.py +0 -253
- clarifai_utils/client/workflow.py +0 -223
- clarifai_utils/constants/model.py +0 -4
- clarifai_utils/constants/search.py +0 -2
- clarifai_utils/datasets/__init__.py +0 -0
- clarifai_utils/datasets/export/__init__.py +0 -0
- clarifai_utils/datasets/export/inputs_annotations.py +0 -222
- clarifai_utils/datasets/upload/__init__.py +0 -0
- clarifai_utils/datasets/upload/base.py +0 -66
- clarifai_utils/datasets/upload/examples/README.md +0 -31
- clarifai_utils/datasets/upload/examples/image_classification/__init__.py +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/__init__.py +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/cifar_small_test.csv +0 -10
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/cifar_small_train.csv +0 -10
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/dataset.py +0 -46
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_700.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_701.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_702.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_703.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_704.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_705.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_706.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_707.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_708.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_709.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/__init__.py +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/dataset.py +0 -39
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/1420783.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/3287885.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/3617075.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/38052.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/39147.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/139558.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/1636096.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/2480925.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/3385808.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/3647386.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/1826869.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/2243245.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/259212.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/2842688.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/3035414.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/1545393.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/2427642.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/3520891.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/377566.jpg +0 -0
- clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/503504.jpg +0 -0
- clarifai_utils/datasets/upload/examples/text_classification/__init__.py +0 -0
- clarifai_utils/datasets/upload/examples/text_classification/imdb_dataset/__init__.py +0 -0
- clarifai_utils/datasets/upload/examples/text_classification/imdb_dataset/dataset.py +0 -42
- clarifai_utils/datasets/upload/examples/text_classification/imdb_dataset/test.csv +0 -201
- clarifai_utils/datasets/upload/examples/text_classification/imdb_dataset/train.csv +0 -201
- clarifai_utils/datasets/upload/features.py +0 -44
- clarifai_utils/datasets/upload/image.py +0 -165
- clarifai_utils/datasets/upload/loaders/README.md +0 -49
- clarifai_utils/datasets/upload/loaders/__init__.py +0 -0
- clarifai_utils/datasets/upload/loaders/coco_captions.py +0 -103
- clarifai_utils/datasets/upload/loaders/coco_detection.py +0 -134
- clarifai_utils/datasets/upload/loaders/coco_segmentation.py +0 -166
- clarifai_utils/datasets/upload/loaders/imagenet_classification.py +0 -59
- clarifai_utils/datasets/upload/loaders/xview_detection.py +0 -148
- clarifai_utils/datasets/upload/text.py +0 -53
- clarifai_utils/datasets/upload/utils.py +0 -63
- clarifai_utils/errors.py +0 -89
- clarifai_utils/models/__init__.py +0 -0
- clarifai_utils/models/api.py +0 -283
- clarifai_utils/models/model_serving/README.md +0 -155
- clarifai_utils/models/model_serving/__init__.py +0 -12
- clarifai_utils/models/model_serving/cli/__init__.py +0 -12
- clarifai_utils/models/model_serving/cli/deploy_cli.py +0 -123
- clarifai_utils/models/model_serving/cli/model_zip.py +0 -61
- clarifai_utils/models/model_serving/cli/repository.py +0 -87
- clarifai_utils/models/model_serving/constants.py +0 -1
- clarifai_utils/models/model_serving/docs/custom_config.md +0 -33
- clarifai_utils/models/model_serving/docs/dependencies.md +0 -11
- clarifai_utils/models/model_serving/docs/inference_parameters.md +0 -134
- clarifai_utils/models/model_serving/docs/model_types.md +0 -20
- clarifai_utils/models/model_serving/docs/output.md +0 -28
- clarifai_utils/models/model_serving/examples/README.md +0 -7
- clarifai_utils/models/model_serving/examples/image_classification/README.md +0 -9
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/__init__.py +0 -0
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/inference.py +0 -56
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/model.py +0 -61
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/README.md +0 -11
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/config.json +0 -42
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/preprocessor_config.json +0 -15
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/config.pbtxt +0 -23
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/labels.txt +0 -9
- clarifai_utils/models/model_serving/examples/image_classification/age_vit/requirements.txt +0 -7
- clarifai_utils/models/model_serving/examples/text_classification/README.md +0 -9
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/__init__.py +0 -0
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/inference.py +0 -55
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/model.py +0 -61
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/README.md +0 -12
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/config.json +0 -34
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/special_tokens_map.json +0 -1
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/config.pbtxt +0 -21
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/labels.txt +0 -3
- clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/requirements.txt +0 -7
- clarifai_utils/models/model_serving/examples/text_embedding/README.md +0 -9
- clarifai_utils/models/model_serving/examples/text_to_image/README.md +0 -9
- clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/1/__init__.py +0 -0
- clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/1/inference.py +0 -52
- clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/1/model.py +0 -60
- clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/config.pbtxt +0 -22
- clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/requirements.txt +0 -6
- clarifai_utils/models/model_serving/examples/text_to_text/README.md +0 -10
- clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/1/__init__.py +0 -0
- clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/1/inference.py +0 -47
- clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/1/model.py +0 -60
- clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/config.pbtxt +0 -20
- clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/requirements.txt +0 -4
- clarifai_utils/models/model_serving/examples/visual_detection/README.md +0 -11
- clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/1/inference.py +0 -72
- clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/1/model.py +0 -61
- clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/config.pbtxt +0 -36
- clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/labels.txt +0 -80
- clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/requirements.txt +0 -12
- clarifai_utils/models/model_serving/examples/visual_embedding/README.md +0 -9
- clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/1/__init__.py +0 -0
- clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/1/inference.py +0 -51
- clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/1/model.py +0 -60
- clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/config.pbtxt +0 -22
- clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/requirements.txt +0 -5
- clarifai_utils/models/model_serving/examples/visual_segmentation/README.md +0 -9
- clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/1/__init__.py +0 -0
- clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/1/inference.py +0 -55
- clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/1/model.py +0 -60
- clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/config.pbtxt +0 -24
- clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/labels.txt +0 -18
- clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/requirements.txt +0 -5
- clarifai_utils/models/model_serving/model_config/__init__.py +0 -14
- clarifai_utils/models/model_serving/model_config/config.py +0 -302
- clarifai_utils/models/model_serving/model_config/inference_parameter.py +0 -124
- clarifai_utils/models/model_serving/model_config/model_types_config/multimodal-embedder.yaml +0 -24
- clarifai_utils/models/model_serving/model_config/model_types_config/text-classifier.yaml +0 -18
- clarifai_utils/models/model_serving/model_config/model_types_config/text-embedder.yaml +0 -18
- clarifai_utils/models/model_serving/model_config/model_types_config/text-to-image.yaml +0 -18
- clarifai_utils/models/model_serving/model_config/model_types_config/text-to-text.yaml +0 -18
- clarifai_utils/models/model_serving/model_config/model_types_config/visual-classifier.yaml +0 -18
- clarifai_utils/models/model_serving/model_config/model_types_config/visual-detector.yaml +0 -28
- clarifai_utils/models/model_serving/model_config/model_types_config/visual-embedder.yaml +0 -18
- clarifai_utils/models/model_serving/model_config/model_types_config/visual-segmenter.yaml +0 -18
- clarifai_utils/models/model_serving/model_config/serializer.py +0 -134
- clarifai_utils/models/model_serving/models/__init__.py +0 -12
- clarifai_utils/models/model_serving/models/default_test.py +0 -275
- clarifai_utils/models/model_serving/models/inference.py +0 -42
- clarifai_utils/models/model_serving/models/model_types.py +0 -265
- clarifai_utils/models/model_serving/models/output.py +0 -124
- clarifai_utils/models/model_serving/models/pb_model.py +0 -74
- clarifai_utils/models/model_serving/models/test.py +0 -64
- clarifai_utils/models/model_serving/pb_model_repository.py +0 -101
- clarifai_utils/modules/README.md +0 -5
- clarifai_utils/modules/__init__.py +0 -0
- clarifai_utils/modules/css.py +0 -60
- clarifai_utils/modules/pages.py +0 -42
- clarifai_utils/modules/style.css +0 -217
- clarifai_utils/runners/__init__.py +0 -0
- clarifai_utils/runners/example.py +0 -33
- clarifai_utils/schema/search.py +0 -69
- clarifai_utils/urls/helper.py +0 -103
- clarifai_utils/utils/__init__.py +0 -0
- clarifai_utils/utils/logging.py +0 -90
- clarifai_utils/utils/misc.py +0 -33
- clarifai_utils/utils/model_train.py +0 -157
- clarifai_utils/versions.py +0 -6
- clarifai_utils/workflows/__init__.py +0 -0
- clarifai_utils/workflows/export.py +0 -68
- clarifai_utils/workflows/utils.py +0 -59
- clarifai_utils/workflows/validate.py +0 -67
- {clarifai-9.10.1.dist-info → clarifai-9.10.3.dist-info}/LICENSE +0 -0
- {clarifai-9.10.1.dist-info → clarifai-9.10.3.dist-info}/WHEEL +0 -0
- {clarifai-9.10.1.dist-info → clarifai-9.10.3.dist-info}/entry_points.txt +0 -0
|
@@ -1,222 +0,0 @@
|
|
|
1
|
-
import json
|
|
2
|
-
import os
|
|
3
|
-
import tempfile
|
|
4
|
-
import zipfile
|
|
5
|
-
from concurrent.futures import ThreadPoolExecutor, as_completed
|
|
6
|
-
from io import BytesIO
|
|
7
|
-
from typing import Any, Dict, Iterator, List, Optional
|
|
8
|
-
|
|
9
|
-
import requests
|
|
10
|
-
from clarifai_grpc.grpc.api import resources_pb2
|
|
11
|
-
from google.protobuf.json_format import MessageToDict
|
|
12
|
-
from PIL import ImageFile
|
|
13
|
-
from tqdm import tqdm
|
|
14
|
-
|
|
15
|
-
from clarifai.errors import UserError
|
|
16
|
-
from clarifai.utils.logging import get_logger
|
|
17
|
-
|
|
18
|
-
logger = get_logger("INFO", __name__)
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
class DatasetExportReader:
|
|
22
|
-
|
|
23
|
-
def __init__(self,
|
|
24
|
-
session: requests.Session,
|
|
25
|
-
archive_url: Optional[str] = None,
|
|
26
|
-
local_archive_path: Optional[str] = None):
|
|
27
|
-
"""Download/Reads the zipfile archive and yields every api.Input object.
|
|
28
|
-
|
|
29
|
-
Args:
|
|
30
|
-
session: requests.Session object
|
|
31
|
-
archive_url: URL of the DatasetVersionExport archive
|
|
32
|
-
local_archive_path: Path to the DatasetVersionExport archive
|
|
33
|
-
"""
|
|
34
|
-
self.input_count = 0
|
|
35
|
-
self.temp_file = None
|
|
36
|
-
self.session = session
|
|
37
|
-
|
|
38
|
-
assert archive_url or local_archive_path, UserError(
|
|
39
|
-
"Either archive_url or local_archive_path must be provided.")
|
|
40
|
-
assert not (archive_url and local_archive_path), UserError(
|
|
41
|
-
"Only one of archive_url or local_archive_path must be provided.")
|
|
42
|
-
|
|
43
|
-
if archive_url:
|
|
44
|
-
logger.info('url: %s' % archive_url)
|
|
45
|
-
self.temp_file = self._download_temp_archive(archive_url)
|
|
46
|
-
self.archive = zipfile.ZipFile(self.temp_file)
|
|
47
|
-
else:
|
|
48
|
-
logger.info("path: %s" % local_archive_path)
|
|
49
|
-
self.archive = zipfile.ZipFile(local_archive_path)
|
|
50
|
-
|
|
51
|
-
self.file_name_list = self.archive.namelist()
|
|
52
|
-
assert "mimetype" in self.file_name_list, "Missing mimetype file in the dataset export archive."
|
|
53
|
-
assert self.archive.read("mimetype") == b"application/x.clarifai-data+protobuf"
|
|
54
|
-
self.file_name_list.remove("mimetype")
|
|
55
|
-
|
|
56
|
-
logger.info("Obtained file name list. %d entries." % len(self.file_name_list))
|
|
57
|
-
self.split_dir = os.path.dirname(self.file_name_list[0]) if len(self.file_name_list) else ""
|
|
58
|
-
|
|
59
|
-
def _download_temp_archive(self, archive_url: str,
|
|
60
|
-
chunk_size: int = 128) -> tempfile.TemporaryFile:
|
|
61
|
-
"""Downloads the temp archive of InputBatches."""
|
|
62
|
-
r = self.session.get(archive_url, stream=True)
|
|
63
|
-
temp_file = tempfile.TemporaryFile()
|
|
64
|
-
for chunk in r.iter_content(chunk_size=chunk_size):
|
|
65
|
-
temp_file.write(chunk)
|
|
66
|
-
|
|
67
|
-
return temp_file
|
|
68
|
-
|
|
69
|
-
def __len__(self) -> int:
|
|
70
|
-
if not self.input_count:
|
|
71
|
-
if self.file_name_list is not None:
|
|
72
|
-
for filename in self.file_name_list:
|
|
73
|
-
self.input_count += int(filename.split('_n')[-1])
|
|
74
|
-
|
|
75
|
-
return self.input_count
|
|
76
|
-
|
|
77
|
-
def __iter__(self) -> Iterator[resources_pb2.Input]:
|
|
78
|
-
"""Loops through all InputBatches in the DatasetVersionExport and yields every api.Input object"""
|
|
79
|
-
if self.file_name_list is not None:
|
|
80
|
-
for filename in self.file_name_list:
|
|
81
|
-
db = resources_pb2.InputBatch().FromString(self.archive.read(filename))
|
|
82
|
-
for db_input in db.inputs:
|
|
83
|
-
yield db_input
|
|
84
|
-
|
|
85
|
-
def __enter__(self) -> 'DatasetExportReader':
|
|
86
|
-
return self
|
|
87
|
-
|
|
88
|
-
def __exit__(self, *args: Any) -> None:
|
|
89
|
-
self.close()
|
|
90
|
-
|
|
91
|
-
def close(self) -> None:
|
|
92
|
-
logger.info("closing file objects.")
|
|
93
|
-
self.archive.close()
|
|
94
|
-
if self.temp_file:
|
|
95
|
-
self.temp_file.close()
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
class InputAnnotationDownloader:
|
|
99
|
-
|
|
100
|
-
def __init__(self,
|
|
101
|
-
session: requests.Session,
|
|
102
|
-
input_iterator: DatasetExportReader,
|
|
103
|
-
num_workers: int = 4):
|
|
104
|
-
"""Downloads the archive from the URL into an archive of inputs, annotations in the directory format
|
|
105
|
-
{split}/inputs and {split}/annotations.
|
|
106
|
-
|
|
107
|
-
Args:
|
|
108
|
-
session: requests.Session object
|
|
109
|
-
input_iterator: Iterable of DatasetExportReader object
|
|
110
|
-
num_workers: Number of threads to use for downloading
|
|
111
|
-
"""
|
|
112
|
-
self.input_iterator = input_iterator
|
|
113
|
-
self.num_workers = min(num_workers, 10) # Max 10 threads
|
|
114
|
-
self.num_inputs_annotations = 0
|
|
115
|
-
self.split_prefix = None
|
|
116
|
-
self.session = session
|
|
117
|
-
self.input_ext = dict(image=".png", text=".txt", audio=".mp3", video=".mp4")
|
|
118
|
-
if isinstance(self.input_iterator, DatasetExportReader):
|
|
119
|
-
self.split_prefix = self.input_iterator.split_dir
|
|
120
|
-
|
|
121
|
-
def _save_image_to_archive(self, new_archive: zipfile.ZipFile, hosted_url: str,
|
|
122
|
-
file_name: str) -> None:
|
|
123
|
-
"""Use PIL ImageFile to return image parsed from the response bytestring (from requests) and append to zip file."""
|
|
124
|
-
p = ImageFile.Parser()
|
|
125
|
-
p.feed(self.session.get(hosted_url).content)
|
|
126
|
-
image = p.close()
|
|
127
|
-
image_file = BytesIO()
|
|
128
|
-
image.save(image_file, 'PNG')
|
|
129
|
-
new_archive.writestr(file_name, image_file.getvalue())
|
|
130
|
-
|
|
131
|
-
def _save_text_to_archive(self, new_archive: zipfile.ZipFile, hosted_url: str,
|
|
132
|
-
file_name: str) -> None:
|
|
133
|
-
"""Gets the text response bytestring (from requests) and append to zip file."""
|
|
134
|
-
text_content = self.session.get(hosted_url).content
|
|
135
|
-
new_archive.writestr(file_name, text_content)
|
|
136
|
-
|
|
137
|
-
def _save_audio_to_archive(self, new_archive: zipfile.ZipFile, hosted_url: str,
|
|
138
|
-
file_name: str) -> None:
|
|
139
|
-
"""Gets the audio response bytestring (from requests) as chunks and append to zip file."""
|
|
140
|
-
audio_response = self.session.get(hosted_url, stream=True)
|
|
141
|
-
audio_stream = BytesIO()
|
|
142
|
-
# Retrieve the audio content in chunks and write to the BytesIO object
|
|
143
|
-
for chunk in audio_response.iter_content(chunk_size=128):
|
|
144
|
-
audio_stream.write(chunk)
|
|
145
|
-
new_archive.writestr(file_name, audio_stream.getvalue())
|
|
146
|
-
|
|
147
|
-
def _save_video_to_archive(self, new_archive: zipfile.ZipFile, hosted_url: str,
|
|
148
|
-
file_name: str) -> None:
|
|
149
|
-
"""Gets the video response bytestring (from requests) as chunks and append to zip file."""
|
|
150
|
-
video_response = self.session.get(hosted_url)
|
|
151
|
-
video_stream = BytesIO()
|
|
152
|
-
# Retrieve the video content in chunks and write to the BytesIO object
|
|
153
|
-
for chunk in video_response.iter_content(chunk_size=128):
|
|
154
|
-
video_stream.write(chunk)
|
|
155
|
-
new_archive.writestr(file_name, video_stream.getvalue())
|
|
156
|
-
|
|
157
|
-
def _save_annotation_to_archive(self, new_archive: zipfile.ZipFile, annot_data: List[Dict],
|
|
158
|
-
file_name: str) -> None:
|
|
159
|
-
"""Gets the annotation response bytestring (from requests) and append to zip file."""
|
|
160
|
-
# Serialize the dictionary to a JSON string
|
|
161
|
-
json_str = json.dumps(annot_data)
|
|
162
|
-
# Convert the JSON string to bytes
|
|
163
|
-
bytes_object = json_str.encode()
|
|
164
|
-
|
|
165
|
-
new_archive.writestr(file_name, bytes_object)
|
|
166
|
-
|
|
167
|
-
def _write_archive(self, input_, new_archive, split: Optional[str] = None) -> None:
|
|
168
|
-
"""Writes the input, annotation archive into prefix dir."""
|
|
169
|
-
data_dict = MessageToDict(input_.data)
|
|
170
|
-
input_type = list(filter(lambda x: x in list(data_dict.keys()),
|
|
171
|
-
list(self.input_ext.keys())))[0]
|
|
172
|
-
hosted = getattr(input_.data, input_type).hosted
|
|
173
|
-
if hosted.prefix:
|
|
174
|
-
assert 'orig' in hosted.sizes
|
|
175
|
-
hosted_url = f"{hosted.prefix}/orig/{hosted.suffix}"
|
|
176
|
-
file_name = os.path.join(split, "inputs", input_.id + self.input_ext[input_type])
|
|
177
|
-
if input_type == "image":
|
|
178
|
-
self._save_image_to_archive(new_archive, hosted_url, file_name)
|
|
179
|
-
elif input_type == "text":
|
|
180
|
-
self._save_text_to_archive(new_archive, hosted_url, file_name)
|
|
181
|
-
elif input_type == "audio":
|
|
182
|
-
self._save_audio_to_archive(new_archive, hosted_url, file_name)
|
|
183
|
-
elif input_type == "video":
|
|
184
|
-
self._save_video_to_archive(new_archive, hosted_url, file_name)
|
|
185
|
-
self.num_inputs_annotations += 1
|
|
186
|
-
|
|
187
|
-
if data_dict.get("concepts") or data_dict.get("regions"):
|
|
188
|
-
file_name = os.path.join(split, "annotations", input_.id + ".json")
|
|
189
|
-
annot_data = data_dict.get("concepts") or data_dict.get("regions")
|
|
190
|
-
|
|
191
|
-
self._save_annotation_to_archive(new_archive, annot_data, file_name)
|
|
192
|
-
self.num_inputs_annotations += 1
|
|
193
|
-
|
|
194
|
-
def _check_output_archive(self, save_path: str) -> None:
|
|
195
|
-
try:
|
|
196
|
-
archive = zipfile.ZipFile(save_path, 'r')
|
|
197
|
-
except zipfile.BadZipFile as e:
|
|
198
|
-
raise e
|
|
199
|
-
assert len(
|
|
200
|
-
archive.namelist()
|
|
201
|
-
) == self.num_inputs_annotations, "Archive has %d inputs+annotations | expecting %d inputs+annotations" % (
|
|
202
|
-
len(archive.namelist()), self.num_inputs_annotations)
|
|
203
|
-
|
|
204
|
-
def download_archive(self, save_path: str, split: Optional[str] = None) -> None:
|
|
205
|
-
"""Downloads the archive from the URL into an archive of inputs, annotations in the directory format
|
|
206
|
-
{split}/inputs and {split}/annotations.
|
|
207
|
-
"""
|
|
208
|
-
with zipfile.ZipFile(save_path, "a") as new_archive:
|
|
209
|
-
with ThreadPoolExecutor(max_workers=self.num_workers) as executor:
|
|
210
|
-
with tqdm(total=len(self.input_iterator), desc='Downloading Dataset') as progress:
|
|
211
|
-
# Submit all jobs to the executor and store the returned futures
|
|
212
|
-
futures = [
|
|
213
|
-
executor.submit(self._write_archive, input_, new_archive, split)
|
|
214
|
-
for input_ in self.input_iterator
|
|
215
|
-
]
|
|
216
|
-
|
|
217
|
-
for _ in as_completed(futures):
|
|
218
|
-
progress.update()
|
|
219
|
-
|
|
220
|
-
self._check_output_archive(save_path)
|
|
221
|
-
logger.info("Downloaded %d inputs+annotations to %s" % (self.num_inputs_annotations,
|
|
222
|
-
save_path))
|
|
File without changes
|
|
@@ -1,66 +0,0 @@
|
|
|
1
|
-
from collections import defaultdict
|
|
2
|
-
from typing import Iterator, List, Tuple, TypeVar, Union
|
|
3
|
-
|
|
4
|
-
from clarifai_grpc.grpc.api import resources_pb2
|
|
5
|
-
|
|
6
|
-
from clarifai.client.input import Inputs
|
|
7
|
-
from clarifai.datasets.upload.features import (TextFeatures, VisualClassificationFeatures,
|
|
8
|
-
VisualDetectionFeatures, VisualSegmentationFeatures)
|
|
9
|
-
|
|
10
|
-
OutputFeaturesType = TypeVar(
|
|
11
|
-
'OutputFeaturesType',
|
|
12
|
-
bound=Union[TextFeatures, VisualClassificationFeatures, VisualDetectionFeatures,
|
|
13
|
-
VisualSegmentationFeatures])
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
class ClarifaiDataset:
|
|
17
|
-
"""Clarifai datasets base class."""
|
|
18
|
-
|
|
19
|
-
def __init__(self, datagen_object: Iterator, dataset_id: str, split: str) -> None:
|
|
20
|
-
self.datagen_object = datagen_object
|
|
21
|
-
self.dataset_id = dataset_id
|
|
22
|
-
self.split = split
|
|
23
|
-
self.all_input_ids = {}
|
|
24
|
-
self._all_input_protos = {}
|
|
25
|
-
self._all_annotation_protos = defaultdict(list)
|
|
26
|
-
self.input_object = Inputs()
|
|
27
|
-
|
|
28
|
-
def __len__(self) -> int:
|
|
29
|
-
"""Get size of all input protos"""
|
|
30
|
-
return len(self.datagen_object)
|
|
31
|
-
|
|
32
|
-
def _to_list(self, input_protos: Iterator) -> List:
|
|
33
|
-
"""Parse protos iterator to list."""
|
|
34
|
-
return list(input_protos)
|
|
35
|
-
|
|
36
|
-
def _extract_protos(self) -> None:
|
|
37
|
-
"""Create input image protos for each data generator item."""
|
|
38
|
-
raise NotImplementedError()
|
|
39
|
-
|
|
40
|
-
def get_protos(self, input_ids: List[int]
|
|
41
|
-
) -> Tuple[List[resources_pb2.Input], List[resources_pb2.Annotation]]:
|
|
42
|
-
"""Get input and annotation protos based on input_ids.
|
|
43
|
-
Args:
|
|
44
|
-
input_ids: List of input IDs to retrieve the protos for.
|
|
45
|
-
Returns:
|
|
46
|
-
Input and Annotation proto iterators for the specified input IDs.
|
|
47
|
-
"""
|
|
48
|
-
input_protos, annotation_protos = self._extract_protos(input_ids)
|
|
49
|
-
|
|
50
|
-
return input_protos, annotation_protos
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
class ClarifaiDataLoader:
|
|
54
|
-
"""Clarifai data loader base class."""
|
|
55
|
-
|
|
56
|
-
def __init__(self, split: str) -> None:
|
|
57
|
-
pass
|
|
58
|
-
|
|
59
|
-
def load_data(self) -> None:
|
|
60
|
-
raise NotImplementedError()
|
|
61
|
-
|
|
62
|
-
def __len__(self) -> int:
|
|
63
|
-
raise NotImplementedError()
|
|
64
|
-
|
|
65
|
-
def __getitem__(self, index: int) -> OutputFeaturesType:
|
|
66
|
-
raise NotImplementedError()
|
|
@@ -1,31 +0,0 @@
|
|
|
1
|
-
## Dataset upload from local directory
|
|
2
|
-
|
|
3
|
-
Examples of how to upload your local directory datasets into clarifai app using `module_dir` feature from `Dataset`.
|
|
4
|
-
|
|
5
|
-
**Note:**
|
|
6
|
-
**Note:**
|
|
7
|
-
|
|
8
|
-
- Ensure that the `CLARIFAI_PAT` environment variable is set.
|
|
9
|
-
- Ensure that the appropriate base workflow is being set for indexing respective input type.
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
## Image Classification - Cifar10
|
|
13
|
-
```python
|
|
14
|
-
from clarifai.client.dataset import Dataset
|
|
15
|
-
dataset = Dataset(user_id="user_id", app_id="app_id", dataset_id="dataset_id")
|
|
16
|
-
dataset.upload_dataset(task="visual_classification", split="train", module_dir="path_to_cifar10_module")
|
|
17
|
-
```
|
|
18
|
-
|
|
19
|
-
## Image Classification - [Food-101](https://data.vision.ee.ethz.ch/cvl/datasets_extra/food-101/)
|
|
20
|
-
```python
|
|
21
|
-
from clarifai.client.dataset import Dataset
|
|
22
|
-
dataset = Dataset(user_id="user_id", app_id="app_id", dataset_id="dataset_id")
|
|
23
|
-
dataset.upload_dataset(task="visual_classification", split="train", module_dir="path_to_food-101_module")
|
|
24
|
-
```
|
|
25
|
-
|
|
26
|
-
## Text Classification - IMDB Reviews
|
|
27
|
-
```python
|
|
28
|
-
from clarifai.client.dataset import Dataset
|
|
29
|
-
dataset = Dataset(user_id="user_id", app_id="app_id", dataset_id="dataset_id")
|
|
30
|
-
dataset.upload_dataset(task="text_clf", split="train", module_dir="path_to_imdb_reviews_module")
|
|
31
|
-
```
|
|
File without changes
|
|
File without changes
|
|
@@ -1,10 +0,0 @@
|
|
|
1
|
-
image_path,label
|
|
2
|
-
images/test_batch_700.jpg,horse
|
|
3
|
-
images/test_batch_701.jpg,bird
|
|
4
|
-
images/test_batch_702.jpg,deer
|
|
5
|
-
images/test_batch_703.jpg,ship
|
|
6
|
-
images/test_batch_704.jpg,horse
|
|
7
|
-
images/test_batch_705.jpg,deer
|
|
8
|
-
images/test_batch_706.jpg,bird
|
|
9
|
-
images/test_batch_707.jpg,truck
|
|
10
|
-
images/test_batch_709.jpg,ship
|
|
@@ -1,10 +0,0 @@
|
|
|
1
|
-
image_path,label
|
|
2
|
-
images/test_batch_700.jpg,horse
|
|
3
|
-
images/test_batch_701.jpg,bird
|
|
4
|
-
images/test_batch_702.jpg,deer
|
|
5
|
-
images/test_batch_703.jpg,ship
|
|
6
|
-
images/test_batch_704.jpg,horse
|
|
7
|
-
images/test_batch_705.jpg,deer
|
|
8
|
-
images/test_batch_706.jpg,bird
|
|
9
|
-
images/test_batch_707.jpg,truck
|
|
10
|
-
images/test_batch_709.jpg,ship
|
|
@@ -1,46 +0,0 @@
|
|
|
1
|
-
#! Cifar10 Dataset
|
|
2
|
-
|
|
3
|
-
import csv
|
|
4
|
-
import os
|
|
5
|
-
|
|
6
|
-
from clarifai.datasets.upload.base import ClarifaiDataLoader
|
|
7
|
-
from clarifai.datasets.upload.features import VisualClassificationFeatures
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
class Cifar10DataLoader(ClarifaiDataLoader):
|
|
11
|
-
"""Cifar10 Dataset."""
|
|
12
|
-
|
|
13
|
-
def __init__(self, split: str = "train"):
|
|
14
|
-
"""Initialize dataset params.
|
|
15
|
-
Args:
|
|
16
|
-
split: "train" or "test"
|
|
17
|
-
"""
|
|
18
|
-
self.split = split
|
|
19
|
-
self.data_dirs = {
|
|
20
|
-
"train": os.path.join(os.path.dirname(__file__), "cifar_small_train.csv"),
|
|
21
|
-
"test": os.path.join(os.path.dirname(__file__), "cifar_small_test.csv")
|
|
22
|
-
}
|
|
23
|
-
self.data = self.load_data()
|
|
24
|
-
|
|
25
|
-
def load_data(self):
|
|
26
|
-
data = []
|
|
27
|
-
with open(self.data_dirs[self.split]) as _file:
|
|
28
|
-
reader = csv.reader(_file)
|
|
29
|
-
next(reader, None) # skip header
|
|
30
|
-
for review in reader:
|
|
31
|
-
data.append((review[0], review[1]))
|
|
32
|
-
return data
|
|
33
|
-
|
|
34
|
-
def __getitem__(self, index):
|
|
35
|
-
item = self.data[index]
|
|
36
|
-
return VisualClassificationFeatures(
|
|
37
|
-
image_path=os.path.join(os.path.dirname(__file__), item[0]),
|
|
38
|
-
label=item[1],
|
|
39
|
-
id=os.path.basename(item[0]).split(".")[0],
|
|
40
|
-
metadata={
|
|
41
|
-
"split": self.split,
|
|
42
|
-
"image_path": item[0]
|
|
43
|
-
})
|
|
44
|
-
|
|
45
|
-
def __len__(self):
|
|
46
|
-
return len(self.data)
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_700.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_701.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_702.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_703.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_704.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_705.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_706.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_707.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_708.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/cifar10/images/test_batch_709.jpg
DELETED
|
Binary file
|
|
File without changes
|
|
@@ -1,39 +0,0 @@
|
|
|
1
|
-
import os
|
|
2
|
-
|
|
3
|
-
from clarifai.datasets.upload.base import ClarifaiDataLoader
|
|
4
|
-
from clarifai.datasets.upload.features import VisualClassificationFeatures
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
class Food101DataLoader(ClarifaiDataLoader):
|
|
8
|
-
"""Food-101 Image Classification Dataset."""
|
|
9
|
-
|
|
10
|
-
def __init__(self, split: str = "train"):
|
|
11
|
-
"""Initialize dataset params.
|
|
12
|
-
Args:
|
|
13
|
-
split: "train" or "test"
|
|
14
|
-
"""
|
|
15
|
-
self.split = split
|
|
16
|
-
self.image_dir = {"train": os.path.join(os.path.dirname(__file__), "images")}
|
|
17
|
-
self.load_data()
|
|
18
|
-
|
|
19
|
-
def load_data(self):
|
|
20
|
-
"""Load data for the food-101 dataset."""
|
|
21
|
-
self.data = []
|
|
22
|
-
class_names = os.listdir(self.image_dir[self.split])
|
|
23
|
-
for class_name in class_names:
|
|
24
|
-
for image in os.listdir(os.path.join(self.image_dir[self.split], class_name)):
|
|
25
|
-
image_path = os.path.join(self.image_dir[self.split], class_name, image)
|
|
26
|
-
self.data.append({
|
|
27
|
-
"image_path": image_path,
|
|
28
|
-
"class_name": class_name,
|
|
29
|
-
})
|
|
30
|
-
|
|
31
|
-
def __getitem__(self, idx):
|
|
32
|
-
data_item = self.data[idx]
|
|
33
|
-
image_path = data_item["image_path"]
|
|
34
|
-
class_name = data_item["class_name"]
|
|
35
|
-
return VisualClassificationFeatures(
|
|
36
|
-
image_path=image_path, label=class_name, id=os.path.basename(image_path).split(".")[0])
|
|
37
|
-
|
|
38
|
-
def __len__(self):
|
|
39
|
-
return len(self.data)
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/1420783.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/3287885.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/3617075.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/38052.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/beignets/39147.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/139558.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/1636096.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/2480925.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/3385808.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/hamburger/3647386.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/1826869.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/2243245.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/259212.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/2842688.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/prime_rib/3035414.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/1545393.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/2427642.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/3520891.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/377566.jpg
DELETED
|
Binary file
|
clarifai_utils/datasets/upload/examples/image_classification/food-101/images/ramen/503504.jpg
DELETED
|
Binary file
|
|
File without changes
|
|
File without changes
|
|
@@ -1,42 +0,0 @@
|
|
|
1
|
-
import csv
|
|
2
|
-
import os
|
|
3
|
-
|
|
4
|
-
from clarifai.datasets.upload.base import ClarifaiDataLoader
|
|
5
|
-
from clarifai.datasets.upload.features import TextFeatures
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
class IMDBMovieReviewsDataLoader(ClarifaiDataLoader):
|
|
9
|
-
"""IMDB 50K Movie Reviews Dataset."""
|
|
10
|
-
|
|
11
|
-
def __init__(self, split: str = "train"):
|
|
12
|
-
"""Initialize dataset params.
|
|
13
|
-
Args:
|
|
14
|
-
split: "train" or "test"
|
|
15
|
-
"""
|
|
16
|
-
self.split = split
|
|
17
|
-
self.data_dirs = {
|
|
18
|
-
"train": os.path.join(os.path.dirname(__file__), "train.csv"),
|
|
19
|
-
"test": os.path.join(os.path.dirname(__file__), "test.csv")
|
|
20
|
-
}
|
|
21
|
-
self.data = []
|
|
22
|
-
|
|
23
|
-
self.load_data()
|
|
24
|
-
|
|
25
|
-
def load_data(self):
|
|
26
|
-
with open(self.data_dirs[self.split]) as _file:
|
|
27
|
-
reader = csv.reader(_file)
|
|
28
|
-
next(reader, None) # skip header
|
|
29
|
-
for review in reader:
|
|
30
|
-
self.data.append({
|
|
31
|
-
"text": review[0],
|
|
32
|
-
"labels": review[1],
|
|
33
|
-
"id": None,
|
|
34
|
-
"metadata": dict(split=self.split)
|
|
35
|
-
})
|
|
36
|
-
|
|
37
|
-
def __getitem__(self, idx):
|
|
38
|
-
item = self.data[idx]
|
|
39
|
-
return TextFeatures(text=item["text"], labels=item["labels"], id=item["id"])
|
|
40
|
-
|
|
41
|
-
def __len__(self):
|
|
42
|
-
return len(self.data)
|