clarifai 9.7.0__tar.gz → 9.7.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {clarifai-9.7.0 → clarifai-9.7.2}/LICENSE +1 -1
- {clarifai-9.7.0 → clarifai-9.7.2}/MANIFEST.in +1 -1
- clarifai-9.7.2/PKG-INFO +176 -0
- clarifai-9.7.2/README.md +160 -0
- clarifai-9.7.2/clarifai/auth/__init__.py +6 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/auth/helper.py +35 -36
- clarifai-9.7.0/clarifai_utils/client/abc.py → clarifai-9.7.2/clarifai/auth/register.py +2 -2
- {clarifai-9.7.0/clarifai_utils/client → clarifai-9.7.2/clarifai/auth}/stub.py +10 -10
- clarifai-9.7.2/clarifai/client/__init__.py +1 -0
- clarifai-9.7.2/clarifai/client/app.py +483 -0
- clarifai-9.7.2/clarifai/client/auth/__init__.py +4 -0
- {clarifai-9.7.0/clarifai_utils → clarifai-9.7.2/clarifai/client}/auth/helper.py +35 -36
- clarifai-9.7.0/clarifai/client/abc.py → clarifai-9.7.2/clarifai/client/auth/register.py +2 -2
- {clarifai-9.7.0/clarifai/client → clarifai-9.7.2/clarifai/client/auth}/stub.py +10 -10
- clarifai-9.7.2/clarifai/client/base.py +112 -0
- clarifai-9.7.2/clarifai/client/dataset.py +290 -0
- clarifai-9.7.2/clarifai/client/input.py +730 -0
- clarifai-9.7.2/clarifai/client/lister.py +41 -0
- clarifai-9.7.2/clarifai/client/model.py +218 -0
- clarifai-9.7.2/clarifai/client/module.py +82 -0
- clarifai-9.7.2/clarifai/client/user.py +125 -0
- clarifai-9.7.2/clarifai/client/workflow.py +194 -0
- clarifai-9.7.2/clarifai/datasets/upload/base.py +66 -0
- clarifai-9.7.2/clarifai/datasets/upload/examples/README.md +31 -0
- clarifai-9.7.2/clarifai/datasets/upload/examples/image_classification/cifar10/dataset.py +42 -0
- clarifai-9.7.2/clarifai/datasets/upload/examples/image_classification/food-101/dataset.py +39 -0
- clarifai-9.7.2/clarifai/datasets/upload/examples/text_classification/imdb_dataset/dataset.py +37 -0
- {clarifai-9.7.0/clarifai/data_upload/datasets → clarifai-9.7.2/clarifai/datasets/upload}/features.py +4 -12
- clarifai-9.7.2/clarifai/datasets/upload/image.py +156 -0
- clarifai-9.7.2/clarifai/datasets/upload/loaders/README.md +49 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/datasets/zoo → clarifai-9.7.2/clarifai/datasets/upload/loaders}/coco_captions.py +24 -21
- {clarifai-9.7.0/clarifai_utils/data_upload/datasets/zoo → clarifai-9.7.2/clarifai/datasets/upload/loaders}/coco_detection.py +46 -42
- clarifai-9.7.2/clarifai/datasets/upload/loaders/coco_segmentation.py +166 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/datasets/zoo → clarifai-9.7.2/clarifai/datasets/upload/loaders}/imagenet_classification.py +22 -12
- {clarifai-9.7.0/clarifai/data_upload/datasets/zoo → clarifai-9.7.2/clarifai/datasets/upload/loaders}/xview_detection.py +44 -53
- clarifai-9.7.2/clarifai/datasets/upload/text.py +50 -0
- clarifai-9.7.2/clarifai/datasets/upload/utils.py +62 -0
- clarifai-9.7.2/clarifai/errors.py +90 -0
- {clarifai-9.7.0/clarifai_utils → clarifai-9.7.2/clarifai}/urls/helper.py +16 -17
- clarifai-9.7.2/clarifai/utils/logging.py +40 -0
- clarifai-9.7.2/clarifai/utils/misc.py +33 -0
- clarifai-9.7.2/clarifai/versions.py +6 -0
- clarifai-9.7.2/clarifai.egg-info/PKG-INFO +176 -0
- clarifai-9.7.2/clarifai.egg-info/SOURCES.txt +358 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai.egg-info/requires.txt +1 -1
- clarifai-9.7.2/clarifai_utils/auth/__init__.py +6 -0
- clarifai-9.7.2/clarifai_utils/auth/helper.py +377 -0
- clarifai-9.7.2/clarifai_utils/auth/register.py +23 -0
- clarifai-9.7.2/clarifai_utils/auth/stub.py +127 -0
- clarifai-9.7.2/clarifai_utils/client/__init__.py +1 -0
- clarifai-9.7.2/clarifai_utils/client/app.py +483 -0
- clarifai-9.7.2/clarifai_utils/client/auth/__init__.py +4 -0
- clarifai-9.7.2/clarifai_utils/client/auth/helper.py +377 -0
- clarifai-9.7.2/clarifai_utils/client/auth/register.py +23 -0
- clarifai-9.7.2/clarifai_utils/client/auth/stub.py +127 -0
- clarifai-9.7.2/clarifai_utils/client/base.py +112 -0
- clarifai-9.7.2/clarifai_utils/client/dataset.py +290 -0
- clarifai-9.7.2/clarifai_utils/client/input.py +730 -0
- clarifai-9.7.2/clarifai_utils/client/lister.py +41 -0
- clarifai-9.7.2/clarifai_utils/client/model.py +218 -0
- clarifai-9.7.2/clarifai_utils/client/module.py +82 -0
- clarifai-9.7.2/clarifai_utils/client/user.py +125 -0
- clarifai-9.7.2/clarifai_utils/client/workflow.py +194 -0
- clarifai-9.7.2/clarifai_utils/datasets/upload/base.py +66 -0
- clarifai-9.7.2/clarifai_utils/datasets/upload/examples/README.md +31 -0
- clarifai-9.7.2/clarifai_utils/datasets/upload/examples/image_classification/cifar10/dataset.py +42 -0
- clarifai-9.7.2/clarifai_utils/datasets/upload/examples/image_classification/food-101/dataset.py +39 -0
- clarifai-9.7.2/clarifai_utils/datasets/upload/examples/text_classification/imdb_dataset/dataset.py +37 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/datasets → clarifai-9.7.2/clarifai_utils/datasets/upload}/features.py +4 -12
- clarifai-9.7.2/clarifai_utils/datasets/upload/image.py +156 -0
- clarifai-9.7.2/clarifai_utils/datasets/upload/loaders/README.md +49 -0
- {clarifai-9.7.0/clarifai/data_upload/datasets/zoo → clarifai-9.7.2/clarifai_utils/datasets/upload/loaders}/coco_captions.py +24 -21
- {clarifai-9.7.0/clarifai/data_upload/datasets/zoo → clarifai-9.7.2/clarifai_utils/datasets/upload/loaders}/coco_detection.py +46 -42
- clarifai-9.7.2/clarifai_utils/datasets/upload/loaders/coco_segmentation.py +166 -0
- {clarifai-9.7.0/clarifai/data_upload/datasets/zoo → clarifai-9.7.2/clarifai_utils/datasets/upload/loaders}/imagenet_classification.py +22 -12
- {clarifai-9.7.0/clarifai_utils/data_upload/datasets/zoo → clarifai-9.7.2/clarifai_utils/datasets/upload/loaders}/xview_detection.py +44 -53
- clarifai-9.7.2/clarifai_utils/datasets/upload/text.py +50 -0
- clarifai-9.7.2/clarifai_utils/datasets/upload/utils.py +62 -0
- clarifai-9.7.2/clarifai_utils/errors.py +90 -0
- {clarifai-9.7.0/clarifai → clarifai-9.7.2/clarifai_utils}/urls/helper.py +16 -17
- clarifai-9.7.2/clarifai_utils/utils/logging.py +40 -0
- clarifai-9.7.2/clarifai_utils/utils/misc.py +33 -0
- clarifai-9.7.2/clarifai_utils/versions.py +6 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/setup.py +4 -4
- clarifai-9.7.2/tests/test_app.py +118 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/tests/test_auth.py +1 -1
- clarifai-9.7.2/tests/test_model_predict.py +124 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/tests/test_stub.py +2 -2
- clarifai-9.7.2/tests/test_workflow_predict.py +49 -0
- clarifai-9.7.0/PKG-INFO +0 -96
- clarifai-9.7.0/README.md +0 -80
- clarifai-9.7.0/clarifai/client/__init__.py +0 -4
- clarifai-9.7.0/clarifai/data_upload/README.md +0 -63
- clarifai-9.7.0/clarifai/data_upload/convert_csv.py +0 -182
- clarifai-9.7.0/clarifai/data_upload/datasets/base.py +0 -87
- clarifai-9.7.0/clarifai/data_upload/datasets/image.py +0 -253
- clarifai-9.7.0/clarifai/data_upload/datasets/text.py +0 -60
- clarifai-9.7.0/clarifai/data_upload/datasets/zoo/README.md +0 -55
- clarifai-9.7.0/clarifai/data_upload/datasets/zoo/coco_segmentation.py +0 -160
- clarifai-9.7.0/clarifai/data_upload/examples/README.md +0 -5
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/cifar10/dataset.py +0 -40
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/dataset.py +0 -39
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/beignets/1036242.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/beignets/1114182.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/beignets/2012944.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/beignets/2464389.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/beignets/478632.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/hamburger/1061270.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/hamburger/1202261.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/hamburger/1381751.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/hamburger/3289634.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/hamburger/862025.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/102197.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/2749372.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/2938268.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/3590861.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/prime_rib/746716.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/ramen/2955110.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/ramen/3208966.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/ramen/3270629.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/ramen/3424562.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101/images/ramen/544680.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2007_000464.xml +0 -39
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2008_000853.xml +0 -28
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2008_003182.xml +0 -54
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2008_008526.xml +0 -67
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2009_004315.xml +0 -28
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2009_004382.xml +0 -28
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2011_000430.xml +0 -28
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2011_001610.xml +0 -46
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2011_006412.xml +0 -99
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/annotations/2012_000690.xml +0 -43
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/dataset.py +0 -76
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2007_000464.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2008_000853.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2008_003182.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2008_008526.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2009_004315.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2009_004382.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2011_000430.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2011_001610.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2011_006412.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc/images/2012_000690.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/annotations/instances_val2017_subset.json +0 -5342
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/dataset.py +0 -107
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000074646.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000086956.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000166563.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000176857.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000182202.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000193245.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000384850.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000409630.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000424349.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco/images/000000573008.jpg +0 -0
- clarifai-9.7.0/clarifai/data_upload/examples/text_classification/imdb_dataset/dataset.py +0 -40
- clarifai-9.7.0/clarifai/data_upload/examples.py +0 -17
- clarifai-9.7.0/clarifai/data_upload/upload.py +0 -356
- clarifai-9.7.0/clarifai/dataset_export/dataset_export_inputs.py +0 -205
- clarifai-9.7.0/clarifai/listing/concepts.py +0 -37
- clarifai-9.7.0/clarifai/listing/datasets.py +0 -37
- clarifai-9.7.0/clarifai/listing/inputs.py +0 -111
- clarifai-9.7.0/clarifai/listing/installed_module_versions.py +0 -40
- clarifai-9.7.0/clarifai/listing/lister.py +0 -200
- clarifai-9.7.0/clarifai/listing/models.py +0 -46
- clarifai-9.7.0/clarifai/listing/module_versions.py +0 -42
- clarifai-9.7.0/clarifai/listing/modules.py +0 -36
- clarifai-9.7.0/clarifai/runners/base.py +0 -140
- clarifai-9.7.0/clarifai/runners/example.py +0 -36
- clarifai-9.7.0/clarifai.egg-info/PKG-INFO +0 -96
- clarifai-9.7.0/clarifai.egg-info/SOURCES.txt +0 -463
- clarifai-9.7.0/clarifai_utils/client/__init__.py +0 -4
- clarifai-9.7.0/clarifai_utils/data_upload/README.md +0 -63
- clarifai-9.7.0/clarifai_utils/data_upload/convert_csv.py +0 -182
- clarifai-9.7.0/clarifai_utils/data_upload/datasets/base.py +0 -87
- clarifai-9.7.0/clarifai_utils/data_upload/datasets/image.py +0 -253
- clarifai-9.7.0/clarifai_utils/data_upload/datasets/text.py +0 -60
- clarifai-9.7.0/clarifai_utils/data_upload/datasets/zoo/README.md +0 -55
- clarifai-9.7.0/clarifai_utils/data_upload/datasets/zoo/coco_segmentation.py +0 -160
- clarifai-9.7.0/clarifai_utils/data_upload/examples/README.md +0 -5
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/cifar10/dataset.py +0 -40
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/dataset.py +0 -39
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/1036242.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/1114182.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/2012944.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/2464389.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/beignets/478632.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1061270.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1202261.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/1381751.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/3289634.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/hamburger/862025.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/102197.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/2749372.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/2938268.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/3590861.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/prime_rib/746716.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/2955110.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/3208966.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/3270629.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/3424562.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101/images/ramen/544680.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2007_000464.xml +0 -39
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2008_000853.xml +0 -28
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2008_003182.xml +0 -54
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2008_008526.xml +0 -67
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2009_004315.xml +0 -28
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2009_004382.xml +0 -28
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2011_000430.xml +0 -28
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2011_001610.xml +0 -46
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2011_006412.xml +0 -99
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/annotations/2012_000690.xml +0 -43
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/dataset.py +0 -76
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2007_000464.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2008_000853.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2008_003182.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2008_008526.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2009_004315.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2009_004382.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2011_000430.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2011_001610.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2011_006412.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc/images/2012_000690.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/annotations/instances_val2017_subset.json +0 -5342
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/dataset.py +0 -107
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000074646.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000086956.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000166563.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000176857.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000182202.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000193245.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000384850.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000409630.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000424349.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco/images/000000573008.jpg +0 -0
- clarifai-9.7.0/clarifai_utils/data_upload/examples/text_classification/imdb_dataset/dataset.py +0 -40
- clarifai-9.7.0/clarifai_utils/data_upload/examples.py +0 -17
- clarifai-9.7.0/clarifai_utils/data_upload/upload.py +0 -356
- clarifai-9.7.0/clarifai_utils/dataset_export/dataset_export_inputs.py +0 -205
- clarifai-9.7.0/clarifai_utils/listing/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/listing/concepts.py +0 -37
- clarifai-9.7.0/clarifai_utils/listing/datasets.py +0 -37
- clarifai-9.7.0/clarifai_utils/listing/inputs.py +0 -111
- clarifai-9.7.0/clarifai_utils/listing/installed_module_versions.py +0 -40
- clarifai-9.7.0/clarifai_utils/listing/lister.py +0 -200
- clarifai-9.7.0/clarifai_utils/listing/models.py +0 -46
- clarifai-9.7.0/clarifai_utils/listing/module_versions.py +0 -42
- clarifai-9.7.0/clarifai_utils/listing/modules.py +0 -36
- clarifai-9.7.0/clarifai_utils/models/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/1/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/1/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/1/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/1/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/modules/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/runners/__init__.py +0 -0
- clarifai-9.7.0/clarifai_utils/runners/base.py +0 -140
- clarifai-9.7.0/clarifai_utils/runners/example.py +0 -36
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/__init__.py +0 -0
- /clarifai-9.7.0/clarifai/auth/__init__.py → /clarifai-9.7.2/clarifai/cli.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload/datasets → clarifai-9.7.2/clarifai/datasets/upload}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload/datasets/zoo → clarifai-9.7.2/clarifai/datasets/upload/examples/image_classification}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/image_classification → clarifai-9.7.2/clarifai/datasets/upload/examples/image_classification/cifar10}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/cifar_small_test.csv +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/cifar_small_train.csv +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_700.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_701.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_702.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_703.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_704.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_705.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_706.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_707.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_708.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/cifar10/images/test_batch_709.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/image_classification/cifar10 → clarifai-9.7.2/clarifai/datasets/upload/examples/image_classification/food-101}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/beignets/1420783.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/beignets/3287885.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/beignets/3617075.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/beignets/38052.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/beignets/39147.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/hamburger/139558.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/hamburger/1636096.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/hamburger/2480925.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/hamburger/3385808.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/hamburger/3647386.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/prime_rib/1826869.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/prime_rib/2243245.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/prime_rib/259212.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/prime_rib/2842688.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/prime_rib/3035414.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/ramen/1545393.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/ramen/2427642.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/ramen/3520891.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/ramen/377566.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/image_classification/food-101/images/ramen/503504.jpg +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/image_classification/food-101 → clarifai-9.7.2/clarifai/datasets/upload/examples/text_classification}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/image_detection → clarifai-9.7.2/clarifai/datasets/upload/examples/text_classification/imdb_dataset}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/text_classification/imdb_dataset/test.csv +0 -0
- {clarifai-9.7.0/clarifai/data_upload → clarifai-9.7.2/clarifai/datasets/upload}/examples/text_classification/imdb_dataset/train.csv +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/image_detection/voc → clarifai-9.7.2/clarifai/datasets/upload/loaders}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation → clarifai-9.7.2/clarifai/models}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/api.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/cli/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/cli/deploy_cli.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/cli/model_zip.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/cli/repository.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/constants.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/docs/dependencies.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/docs/model_types.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/docs/output.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/envs/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/README.md +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/image_segmentation/coco → clarifai-9.7.2/clarifai/models/model_serving/examples/image_classification/age_vit/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/config.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/preprocessor_config.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/image_classification/age_vit/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/README.md +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/text_classification → clarifai-9.7.2/clarifai/models/model_serving/examples/text_classification/xlm-roberta/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/config.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/special_tokens_map.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_classification/xlm-roberta/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_embedding/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_image/README.md +0 -0
- {clarifai-9.7.0/clarifai/data_upload/examples/text_classification/imdb_dataset → clarifai-9.7.2/clarifai/models/model_serving/examples/text_to_image/sd-v1.5/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_image/sd-v1.5/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_image/sd-v1.5/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_image/sd-v1.5/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_image/sd-v1.5/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_image/sd-v1.5/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_text/README.md +0 -0
- {clarifai-9.7.0/clarifai/listing → clarifai-9.7.2/clarifai/models/model_serving/examples/text_to_text/bart-summarize/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_text/bart-summarize/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_text/bart-summarize/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_text/bart-summarize/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_text/bart-summarize/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/text_to_text/bart-summarize/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_detection/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_detection/yolov5x/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_detection/yolov5x/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_detection/yolov5x/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_detection/yolov5x/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_detection/yolov5x/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_detection/yolov5x/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_embedding/README.md +0 -0
- {clarifai-9.7.0/clarifai/models → clarifai-9.7.2/clarifai/models/model_serving/examples/visual_embedding/vit-base/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_embedding/vit-base/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_embedding/vit-base/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_embedding/vit-base/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_embedding/vit-base/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_embedding/vit-base/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_segmentation/README.md +0 -0
- {clarifai-9.7.0/clarifai/models/model_serving/examples/image_classification/age_vit → clarifai-9.7.2/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2}/1/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/model_config/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/model_config/deploy.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/model_config/serializer.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/model_config/triton_config.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/models/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/models/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/models/model_types.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/models/output.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/models/pb_model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/models/model_serving/pb_model_repository.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/modules/README.md +0 -0
- {clarifai-9.7.0/clarifai/models/model_serving/examples/text_classification/xlm-roberta/1 → clarifai-9.7.2/clarifai/modules}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/modules/css.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/modules/pages.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai/modules/style.css +0 -0
- {clarifai-9.7.0/clarifai/models/model_serving/examples/text_to_image/sd-v1.5/1 → clarifai-9.7.2/clarifai/utils}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai.egg-info/dependency_links.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai.egg-info/entry_points.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai.egg-info/top_level.txt +0 -0
- {clarifai-9.7.0/clarifai/models/model_serving/examples/text_to_text/bart-summarize/1 → clarifai-9.7.2/clarifai_utils}/__init__.py +0 -0
- /clarifai-9.7.0/clarifai/models/model_serving/examples/visual_embedding/vit-base/1/__init__.py → /clarifai-9.7.2/clarifai_utils/cli.py +0 -0
- {clarifai-9.7.0/clarifai/models/model_serving/examples/visual_segmentation/segformer-b2/1 → clarifai-9.7.2/clarifai_utils/datasets}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/modules → clarifai-9.7.2/clarifai_utils/datasets/upload}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai/runners → clarifai-9.7.2/clarifai_utils/datasets/upload/examples/image_classification}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai_utils → clarifai-9.7.2/clarifai_utils/datasets/upload/examples/image_classification/cifar10}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/cifar_small_test.csv +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/cifar_small_train.csv +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_700.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_701.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_702.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_703.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_704.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_705.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_706.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_707.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_708.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/cifar10/images/test_batch_709.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/auth → clarifai-9.7.2/clarifai_utils/datasets/upload/examples/image_classification/food-101}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/beignets/1420783.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/beignets/3287885.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/beignets/3617075.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/beignets/38052.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/beignets/39147.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/hamburger/139558.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/hamburger/1636096.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/hamburger/2480925.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/hamburger/3385808.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/hamburger/3647386.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/prime_rib/1826869.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/prime_rib/2243245.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/prime_rib/259212.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/prime_rib/2842688.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/prime_rib/3035414.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/ramen/1545393.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/ramen/2427642.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/ramen/3520891.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/ramen/377566.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/image_classification/food-101/images/ramen/503504.jpg +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload/examples/text_classification}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/datasets → clarifai-9.7.2/clarifai_utils/datasets/upload/examples/text_classification/imdb_dataset}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/text_classification/imdb_dataset/test.csv +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload → clarifai-9.7.2/clarifai_utils/datasets/upload}/examples/text_classification/imdb_dataset/train.csv +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/datasets/zoo → clarifai-9.7.2/clarifai_utils/datasets/upload/loaders}/__init__.py +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification → clarifai-9.7.2/clarifai_utils/models}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/api.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/cli/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/cli/deploy_cli.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/cli/model_zip.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/cli/repository.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/constants.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/docs/dependencies.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/docs/model_types.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/docs/output.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/envs/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/README.md +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/cifar10 → clarifai-9.7.2/clarifai_utils/models/model_serving/examples/image_classification/age_vit/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/config.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/1/vit-age-classifier/preprocessor_config.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/image_classification/age_vit/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/README.md +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/image_classification/food-101 → clarifai-9.7.2/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/config.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/1/twitter-xlm-roberta-base-sentiment/special_tokens_map.json +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_classification/xlm-roberta/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_embedding/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_image/README.md +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection → clarifai-9.7.2/clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_image/sd-v1.5/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_text/README.md +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/image_detection/voc → clarifai-9.7.2/clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/text_to_text/bart-summarize/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_detection/README.md +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_detection/yolov5x/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_embedding/README.md +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation → clarifai-9.7.2/clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_embedding/vit-base/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_segmentation/README.md +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/image_segmentation/coco → clarifai-9.7.2/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/1}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/1/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/1/model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/config.pbtxt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/labels.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/requirements.txt +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/examples/visual_segmentation/segformer-b2/triton_conda.yaml +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/model_config/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/model_config/deploy.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/model_config/serializer.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/model_config/triton_config.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/models/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/models/inference.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/models/model_types.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/models/output.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/models/pb_model.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/models/model_serving/pb_model_repository.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/modules/README.md +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/text_classification → clarifai-9.7.2/clarifai_utils/modules}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/modules/css.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/modules/pages.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/clarifai_utils/modules/style.css +0 -0
- {clarifai-9.7.0/clarifai_utils/data_upload/examples/text_classification/imdb_dataset → clarifai-9.7.2/clarifai_utils/utils}/__init__.py +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/setup.cfg +0 -0
- {clarifai-9.7.0 → clarifai-9.7.2}/tests/test_modules.py +0 -0
clarifai-9.7.2/PKG-INFO
ADDED
|
@@ -0,0 +1,176 @@
|
|
|
1
|
+
Metadata-Version: 2.1
|
|
2
|
+
Name: clarifai
|
|
3
|
+
Version: 9.7.2
|
|
4
|
+
Summary: Clarifai Python SDK
|
|
5
|
+
Home-page: https://github.com/Clarifai/clarifai-python
|
|
6
|
+
Author: Clarifai
|
|
7
|
+
Author-email: support@clarifai.com
|
|
8
|
+
License: Apache 2.0
|
|
9
|
+
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
|
|
10
|
+
Classifier: Programming Language :: Python :: 3
|
|
11
|
+
Classifier: License :: OSI Approved :: Apache Software License
|
|
12
|
+
Classifier: Operating System :: OS Independent
|
|
13
|
+
Requires-Python: >=3.8
|
|
14
|
+
Description-Content-Type: text/markdown
|
|
15
|
+
License-File: LICENSE
|
|
16
|
+
|
|
17
|
+

|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
|
|
21
|
+
# Clarifai API Python Client
|
|
22
|
+
|
|
23
|
+
This is the official Python client for interacting with our powerful recognition [API](https://docs.clarifai.com).
|
|
24
|
+
The Clarifai Python SDK offers a comprehensive set of tools to integrate Clarifai's AI-powered image, video, and text recognition capabilities into your applications. With just a few lines of code, you can leverage cutting-edge artificial intelligence to unlock valuable insights from visual and textual content.
|
|
25
|
+
|
|
26
|
+
* Try the Clarifai demo at: https://clarifai.com/demo
|
|
27
|
+
* Sign up for a free account at: https://clarifai.com/developer/account/signup/
|
|
28
|
+
* Read the developer guide at: https://clarifai.com/developer/guide/
|
|
29
|
+
|
|
30
|
+
## Getting started
|
|
31
|
+
Clarifai uses Personal Access Tokens(PATs) to validate requests. You can create and manage PATs under your Clarifai account security settings.
|
|
32
|
+
|
|
33
|
+
Export your PAT as an environment variable. Then, import and initialize the API Client.
|
|
34
|
+
|
|
35
|
+
|
|
36
|
+
```cmd
|
|
37
|
+
export CLARIFAI_PAT={your personal access token}
|
|
38
|
+
```
|
|
39
|
+
|
|
40
|
+
```python
|
|
41
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
42
|
+
from clarifai.client.user import User
|
|
43
|
+
client = User(user_id="user_id")
|
|
44
|
+
|
|
45
|
+
# Get all apps
|
|
46
|
+
apps = client.list_apps()
|
|
47
|
+
|
|
48
|
+
# Create app and dataset
|
|
49
|
+
app = client.create_app(app_id="demo_app", base_workflow="Universal")
|
|
50
|
+
dataset = app.create_dataset(dataset_id="demo_dataset")
|
|
51
|
+
|
|
52
|
+
# execute data upload to Clarifai app dataset
|
|
53
|
+
dataset.upload_dataset(task='visual_segmentation', split="train", dataset_loader='coco_segmentation')
|
|
54
|
+
|
|
55
|
+
#upload text from csv
|
|
56
|
+
dataset.upload_from_csv(csv_path='csv_path', labels=True)
|
|
57
|
+
|
|
58
|
+
#upload data from folder
|
|
59
|
+
dataset.upload_from_folder(folder_path='folder_path', input_type='text', labels=True)
|
|
60
|
+
```
|
|
61
|
+
|
|
62
|
+
|
|
63
|
+
## Interacting with Inputs
|
|
64
|
+
|
|
65
|
+
```python
|
|
66
|
+
from clarifai.client.user import User
|
|
67
|
+
app = User(user_id="user_id").app(app_id="app_id")
|
|
68
|
+
input_obj = app.inputs()
|
|
69
|
+
|
|
70
|
+
#input upload from url
|
|
71
|
+
input_obj.upload_from_url(input_id = 'demo', image_url='https://samples.clarifai.com/metro-north.jpg')
|
|
72
|
+
|
|
73
|
+
#input upload from filename
|
|
74
|
+
input_obj.upload_from_file(input_id = 'demo', video_file='demo.mp4')
|
|
75
|
+
|
|
76
|
+
#listing inputs
|
|
77
|
+
input_obj.list_inputs()
|
|
78
|
+
|
|
79
|
+
# text upload
|
|
80
|
+
input_obj.upload_text(input_id = 'demo', raw_text = 'This is a test')
|
|
81
|
+
```
|
|
82
|
+
|
|
83
|
+
|
|
84
|
+
## Interacting with Models
|
|
85
|
+
|
|
86
|
+
### Model Predict
|
|
87
|
+
```python
|
|
88
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
89
|
+
from clarifai.client.model import Model
|
|
90
|
+
|
|
91
|
+
# Model Predict
|
|
92
|
+
model = Model("https://clarifai.com/anthropic/completion/models/claude-v2").predict_by_bytes(b"Write a tweet on future of AI", "text")
|
|
93
|
+
|
|
94
|
+
model = Model(user_id="user_id", app_id="app_id", model_id="model_id")
|
|
95
|
+
model_prediction = model.predict_by_url(url="url", input_type="image") # Supports image, text, audio, video
|
|
96
|
+
|
|
97
|
+
# Customizing Model Inference Output
|
|
98
|
+
model = Model(user_id="user_id", app_id="app_id", model_id="model_id",
|
|
99
|
+
output_config={"min_value": 0.98}) # Return predictions having prediction confidence > 0.98
|
|
100
|
+
model_prediction = model.predict_by_filepath(filepath="local_filepath", input_type="text") # Supports image, text, audio, video
|
|
101
|
+
|
|
102
|
+
model = Model(user_id="user_id", app_id="app_id", model_id="model_id",
|
|
103
|
+
output_config={"sample_ms": 2000}) # Return predictions for specified interval
|
|
104
|
+
model_prediction = model.predict_by_url(url="VIDEO_URL", input_type="video")
|
|
105
|
+
```
|
|
106
|
+
### Models Listing
|
|
107
|
+
```python
|
|
108
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
109
|
+
|
|
110
|
+
# List all model versions
|
|
111
|
+
all_model_versions = model.list_versions()
|
|
112
|
+
|
|
113
|
+
# Go to specific model version
|
|
114
|
+
model_v1 = client.app("app_id").model(model_id="model_id", model_version_id="model_version_id")
|
|
115
|
+
|
|
116
|
+
# List all models in an app
|
|
117
|
+
all_models = app.list_models()
|
|
118
|
+
|
|
119
|
+
# List all models in community filtered by model_type, description
|
|
120
|
+
all_llm_community_models = App().list_models(filter_by={"query": "LLM",
|
|
121
|
+
"model_type_id": "text-to-text"}, only_in_app=False)
|
|
122
|
+
```
|
|
123
|
+
|
|
124
|
+
## Interacting with Workflows
|
|
125
|
+
|
|
126
|
+
### Workflow Predict
|
|
127
|
+
```python
|
|
128
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
129
|
+
from clarifai.client.workflow import Workflow
|
|
130
|
+
|
|
131
|
+
# Workflow Predict
|
|
132
|
+
workflow = Workflow("workflow_url") # Example: https://clarifai.com/clarifai/main/workflows/Face-Sentiment
|
|
133
|
+
workflow_prediction = workflow.predict_by_url(url="url", input_type="image") # Supports image, text, audio, video
|
|
134
|
+
|
|
135
|
+
# Customizing Workflow Inference Output
|
|
136
|
+
workflow = Workflow(user_id="user_id", app_id="app_id", workflow_id="workflow_id",
|
|
137
|
+
output_config={"min_value": 0.98}) # Return predictions having prediction confidence > 0.98
|
|
138
|
+
workflow_prediction = workflow.predict_by_filepath(filepath="local_filepath", input_type="text") # Supports image, text, audio, video
|
|
139
|
+
```
|
|
140
|
+
|
|
141
|
+
### Workflows Listing
|
|
142
|
+
```python
|
|
143
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
144
|
+
|
|
145
|
+
# List all workflow versions
|
|
146
|
+
all_workflow_versions = workflow.list_versions()
|
|
147
|
+
|
|
148
|
+
# Go to specific workflow version
|
|
149
|
+
workflow_v1 = Workflow(workflow_id="workflow_id", workflow_version=dict(id="workflow_version_id"), app_id="app_id", user_id="user_id")
|
|
150
|
+
|
|
151
|
+
# List all workflow in an app
|
|
152
|
+
all_workflow = app.list_workflow()
|
|
153
|
+
|
|
154
|
+
# List all workflow in community filtered by description
|
|
155
|
+
all_face_community_workflows = App().list_workflows(filter_by={"query": "face"}, only_in_app=False) # Get all face related workflows
|
|
156
|
+
```
|
|
157
|
+
|
|
158
|
+
## Interacting with Modules
|
|
159
|
+
|
|
160
|
+
```python
|
|
161
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
162
|
+
from clarifai.client.app import App
|
|
163
|
+
app = App(user_id="user_id", app_id="app_id")
|
|
164
|
+
|
|
165
|
+
# create a new module
|
|
166
|
+
module = app.create_module(module_id="module_id", description="module_description")
|
|
167
|
+
|
|
168
|
+
# List all modules in an app
|
|
169
|
+
all_modules = app.list_modules()
|
|
170
|
+
|
|
171
|
+
# List all module versions
|
|
172
|
+
all_module_versions = module.list_versions()
|
|
173
|
+
|
|
174
|
+
# Delete a module
|
|
175
|
+
app.delete_module(module_id="module_id")
|
|
176
|
+
```
|
clarifai-9.7.2/README.md
ADDED
|
@@ -0,0 +1,160 @@
|
|
|
1
|
+

|
|
2
|
+
|
|
3
|
+
|
|
4
|
+
|
|
5
|
+
# Clarifai API Python Client
|
|
6
|
+
|
|
7
|
+
This is the official Python client for interacting with our powerful recognition [API](https://docs.clarifai.com).
|
|
8
|
+
The Clarifai Python SDK offers a comprehensive set of tools to integrate Clarifai's AI-powered image, video, and text recognition capabilities into your applications. With just a few lines of code, you can leverage cutting-edge artificial intelligence to unlock valuable insights from visual and textual content.
|
|
9
|
+
|
|
10
|
+
* Try the Clarifai demo at: https://clarifai.com/demo
|
|
11
|
+
* Sign up for a free account at: https://clarifai.com/developer/account/signup/
|
|
12
|
+
* Read the developer guide at: https://clarifai.com/developer/guide/
|
|
13
|
+
|
|
14
|
+
## Getting started
|
|
15
|
+
Clarifai uses Personal Access Tokens(PATs) to validate requests. You can create and manage PATs under your Clarifai account security settings.
|
|
16
|
+
|
|
17
|
+
Export your PAT as an environment variable. Then, import and initialize the API Client.
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
```cmd
|
|
21
|
+
export CLARIFAI_PAT={your personal access token}
|
|
22
|
+
```
|
|
23
|
+
|
|
24
|
+
```python
|
|
25
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
26
|
+
from clarifai.client.user import User
|
|
27
|
+
client = User(user_id="user_id")
|
|
28
|
+
|
|
29
|
+
# Get all apps
|
|
30
|
+
apps = client.list_apps()
|
|
31
|
+
|
|
32
|
+
# Create app and dataset
|
|
33
|
+
app = client.create_app(app_id="demo_app", base_workflow="Universal")
|
|
34
|
+
dataset = app.create_dataset(dataset_id="demo_dataset")
|
|
35
|
+
|
|
36
|
+
# execute data upload to Clarifai app dataset
|
|
37
|
+
dataset.upload_dataset(task='visual_segmentation', split="train", dataset_loader='coco_segmentation')
|
|
38
|
+
|
|
39
|
+
#upload text from csv
|
|
40
|
+
dataset.upload_from_csv(csv_path='csv_path', labels=True)
|
|
41
|
+
|
|
42
|
+
#upload data from folder
|
|
43
|
+
dataset.upload_from_folder(folder_path='folder_path', input_type='text', labels=True)
|
|
44
|
+
```
|
|
45
|
+
|
|
46
|
+
|
|
47
|
+
## Interacting with Inputs
|
|
48
|
+
|
|
49
|
+
```python
|
|
50
|
+
from clarifai.client.user import User
|
|
51
|
+
app = User(user_id="user_id").app(app_id="app_id")
|
|
52
|
+
input_obj = app.inputs()
|
|
53
|
+
|
|
54
|
+
#input upload from url
|
|
55
|
+
input_obj.upload_from_url(input_id = 'demo', image_url='https://samples.clarifai.com/metro-north.jpg')
|
|
56
|
+
|
|
57
|
+
#input upload from filename
|
|
58
|
+
input_obj.upload_from_file(input_id = 'demo', video_file='demo.mp4')
|
|
59
|
+
|
|
60
|
+
#listing inputs
|
|
61
|
+
input_obj.list_inputs()
|
|
62
|
+
|
|
63
|
+
# text upload
|
|
64
|
+
input_obj.upload_text(input_id = 'demo', raw_text = 'This is a test')
|
|
65
|
+
```
|
|
66
|
+
|
|
67
|
+
|
|
68
|
+
## Interacting with Models
|
|
69
|
+
|
|
70
|
+
### Model Predict
|
|
71
|
+
```python
|
|
72
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
73
|
+
from clarifai.client.model import Model
|
|
74
|
+
|
|
75
|
+
# Model Predict
|
|
76
|
+
model = Model("https://clarifai.com/anthropic/completion/models/claude-v2").predict_by_bytes(b"Write a tweet on future of AI", "text")
|
|
77
|
+
|
|
78
|
+
model = Model(user_id="user_id", app_id="app_id", model_id="model_id")
|
|
79
|
+
model_prediction = model.predict_by_url(url="url", input_type="image") # Supports image, text, audio, video
|
|
80
|
+
|
|
81
|
+
# Customizing Model Inference Output
|
|
82
|
+
model = Model(user_id="user_id", app_id="app_id", model_id="model_id",
|
|
83
|
+
output_config={"min_value": 0.98}) # Return predictions having prediction confidence > 0.98
|
|
84
|
+
model_prediction = model.predict_by_filepath(filepath="local_filepath", input_type="text") # Supports image, text, audio, video
|
|
85
|
+
|
|
86
|
+
model = Model(user_id="user_id", app_id="app_id", model_id="model_id",
|
|
87
|
+
output_config={"sample_ms": 2000}) # Return predictions for specified interval
|
|
88
|
+
model_prediction = model.predict_by_url(url="VIDEO_URL", input_type="video")
|
|
89
|
+
```
|
|
90
|
+
### Models Listing
|
|
91
|
+
```python
|
|
92
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
93
|
+
|
|
94
|
+
# List all model versions
|
|
95
|
+
all_model_versions = model.list_versions()
|
|
96
|
+
|
|
97
|
+
# Go to specific model version
|
|
98
|
+
model_v1 = client.app("app_id").model(model_id="model_id", model_version_id="model_version_id")
|
|
99
|
+
|
|
100
|
+
# List all models in an app
|
|
101
|
+
all_models = app.list_models()
|
|
102
|
+
|
|
103
|
+
# List all models in community filtered by model_type, description
|
|
104
|
+
all_llm_community_models = App().list_models(filter_by={"query": "LLM",
|
|
105
|
+
"model_type_id": "text-to-text"}, only_in_app=False)
|
|
106
|
+
```
|
|
107
|
+
|
|
108
|
+
## Interacting with Workflows
|
|
109
|
+
|
|
110
|
+
### Workflow Predict
|
|
111
|
+
```python
|
|
112
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
113
|
+
from clarifai.client.workflow import Workflow
|
|
114
|
+
|
|
115
|
+
# Workflow Predict
|
|
116
|
+
workflow = Workflow("workflow_url") # Example: https://clarifai.com/clarifai/main/workflows/Face-Sentiment
|
|
117
|
+
workflow_prediction = workflow.predict_by_url(url="url", input_type="image") # Supports image, text, audio, video
|
|
118
|
+
|
|
119
|
+
# Customizing Workflow Inference Output
|
|
120
|
+
workflow = Workflow(user_id="user_id", app_id="app_id", workflow_id="workflow_id",
|
|
121
|
+
output_config={"min_value": 0.98}) # Return predictions having prediction confidence > 0.98
|
|
122
|
+
workflow_prediction = workflow.predict_by_filepath(filepath="local_filepath", input_type="text") # Supports image, text, audio, video
|
|
123
|
+
```
|
|
124
|
+
|
|
125
|
+
### Workflows Listing
|
|
126
|
+
```python
|
|
127
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
128
|
+
|
|
129
|
+
# List all workflow versions
|
|
130
|
+
all_workflow_versions = workflow.list_versions()
|
|
131
|
+
|
|
132
|
+
# Go to specific workflow version
|
|
133
|
+
workflow_v1 = Workflow(workflow_id="workflow_id", workflow_version=dict(id="workflow_version_id"), app_id="app_id", user_id="user_id")
|
|
134
|
+
|
|
135
|
+
# List all workflow in an app
|
|
136
|
+
all_workflow = app.list_workflow()
|
|
137
|
+
|
|
138
|
+
# List all workflow in community filtered by description
|
|
139
|
+
all_face_community_workflows = App().list_workflows(filter_by={"query": "face"}, only_in_app=False) # Get all face related workflows
|
|
140
|
+
```
|
|
141
|
+
|
|
142
|
+
## Interacting with Modules
|
|
143
|
+
|
|
144
|
+
```python
|
|
145
|
+
# Note: CLARIFAI_PAT must be set as env variable.
|
|
146
|
+
from clarifai.client.app import App
|
|
147
|
+
app = App(user_id="user_id", app_id="app_id")
|
|
148
|
+
|
|
149
|
+
# create a new module
|
|
150
|
+
module = app.create_module(module_id="module_id", description="module_description")
|
|
151
|
+
|
|
152
|
+
# List all modules in an app
|
|
153
|
+
all_modules = app.list_modules()
|
|
154
|
+
|
|
155
|
+
# List all module versions
|
|
156
|
+
all_module_versions = module.list_versions()
|
|
157
|
+
|
|
158
|
+
# Delete a module
|
|
159
|
+
app.delete_module(module_id="module_id")
|
|
160
|
+
```
|
|
@@ -14,12 +14,14 @@ base_https_cache = {}
|
|
|
14
14
|
ui_https_cache = {}
|
|
15
15
|
|
|
16
16
|
|
|
17
|
-
def clear_cache():
|
|
17
|
+
def clear_cache() -> None:
|
|
18
|
+
"""Clears the cache."""
|
|
18
19
|
base_https_cache.clear()
|
|
19
20
|
ui_https_cache.clear()
|
|
20
21
|
|
|
21
22
|
|
|
22
|
-
def https_cache(cache, url):
|
|
23
|
+
def https_cache(cache: dict, url: str) -> str:
|
|
24
|
+
"""This is a helper function to cache whether a url is http or https."""
|
|
23
25
|
HTTPS = True
|
|
24
26
|
HTTP = False
|
|
25
27
|
|
|
@@ -72,17 +74,17 @@ class ClarifaiAuthHelper:
|
|
|
72
74
|
Note: only one of personal access token (pat) or a session token (token) can be provided.
|
|
73
75
|
Always use PATs in your code and never session tokens, those are only provided internal UI code.
|
|
74
76
|
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
77
|
+
Args:
|
|
78
|
+
user_id: a user id who owns the resource you want to make calls to.
|
|
79
|
+
app_id: an app id for the application that owns the resource you want to interact with
|
|
80
|
+
pat: a personal access token.
|
|
81
|
+
token: a session token (internal use only, always use a PAT).
|
|
82
|
+
base: a url to the API endpoint to hit. Examples include api.clarifai.com,
|
|
83
|
+
https://api.clarifai.com (default), https://host:port, http://host:port, host:port (will be treated as http, not https). It's highly recommended to include the http:// or https:// otherwise we need to check the endpoint to determine if it has SSL during this __init__
|
|
84
|
+
ui: a url to the UI. Examples include clarifai.com,
|
|
85
|
+
https://clarifai.com (default), https://host:port, http://host:port, host:port (will be treated as http, not https). It's highly recommended to include the http:// or https:// otherwise we need to check the endpoint to determine if it has SSL during this __init__
|
|
86
|
+
validate: whether to validate the inputs. This is useful for overriding vars then validating
|
|
87
|
+
"""
|
|
86
88
|
|
|
87
89
|
self.user_id = user_id
|
|
88
90
|
self.app_id = app_id
|
|
@@ -182,10 +184,10 @@ class ClarifaiAuthHelper:
|
|
|
182
184
|
optionally:
|
|
183
185
|
base: as 'base' in query_params.
|
|
184
186
|
|
|
185
|
-
|
|
186
|
-
|
|
187
|
-
|
|
188
|
-
|
|
187
|
+
Args:
|
|
188
|
+
query_params: the streamlit.experimental_get_query_params() response or an empty dict to fall
|
|
189
|
+
back to using env vars.
|
|
190
|
+
"""
|
|
189
191
|
error_description = """
|
|
190
192
|
Please check the following required query params are in the url:
|
|
191
193
|
- 'user_id': the user ID accessing the module.
|
|
@@ -225,7 +227,7 @@ Additionally, these optional params are supported:
|
|
|
225
227
|
token: CLARIFAI_SESSION_TOKEN env var.
|
|
226
228
|
pat: CLARIFAI_PAT env var.
|
|
227
229
|
base: CLARIFAI_API_BASE env var.
|
|
228
|
-
|
|
230
|
+
"""
|
|
229
231
|
error_description = """
|
|
230
232
|
Please check the following required vars are in your env:
|
|
231
233
|
- 'CLARIFAI_USER_ID': the user ID accessing the module.
|
|
@@ -249,28 +251,26 @@ Additionally, these optional params are supported:
|
|
|
249
251
|
user_id: str = None,
|
|
250
252
|
app_id: str = None,
|
|
251
253
|
) -> resources_pb2.UserAppIDSet:
|
|
252
|
-
"""
|
|
253
|
-
Get the gRPC metadata that contains either the session token or the PAT to use.
|
|
254
|
+
"""Get the gRPC metadata that contains either the session token or the PAT to use.
|
|
254
255
|
|
|
255
|
-
|
|
256
|
-
|
|
257
|
-
|
|
256
|
+
Args:
|
|
257
|
+
user_id: optional user_id to override the default
|
|
258
|
+
app_id: optional app_id to override the default.
|
|
258
259
|
|
|
259
|
-
|
|
260
|
-
|
|
261
|
-
|
|
260
|
+
Returns:
|
|
261
|
+
metadata: the metadata need to send with all grpc API calls in the API client.
|
|
262
|
+
"""
|
|
262
263
|
user_id = self.user_id if user_id is None else user_id
|
|
263
264
|
app_id = self.app_id if app_id is None else app_id
|
|
264
265
|
return resources_pb2.UserAppIDSet(user_id=user_id, app_id=app_id)
|
|
265
266
|
|
|
266
267
|
@property
|
|
267
268
|
def metadata(self):
|
|
268
|
-
"""
|
|
269
|
-
Get the gRPC metadata that contains either the session token or the PAT to use.
|
|
269
|
+
"""Get the gRPC metadata that contains either the session token or the PAT to use.
|
|
270
270
|
|
|
271
|
-
|
|
272
|
-
|
|
273
|
-
|
|
271
|
+
Returns:
|
|
272
|
+
metadata: the metadata need to send with all grpc API calls in the API client.
|
|
273
|
+
"""
|
|
274
274
|
if self._pat != "":
|
|
275
275
|
return (("authorization", "Key %s" % self._pat),)
|
|
276
276
|
elif self._token != "":
|
|
@@ -279,12 +279,11 @@ Additionally, these optional params are supported:
|
|
|
279
279
|
raise Exception("'token' or 'pat' needed to be provided in the query params or env vars.")
|
|
280
280
|
|
|
281
281
|
def get_stub(self) -> service_pb2_grpc.V2Stub:
|
|
282
|
-
"""
|
|
283
|
-
Get the API gRPC stub using the right channel based on the API endpoint base.
|
|
282
|
+
"""Get the API gRPC stub using the right channel based on the API endpoint base.
|
|
284
283
|
|
|
285
|
-
|
|
286
|
-
|
|
287
|
-
|
|
284
|
+
Returns:
|
|
285
|
+
stub: The service_pb2_grpc.V2Stub stub for the API.
|
|
286
|
+
"""
|
|
288
287
|
if self._base not in base_https_cache:
|
|
289
288
|
raise Exception("Cannot determine if base %s is https" % self._base)
|
|
290
289
|
|
|
@@ -5,11 +5,11 @@ from clarifai_grpc.grpc.api import service_pb2_grpc
|
|
|
5
5
|
|
|
6
6
|
|
|
7
7
|
class V2Stub(abc.ABC):
|
|
8
|
-
|
|
8
|
+
"""Abstract base class of clarifai api rpc client stubs"""
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
class RpcCallable(abc.ABC):
|
|
12
|
-
|
|
12
|
+
"""Abstract base class of clarifai api rpc callables"""
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
# add grpc classes as subclasses of the abcs, so they also succeed in isinstance calls
|
|
@@ -5,8 +5,8 @@ from concurrent.futures import ThreadPoolExecutor
|
|
|
5
5
|
import grpc
|
|
6
6
|
from clarifai_grpc.grpc.api.status import status_code_pb2
|
|
7
7
|
|
|
8
|
-
from clarifai.auth.helper import ClarifaiAuthHelper
|
|
9
|
-
from clarifai.client.
|
|
8
|
+
from clarifai.client.auth.helper import ClarifaiAuthHelper
|
|
9
|
+
from clarifai.client.auth.register import RpcCallable, V2Stub
|
|
10
10
|
|
|
11
11
|
throttle_status_codes = {
|
|
12
12
|
status_code_pb2.CONN_THROTTLED,
|
|
@@ -20,15 +20,15 @@ retry_codes_grpc = {
|
|
|
20
20
|
_threadpool = ThreadPoolExecutor(100)
|
|
21
21
|
|
|
22
22
|
|
|
23
|
-
def create_stub(auth_helper=None, max_retry_attempts=10):
|
|
24
|
-
|
|
23
|
+
def create_stub(auth_helper: ClarifaiAuthHelper = None, max_retry_attempts: int = 10) -> V2Stub:
|
|
24
|
+
"""
|
|
25
25
|
Create client stub that handles authorization and basic retries for
|
|
26
26
|
unavailable or throttled connections.
|
|
27
27
|
|
|
28
28
|
Args:
|
|
29
29
|
auth_helper: ClarifaiAuthHelper to use for auth metadata (default: from env)
|
|
30
30
|
max_retry_attempts: max attempts to retry rpcs with retryable failures
|
|
31
|
-
|
|
31
|
+
"""
|
|
32
32
|
stub = AuthorizedStub(auth_helper)
|
|
33
33
|
if max_retry_attempts > 0:
|
|
34
34
|
return RetryStub(stub, max_retry_attempts)
|
|
@@ -36,9 +36,7 @@ def create_stub(auth_helper=None, max_retry_attempts=10):
|
|
|
36
36
|
|
|
37
37
|
|
|
38
38
|
class AuthorizedStub(V2Stub):
|
|
39
|
-
|
|
40
|
-
V2Stub proxy that inserts metadata authorization in rpc calls.
|
|
41
|
-
'''
|
|
39
|
+
"""V2Stub proxy that inserts metadata authorization in rpc calls."""
|
|
42
40
|
|
|
43
41
|
def __init__(self, auth_helper: ClarifaiAuthHelper = None):
|
|
44
42
|
if auth_helper is None:
|
|
@@ -54,6 +52,7 @@ class AuthorizedStub(V2Stub):
|
|
|
54
52
|
|
|
55
53
|
|
|
56
54
|
class _AuthorizedRpcCallable(RpcCallable):
|
|
55
|
+
"""Adds metadata(authorization header) to rpc calls"""
|
|
57
56
|
|
|
58
57
|
def __init__(self, func, metadata):
|
|
59
58
|
self.f = func
|
|
@@ -75,9 +74,9 @@ class _AuthorizedRpcCallable(RpcCallable):
|
|
|
75
74
|
|
|
76
75
|
|
|
77
76
|
class RetryStub(V2Stub):
|
|
78
|
-
|
|
77
|
+
"""
|
|
79
78
|
V2Stub proxy that retries requests (currently on unavailable server or throttle codes)
|
|
80
|
-
|
|
79
|
+
"""
|
|
81
80
|
|
|
82
81
|
def __init__(self, stub, max_attempts=10, backoff_time=5):
|
|
83
82
|
self.stub = stub
|
|
@@ -92,6 +91,7 @@ class RetryStub(V2Stub):
|
|
|
92
91
|
|
|
93
92
|
|
|
94
93
|
class _RetryRpcCallable(RpcCallable):
|
|
94
|
+
"""Retries rpc calls on unavailable server or throttle codes"""
|
|
95
95
|
|
|
96
96
|
def __init__(self, func, max_attempts, backoff_time):
|
|
97
97
|
self.f = func
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
# Note: This is just added for backwards compatibility. It will be removed in the future.
|