megadetector 5.0.10__py3-none-any.whl → 5.0.11__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of megadetector might be problematic. Click here for more details.
- {megadetector-5.0.10.dist-info → megadetector-5.0.11.dist-info}/LICENSE +0 -0
- {megadetector-5.0.10.dist-info → megadetector-5.0.11.dist-info}/METADATA +12 -11
- megadetector-5.0.11.dist-info/RECORD +5 -0
- megadetector-5.0.11.dist-info/top_level.txt +1 -0
- api/__init__.py +0 -0
- api/batch_processing/__init__.py +0 -0
- api/batch_processing/api_core/__init__.py +0 -0
- api/batch_processing/api_core/batch_service/__init__.py +0 -0
- api/batch_processing/api_core/batch_service/score.py +0 -439
- api/batch_processing/api_core/server.py +0 -294
- api/batch_processing/api_core/server_api_config.py +0 -98
- api/batch_processing/api_core/server_app_config.py +0 -55
- api/batch_processing/api_core/server_batch_job_manager.py +0 -220
- api/batch_processing/api_core/server_job_status_table.py +0 -152
- api/batch_processing/api_core/server_orchestration.py +0 -360
- api/batch_processing/api_core/server_utils.py +0 -92
- api/batch_processing/api_core_support/__init__.py +0 -0
- api/batch_processing/api_core_support/aggregate_results_manually.py +0 -46
- api/batch_processing/api_support/__init__.py +0 -0
- api/batch_processing/api_support/summarize_daily_activity.py +0 -152
- api/batch_processing/data_preparation/__init__.py +0 -0
- api/batch_processing/data_preparation/manage_local_batch.py +0 -2391
- api/batch_processing/data_preparation/manage_video_batch.py +0 -327
- api/batch_processing/integration/digiKam/setup.py +0 -6
- api/batch_processing/integration/digiKam/xmp_integration.py +0 -465
- api/batch_processing/integration/eMammal/test_scripts/config_template.py +0 -5
- api/batch_processing/integration/eMammal/test_scripts/push_annotations_to_emammal.py +0 -126
- api/batch_processing/integration/eMammal/test_scripts/select_images_for_testing.py +0 -55
- api/batch_processing/postprocessing/__init__.py +0 -0
- api/batch_processing/postprocessing/add_max_conf.py +0 -64
- api/batch_processing/postprocessing/categorize_detections_by_size.py +0 -163
- api/batch_processing/postprocessing/combine_api_outputs.py +0 -249
- api/batch_processing/postprocessing/compare_batch_results.py +0 -958
- api/batch_processing/postprocessing/convert_output_format.py +0 -397
- api/batch_processing/postprocessing/load_api_results.py +0 -195
- api/batch_processing/postprocessing/md_to_coco.py +0 -310
- api/batch_processing/postprocessing/md_to_labelme.py +0 -330
- api/batch_processing/postprocessing/merge_detections.py +0 -401
- api/batch_processing/postprocessing/postprocess_batch_results.py +0 -1904
- api/batch_processing/postprocessing/remap_detection_categories.py +0 -170
- api/batch_processing/postprocessing/render_detection_confusion_matrix.py +0 -661
- api/batch_processing/postprocessing/repeat_detection_elimination/find_repeat_detections.py +0 -211
- api/batch_processing/postprocessing/repeat_detection_elimination/remove_repeat_detections.py +0 -82
- api/batch_processing/postprocessing/repeat_detection_elimination/repeat_detections_core.py +0 -1631
- api/batch_processing/postprocessing/separate_detections_into_folders.py +0 -731
- api/batch_processing/postprocessing/subset_json_detector_output.py +0 -696
- api/batch_processing/postprocessing/top_folders_to_bottom.py +0 -223
- api/synchronous/__init__.py +0 -0
- api/synchronous/api_core/animal_detection_api/__init__.py +0 -0
- api/synchronous/api_core/animal_detection_api/api_backend.py +0 -152
- api/synchronous/api_core/animal_detection_api/api_frontend.py +0 -266
- api/synchronous/api_core/animal_detection_api/config.py +0 -35
- api/synchronous/api_core/animal_detection_api/data_management/annotations/annotation_constants.py +0 -47
- api/synchronous/api_core/animal_detection_api/detection/detector_training/copy_checkpoints.py +0 -43
- api/synchronous/api_core/animal_detection_api/detection/detector_training/model_main_tf2.py +0 -114
- api/synchronous/api_core/animal_detection_api/detection/process_video.py +0 -543
- api/synchronous/api_core/animal_detection_api/detection/pytorch_detector.py +0 -304
- api/synchronous/api_core/animal_detection_api/detection/run_detector.py +0 -627
- api/synchronous/api_core/animal_detection_api/detection/run_detector_batch.py +0 -1029
- api/synchronous/api_core/animal_detection_api/detection/run_inference_with_yolov5_val.py +0 -581
- api/synchronous/api_core/animal_detection_api/detection/run_tiled_inference.py +0 -754
- api/synchronous/api_core/animal_detection_api/detection/tf_detector.py +0 -165
- api/synchronous/api_core/animal_detection_api/detection/video_utils.py +0 -495
- api/synchronous/api_core/animal_detection_api/md_utils/azure_utils.py +0 -174
- api/synchronous/api_core/animal_detection_api/md_utils/ct_utils.py +0 -262
- api/synchronous/api_core/animal_detection_api/md_utils/directory_listing.py +0 -251
- api/synchronous/api_core/animal_detection_api/md_utils/matlab_porting_tools.py +0 -97
- api/synchronous/api_core/animal_detection_api/md_utils/path_utils.py +0 -416
- api/synchronous/api_core/animal_detection_api/md_utils/process_utils.py +0 -110
- api/synchronous/api_core/animal_detection_api/md_utils/sas_blob_utils.py +0 -509
- api/synchronous/api_core/animal_detection_api/md_utils/string_utils.py +0 -59
- api/synchronous/api_core/animal_detection_api/md_utils/url_utils.py +0 -144
- api/synchronous/api_core/animal_detection_api/md_utils/write_html_image_list.py +0 -226
- api/synchronous/api_core/animal_detection_api/md_visualization/visualization_utils.py +0 -841
- api/synchronous/api_core/tests/__init__.py +0 -0
- api/synchronous/api_core/tests/load_test.py +0 -110
- classification/__init__.py +0 -0
- classification/aggregate_classifier_probs.py +0 -108
- classification/analyze_failed_images.py +0 -227
- classification/cache_batchapi_outputs.py +0 -198
- classification/create_classification_dataset.py +0 -627
- classification/crop_detections.py +0 -516
- classification/csv_to_json.py +0 -226
- classification/detect_and_crop.py +0 -855
- classification/efficientnet/__init__.py +0 -9
- classification/efficientnet/model.py +0 -415
- classification/efficientnet/utils.py +0 -610
- classification/evaluate_model.py +0 -520
- classification/identify_mislabeled_candidates.py +0 -152
- classification/json_to_azcopy_list.py +0 -63
- classification/json_validator.py +0 -695
- classification/map_classification_categories.py +0 -276
- classification/merge_classification_detection_output.py +0 -506
- classification/prepare_classification_script.py +0 -194
- classification/prepare_classification_script_mc.py +0 -228
- classification/run_classifier.py +0 -286
- classification/save_mislabeled.py +0 -110
- classification/train_classifier.py +0 -825
- classification/train_classifier_tf.py +0 -724
- classification/train_utils.py +0 -322
- data_management/__init__.py +0 -0
- data_management/annotations/__init__.py +0 -0
- data_management/annotations/annotation_constants.py +0 -34
- data_management/camtrap_dp_to_coco.py +0 -238
- data_management/cct_json_utils.py +0 -395
- data_management/cct_to_md.py +0 -176
- data_management/cct_to_wi.py +0 -289
- data_management/coco_to_labelme.py +0 -272
- data_management/coco_to_yolo.py +0 -662
- data_management/databases/__init__.py +0 -0
- data_management/databases/add_width_and_height_to_db.py +0 -33
- data_management/databases/combine_coco_camera_traps_files.py +0 -206
- data_management/databases/integrity_check_json_db.py +0 -477
- data_management/databases/subset_json_db.py +0 -115
- data_management/generate_crops_from_cct.py +0 -149
- data_management/get_image_sizes.py +0 -188
- data_management/importers/add_nacti_sizes.py +0 -52
- data_management/importers/add_timestamps_to_icct.py +0 -79
- data_management/importers/animl_results_to_md_results.py +0 -158
- data_management/importers/auckland_doc_test_to_json.py +0 -372
- data_management/importers/auckland_doc_to_json.py +0 -200
- data_management/importers/awc_to_json.py +0 -189
- data_management/importers/bellevue_to_json.py +0 -273
- data_management/importers/cacophony-thermal-importer.py +0 -796
- data_management/importers/carrizo_shrubfree_2018.py +0 -268
- data_management/importers/carrizo_trail_cam_2017.py +0 -287
- data_management/importers/cct_field_adjustments.py +0 -57
- data_management/importers/channel_islands_to_cct.py +0 -913
- data_management/importers/eMammal/copy_and_unzip_emammal.py +0 -180
- data_management/importers/eMammal/eMammal_helpers.py +0 -249
- data_management/importers/eMammal/make_eMammal_json.py +0 -223
- data_management/importers/ena24_to_json.py +0 -275
- data_management/importers/filenames_to_json.py +0 -385
- data_management/importers/helena_to_cct.py +0 -282
- data_management/importers/idaho-camera-traps.py +0 -1407
- data_management/importers/idfg_iwildcam_lila_prep.py +0 -294
- data_management/importers/jb_csv_to_json.py +0 -150
- data_management/importers/mcgill_to_json.py +0 -250
- data_management/importers/missouri_to_json.py +0 -489
- data_management/importers/nacti_fieldname_adjustments.py +0 -79
- data_management/importers/noaa_seals_2019.py +0 -181
- data_management/importers/pc_to_json.py +0 -365
- data_management/importers/plot_wni_giraffes.py +0 -123
- data_management/importers/prepare-noaa-fish-data-for-lila.py +0 -359
- data_management/importers/prepare_zsl_imerit.py +0 -131
- data_management/importers/rspb_to_json.py +0 -356
- data_management/importers/save_the_elephants_survey_A.py +0 -320
- data_management/importers/save_the_elephants_survey_B.py +0 -332
- data_management/importers/snapshot_safari_importer.py +0 -758
- data_management/importers/snapshot_safari_importer_reprise.py +0 -665
- data_management/importers/snapshot_serengeti_lila.py +0 -1067
- data_management/importers/snapshotserengeti/make_full_SS_json.py +0 -150
- data_management/importers/snapshotserengeti/make_per_season_SS_json.py +0 -153
- data_management/importers/sulross_get_exif.py +0 -65
- data_management/importers/timelapse_csv_set_to_json.py +0 -490
- data_management/importers/ubc_to_json.py +0 -399
- data_management/importers/umn_to_json.py +0 -507
- data_management/importers/wellington_to_json.py +0 -263
- data_management/importers/wi_to_json.py +0 -441
- data_management/importers/zamba_results_to_md_results.py +0 -181
- data_management/labelme_to_coco.py +0 -548
- data_management/labelme_to_yolo.py +0 -272
- data_management/lila/__init__.py +0 -0
- data_management/lila/add_locations_to_island_camera_traps.py +0 -97
- data_management/lila/add_locations_to_nacti.py +0 -147
- data_management/lila/create_lila_blank_set.py +0 -557
- data_management/lila/create_lila_test_set.py +0 -151
- data_management/lila/create_links_to_md_results_files.py +0 -106
- data_management/lila/download_lila_subset.py +0 -177
- data_management/lila/generate_lila_per_image_labels.py +0 -515
- data_management/lila/get_lila_annotation_counts.py +0 -170
- data_management/lila/get_lila_image_counts.py +0 -111
- data_management/lila/lila_common.py +0 -300
- data_management/lila/test_lila_metadata_urls.py +0 -132
- data_management/ocr_tools.py +0 -874
- data_management/read_exif.py +0 -681
- data_management/remap_coco_categories.py +0 -84
- data_management/remove_exif.py +0 -66
- data_management/resize_coco_dataset.py +0 -189
- data_management/wi_download_csv_to_coco.py +0 -246
- data_management/yolo_output_to_md_output.py +0 -441
- data_management/yolo_to_coco.py +0 -676
- detection/__init__.py +0 -0
- detection/detector_training/__init__.py +0 -0
- detection/detector_training/model_main_tf2.py +0 -114
- detection/process_video.py +0 -703
- detection/pytorch_detector.py +0 -337
- detection/run_detector.py +0 -779
- detection/run_detector_batch.py +0 -1219
- detection/run_inference_with_yolov5_val.py +0 -917
- detection/run_tiled_inference.py +0 -935
- detection/tf_detector.py +0 -188
- detection/video_utils.py +0 -606
- docs/source/conf.py +0 -43
- md_utils/__init__.py +0 -0
- md_utils/azure_utils.py +0 -174
- md_utils/ct_utils.py +0 -612
- md_utils/directory_listing.py +0 -246
- md_utils/md_tests.py +0 -968
- md_utils/path_utils.py +0 -1044
- md_utils/process_utils.py +0 -157
- md_utils/sas_blob_utils.py +0 -509
- md_utils/split_locations_into_train_val.py +0 -228
- md_utils/string_utils.py +0 -92
- md_utils/url_utils.py +0 -323
- md_utils/write_html_image_list.py +0 -225
- md_visualization/__init__.py +0 -0
- md_visualization/plot_utils.py +0 -293
- md_visualization/render_images_with_thumbnails.py +0 -275
- md_visualization/visualization_utils.py +0 -1537
- md_visualization/visualize_db.py +0 -551
- md_visualization/visualize_detector_output.py +0 -406
- megadetector-5.0.10.dist-info/RECORD +0 -224
- megadetector-5.0.10.dist-info/top_level.txt +0 -8
- taxonomy_mapping/__init__.py +0 -0
- taxonomy_mapping/map_lila_taxonomy_to_wi_taxonomy.py +0 -491
- taxonomy_mapping/map_new_lila_datasets.py +0 -154
- taxonomy_mapping/prepare_lila_taxonomy_release.py +0 -142
- taxonomy_mapping/preview_lila_taxonomy.py +0 -591
- taxonomy_mapping/retrieve_sample_image.py +0 -71
- taxonomy_mapping/simple_image_download.py +0 -218
- taxonomy_mapping/species_lookup.py +0 -834
- taxonomy_mapping/taxonomy_csv_checker.py +0 -159
- taxonomy_mapping/taxonomy_graph.py +0 -346
- taxonomy_mapping/validate_lila_category_mappings.py +0 -83
- {megadetector-5.0.10.dist-info → megadetector-5.0.11.dist-info}/WHEEL +0 -0
|
@@ -1,223 +0,0 @@
|
|
|
1
|
-
"""
|
|
2
|
-
|
|
3
|
-
top_folders_to_bottom.py
|
|
4
|
-
|
|
5
|
-
Given a base folder with files like:
|
|
6
|
-
|
|
7
|
-
* A/1/2/a.jpg
|
|
8
|
-
* B/3/4/b.jpg
|
|
9
|
-
|
|
10
|
-
...moves the top-level folders to the bottom in a new output folder, i.e., creates:
|
|
11
|
-
|
|
12
|
-
* 1/2/A/a.jpg
|
|
13
|
-
* 3/4/B/b.jpg
|
|
14
|
-
|
|
15
|
-
In practice, this is used to make this:
|
|
16
|
-
|
|
17
|
-
animal/camera01/image01.jpg
|
|
18
|
-
|
|
19
|
-
...look like:
|
|
20
|
-
|
|
21
|
-
camera01/animal/image01.jpg
|
|
22
|
-
|
|
23
|
-
"""
|
|
24
|
-
|
|
25
|
-
#%% Constants and imports
|
|
26
|
-
|
|
27
|
-
import os
|
|
28
|
-
import sys
|
|
29
|
-
import shutil
|
|
30
|
-
import argparse
|
|
31
|
-
|
|
32
|
-
from pathlib import Path
|
|
33
|
-
from tqdm import tqdm
|
|
34
|
-
|
|
35
|
-
from functools import partial
|
|
36
|
-
from multiprocessing.pool import ThreadPool
|
|
37
|
-
|
|
38
|
-
from md_utils.path_utils import path_is_abs
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
#%% Classes
|
|
42
|
-
|
|
43
|
-
class TopFoldersToBottomOptions:
|
|
44
|
-
"""
|
|
45
|
-
Options used to parameterize top_folders_to_bottom()
|
|
46
|
-
"""
|
|
47
|
-
|
|
48
|
-
def __init__(self,input_folder,output_folder,copy=True,n_threads=1):
|
|
49
|
-
|
|
50
|
-
#: Whether to copy (True) vs. move (False) false when re-organizing
|
|
51
|
-
self.copy = copy
|
|
52
|
-
|
|
53
|
-
#: Number of worker threads to use, or <1 to disable parallelization
|
|
54
|
-
self.n_threads = n_threads
|
|
55
|
-
|
|
56
|
-
#: Input folder
|
|
57
|
-
self.input_folder = input_folder
|
|
58
|
-
|
|
59
|
-
#: Output folder
|
|
60
|
-
self.output_folder = output_folder
|
|
61
|
-
|
|
62
|
-
#: If this is False and an output file exists, throw an error
|
|
63
|
-
self.overwrite = False
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
#%% Main functions
|
|
67
|
-
|
|
68
|
-
def _process_file(relative_filename,options,execute=True):
|
|
69
|
-
|
|
70
|
-
assert ('/' in relative_filename) and \
|
|
71
|
-
('\\' not in relative_filename) and \
|
|
72
|
-
(not path_is_abs(relative_filename))
|
|
73
|
-
|
|
74
|
-
# Find top-level folder
|
|
75
|
-
tokens = relative_filename.split('/')
|
|
76
|
-
topmost_folder = tokens.pop(0)
|
|
77
|
-
tokens.insert(len(tokens)-1,topmost_folder)
|
|
78
|
-
|
|
79
|
-
# Find file/folder names
|
|
80
|
-
output_relative_path = '/'.join(tokens)
|
|
81
|
-
output_relative_folder = '/'.join(tokens[0:-1])
|
|
82
|
-
|
|
83
|
-
output_absolute_folder = os.path.join(options.output_folder,output_relative_folder)
|
|
84
|
-
output_absolute_path = os.path.join(options.output_folder,output_relative_path)
|
|
85
|
-
|
|
86
|
-
if execute:
|
|
87
|
-
|
|
88
|
-
os.makedirs(output_absolute_folder,exist_ok=True)
|
|
89
|
-
|
|
90
|
-
input_absolute_path = os.path.join(options.input_folder,relative_filename)
|
|
91
|
-
|
|
92
|
-
if not options.overwrite:
|
|
93
|
-
assert not os.path.isfile(output_absolute_path), \
|
|
94
|
-
'Error: output file {} exists'.format(output_absolute_path)
|
|
95
|
-
|
|
96
|
-
# Move or copy
|
|
97
|
-
if options.copy:
|
|
98
|
-
shutil.copy(input_absolute_path, output_absolute_path)
|
|
99
|
-
else:
|
|
100
|
-
shutil.move(input_absolute_path, output_absolute_path)
|
|
101
|
-
|
|
102
|
-
return output_absolute_path
|
|
103
|
-
|
|
104
|
-
# ...def _process_file()
|
|
105
|
-
|
|
106
|
-
|
|
107
|
-
def top_folders_to_bottom(options):
|
|
108
|
-
"""
|
|
109
|
-
top_folders_to_bottom.py
|
|
110
|
-
|
|
111
|
-
Given a base folder with files like:
|
|
112
|
-
|
|
113
|
-
* A/1/2/a.jpg
|
|
114
|
-
* B/3/4/b.jpg
|
|
115
|
-
|
|
116
|
-
...moves the top-level folders to the bottom in a new output folder, i.e., creates:
|
|
117
|
-
|
|
118
|
-
* 1/2/A/a.jpg
|
|
119
|
-
* 3/4/B/b.jpg
|
|
120
|
-
|
|
121
|
-
In practice, this is used to make this:
|
|
122
|
-
|
|
123
|
-
animal/camera01/image01.jpg
|
|
124
|
-
|
|
125
|
-
...look like:
|
|
126
|
-
|
|
127
|
-
camera01/animal/image01.jpg
|
|
128
|
-
|
|
129
|
-
Args:
|
|
130
|
-
options (TopFoldersToBottomOptions): See TopFoldersToBottomOptions for parameter details.
|
|
131
|
-
|
|
132
|
-
"""
|
|
133
|
-
os.makedirs(options.output_folder,exist_ok=True)
|
|
134
|
-
|
|
135
|
-
# Enumerate input folder
|
|
136
|
-
print('Enumerating files...')
|
|
137
|
-
files = list(Path(options.input_folder).rglob('*'))
|
|
138
|
-
files = [p for p in files if not p.is_dir()]
|
|
139
|
-
files = [str(s) for s in files]
|
|
140
|
-
print('Enumerated {} files'.format(len(files)))
|
|
141
|
-
|
|
142
|
-
# Convert absolute paths to relative paths
|
|
143
|
-
relative_files = [os.path.relpath(s,options.input_folder) for s in files]
|
|
144
|
-
|
|
145
|
-
# Standardize delimiters
|
|
146
|
-
relative_files = [s.replace('\\','/') for s in relative_files]
|
|
147
|
-
|
|
148
|
-
base_files = [s for s in relative_files if '/' not in s]
|
|
149
|
-
if len(base_files) > 0:
|
|
150
|
-
print('Warning: ignoring {} files in the base folder'.format(len(base_files)))
|
|
151
|
-
relative_files = [s for s in relative_files if '/' in s]
|
|
152
|
-
|
|
153
|
-
# Make sure each input file maps to a unique output file
|
|
154
|
-
absolute_output_files = [_process_file(s, options, execute=False) for s in relative_files]
|
|
155
|
-
assert len(absolute_output_files) == len(set(absolute_output_files)),\
|
|
156
|
-
"Error: input filenames don't map to unique output filenames"
|
|
157
|
-
|
|
158
|
-
# relative_filename = relative_files[0]
|
|
159
|
-
|
|
160
|
-
# Loop
|
|
161
|
-
if options.n_threads <= 1:
|
|
162
|
-
|
|
163
|
-
for relative_filename in tqdm(relative_files):
|
|
164
|
-
_process_file(relative_filename,options)
|
|
165
|
-
|
|
166
|
-
else:
|
|
167
|
-
|
|
168
|
-
print('Starting a pool with {} threads'.format(options.n_threads))
|
|
169
|
-
pool = ThreadPool(options.n_threads)
|
|
170
|
-
process_file_with_options = partial(_process_file, options=options)
|
|
171
|
-
_ = list(tqdm(pool.imap(process_file_with_options, relative_files), total=len(relative_files)))
|
|
172
|
-
|
|
173
|
-
# ...def top_folders_to_bottom()
|
|
174
|
-
|
|
175
|
-
|
|
176
|
-
#%% Interactive driver
|
|
177
|
-
|
|
178
|
-
if False:
|
|
179
|
-
|
|
180
|
-
pass
|
|
181
|
-
|
|
182
|
-
#%%
|
|
183
|
-
|
|
184
|
-
input_folder = r"G:\temp\output"
|
|
185
|
-
output_folder = r"G:\temp\output-inverted"
|
|
186
|
-
options = TopFoldersToBottomOptions(input_folder,output_folder,copy=True,n_threads=10)
|
|
187
|
-
|
|
188
|
-
#%%
|
|
189
|
-
|
|
190
|
-
top_folders_to_bottom(options)
|
|
191
|
-
|
|
192
|
-
|
|
193
|
-
#%% Command-line driver
|
|
194
|
-
|
|
195
|
-
# python top_folders_to_bottom.py "g:\temp\separated_images" "g:\temp\separated_images_inverted" --n_threads 100
|
|
196
|
-
|
|
197
|
-
def main():
|
|
198
|
-
|
|
199
|
-
parser = argparse.ArgumentParser()
|
|
200
|
-
parser.add_argument('input_folder', type=str, help='Input image folder')
|
|
201
|
-
parser.add_argument('output_folder', type=str, help='Output image folder')
|
|
202
|
-
|
|
203
|
-
parser.add_argument('--copy', action='store_true',
|
|
204
|
-
help='Copy images, instead of moving (moving is the default)')
|
|
205
|
-
parser.add_argument('--overwrite', action='store_true',
|
|
206
|
-
help='Allow image overwrite (default=False)')
|
|
207
|
-
parser.add_argument('--n_threads', type=int, default=1,
|
|
208
|
-
help='Number of threads to use for parallel operation (default=1)')
|
|
209
|
-
|
|
210
|
-
if len(sys.argv[1:])==0:
|
|
211
|
-
parser.print_help()
|
|
212
|
-
parser.exit()
|
|
213
|
-
|
|
214
|
-
args = parser.parse_args()
|
|
215
|
-
|
|
216
|
-
# Convert to an options object
|
|
217
|
-
options = TopFoldersToBottomOptions(
|
|
218
|
-
args.input_folder,args.output_folder,copy=args.copy,n_threads=args.n_threads)
|
|
219
|
-
|
|
220
|
-
top_folders_to_bottom(options)
|
|
221
|
-
|
|
222
|
-
if __name__ == '__main__':
|
|
223
|
-
main()
|
api/synchronous/__init__.py
DELETED
|
File without changes
|
|
File without changes
|
|
@@ -1,152 +0,0 @@
|
|
|
1
|
-
#
|
|
2
|
-
# api_backend.py
|
|
3
|
-
#
|
|
4
|
-
# Defines the model execution service, which pulls requests (one or more images)
|
|
5
|
-
# from the shared Redis queue, and runs them through the TF model.
|
|
6
|
-
#
|
|
7
|
-
|
|
8
|
-
#%% Imports
|
|
9
|
-
|
|
10
|
-
import os
|
|
11
|
-
import json
|
|
12
|
-
import time
|
|
13
|
-
import redis
|
|
14
|
-
import argparse
|
|
15
|
-
import PIL
|
|
16
|
-
|
|
17
|
-
from io import BytesIO
|
|
18
|
-
|
|
19
|
-
from detection.run_detector import load_detector, convert_to_tf_coords
|
|
20
|
-
import config
|
|
21
|
-
import md_visualization.visualization_utils as vis_utils
|
|
22
|
-
|
|
23
|
-
#%% Initialization
|
|
24
|
-
|
|
25
|
-
db = redis.StrictRedis(host=config.REDIS_HOST, port=config.REDIS_PORT)
|
|
26
|
-
current_directory = os.path.dirname(os.path.realpath(__file__))
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
#%% Main loop
|
|
30
|
-
|
|
31
|
-
def detect_process():
|
|
32
|
-
|
|
33
|
-
while True:
|
|
34
|
-
|
|
35
|
-
# TODO: convert to a blocking read and eliminate the sleep() statement in this loop
|
|
36
|
-
serialized_entry = db.lpop(config.REDIS_QUEUE_NAME)
|
|
37
|
-
all_detection_results = []
|
|
38
|
-
inference_time_detector = []
|
|
39
|
-
|
|
40
|
-
if serialized_entry:
|
|
41
|
-
|
|
42
|
-
entry = json.loads(serialized_entry)
|
|
43
|
-
id = entry['id']
|
|
44
|
-
print('Processing images from request id:', id)
|
|
45
|
-
return_confidence_threshold = entry['return_confidence_threshold']
|
|
46
|
-
|
|
47
|
-
try:
|
|
48
|
-
|
|
49
|
-
temp_direc = os.path.join(config.TEMP_FOLDER,id)
|
|
50
|
-
assert os.path.isdir(temp_direc), 'Could not find temporary folder {}'.format(temp_direc)
|
|
51
|
-
|
|
52
|
-
for filename in os.listdir(temp_direc):
|
|
53
|
-
|
|
54
|
-
image_path = f'{temp_direc}/{filename}'
|
|
55
|
-
print('Reading image from {}'.format(image_path))
|
|
56
|
-
image = open(image_path, 'rb')
|
|
57
|
-
image = vis_utils.load_image(image)
|
|
58
|
-
|
|
59
|
-
start_time = time.time()
|
|
60
|
-
result = detector.generate_detections_one_image(image, filename, detection_threshold=config.DEFAULT_CONFIDENCE_THRESHOLD)
|
|
61
|
-
all_detection_results.append(result)
|
|
62
|
-
|
|
63
|
-
elapsed = time.time() - start_time
|
|
64
|
-
inference_time_detector.append(elapsed)
|
|
65
|
-
|
|
66
|
-
except Exception as e:
|
|
67
|
-
|
|
68
|
-
print('Detection error: ' + str(e))
|
|
69
|
-
|
|
70
|
-
db.set(entry['id'], json.dumps({
|
|
71
|
-
'status': 500,
|
|
72
|
-
'error': 'Detection error: ' + str(e)
|
|
73
|
-
}))
|
|
74
|
-
|
|
75
|
-
continue
|
|
76
|
-
|
|
77
|
-
# Filter the detections by the confidence threshold
|
|
78
|
-
#
|
|
79
|
-
# Each result is [ymin, xmin, ymax, xmax, confidence, category]
|
|
80
|
-
#
|
|
81
|
-
# Coordinates are relative, with the origin in the upper-left
|
|
82
|
-
detections = {}
|
|
83
|
-
|
|
84
|
-
try:
|
|
85
|
-
|
|
86
|
-
for result in all_detection_results:
|
|
87
|
-
|
|
88
|
-
image_name = result['file']
|
|
89
|
-
_detections = result.get('detections', None)
|
|
90
|
-
detections[image_name] = []
|
|
91
|
-
|
|
92
|
-
if _detections is None:
|
|
93
|
-
continue
|
|
94
|
-
|
|
95
|
-
for d in _detections:
|
|
96
|
-
if d['conf'] > return_confidence_threshold:
|
|
97
|
-
res = convert_to_tf_coords(d['bbox'])
|
|
98
|
-
res.append(d['conf'])
|
|
99
|
-
res.append(int(d['category']))
|
|
100
|
-
detections[image_name].append(res)
|
|
101
|
-
|
|
102
|
-
db.set(entry['id'], json.dumps({
|
|
103
|
-
'status': 200,
|
|
104
|
-
'detections': detections,
|
|
105
|
-
'inference_time_detector': inference_time_detector
|
|
106
|
-
}))
|
|
107
|
-
|
|
108
|
-
except Exception as e:
|
|
109
|
-
print('Error consolidating the detection boxes: ' + str(e))
|
|
110
|
-
|
|
111
|
-
db.set(entry['id'], json.dumps({
|
|
112
|
-
'status': 500,
|
|
113
|
-
'error': 'Error consolidating the detection boxes:' + str(e)
|
|
114
|
-
}))
|
|
115
|
-
|
|
116
|
-
# ...if serialized_entry
|
|
117
|
-
|
|
118
|
-
else:
|
|
119
|
-
time.sleep(0.005)
|
|
120
|
-
|
|
121
|
-
# ...while(True)
|
|
122
|
-
|
|
123
|
-
# ...def detect_process()
|
|
124
|
-
|
|
125
|
-
|
|
126
|
-
#%% Command-line driver
|
|
127
|
-
|
|
128
|
-
if __name__ == '__main__':
|
|
129
|
-
|
|
130
|
-
parser = argparse.ArgumentParser(description='api backend')
|
|
131
|
-
|
|
132
|
-
# use --non-docker if you are testing without Docker
|
|
133
|
-
#
|
|
134
|
-
# python api_frontend.py --non-docker
|
|
135
|
-
parser.add_argument('--non-docker', action='store_true', default=False)
|
|
136
|
-
args = parser.parse_args()
|
|
137
|
-
|
|
138
|
-
if args.non_docker:
|
|
139
|
-
model_path = config.DETECTOR_MODEL_PATH_DEBUG
|
|
140
|
-
else:
|
|
141
|
-
model_path = config.DETECTOR_MODEL_PATH
|
|
142
|
-
|
|
143
|
-
detector = load_detector(model_path)
|
|
144
|
-
|
|
145
|
-
# run detections on a test image to load the model
|
|
146
|
-
print('Running initial detection to load model...')
|
|
147
|
-
test_image = PIL.Image.new(mode="RGB", size=(200, 200))
|
|
148
|
-
result = detector.generate_detections_one_image(test_image, "test_image", detection_threshold=config.DEFAULT_CONFIDENCE_THRESHOLD)
|
|
149
|
-
print(result)
|
|
150
|
-
print('\n')
|
|
151
|
-
|
|
152
|
-
detect_process()
|
|
@@ -1,266 +0,0 @@
|
|
|
1
|
-
#
|
|
2
|
-
# api_frontend.py
|
|
3
|
-
#
|
|
4
|
-
# Defines the Flask app, which takes requests (one or more images) from
|
|
5
|
-
# remote callers and pushes the images onto the shared Redis queue, to be processed
|
|
6
|
-
# by the main service in api_backend.py .
|
|
7
|
-
#
|
|
8
|
-
|
|
9
|
-
#%% Imports
|
|
10
|
-
|
|
11
|
-
import os
|
|
12
|
-
import json
|
|
13
|
-
import time
|
|
14
|
-
import uuid
|
|
15
|
-
import redis
|
|
16
|
-
import shutil
|
|
17
|
-
import argparse
|
|
18
|
-
import traceback
|
|
19
|
-
|
|
20
|
-
from io import BytesIO
|
|
21
|
-
from flask import Flask, Response, jsonify, make_response, request
|
|
22
|
-
from requests_toolbelt.multipart.encoder import MultipartEncoder
|
|
23
|
-
|
|
24
|
-
import md_visualization.visualization_utils as vis_utils
|
|
25
|
-
import config
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
#%% Initialization
|
|
29
|
-
|
|
30
|
-
app = Flask(__name__)
|
|
31
|
-
db = redis.StrictRedis(host=config.REDIS_HOST, port=config.REDIS_PORT)
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
#%% Support functions
|
|
35
|
-
|
|
36
|
-
def _make_error_object(error_code, error_message):
|
|
37
|
-
|
|
38
|
-
# Make a dict that the request_processing_function can return to the endpoint
|
|
39
|
-
# function to notify it of an error
|
|
40
|
-
return {
|
|
41
|
-
'error_message': error_message,
|
|
42
|
-
'error_code': error_code
|
|
43
|
-
}
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
def _make_error_response(error_code, error_message):
|
|
47
|
-
|
|
48
|
-
return make_response(jsonify({'error': error_message}), error_code)
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
def has_access(request):
|
|
52
|
-
|
|
53
|
-
if not os.path.exists(config.API_KEYS_FILE):
|
|
54
|
-
return True
|
|
55
|
-
else:
|
|
56
|
-
if not request.headers.get('key'):
|
|
57
|
-
print('Key header not available')
|
|
58
|
-
return False
|
|
59
|
-
else:
|
|
60
|
-
API_key = request.headers.get('key').strip().lower()
|
|
61
|
-
with open(config.API_KEYS_FILE, "r") as f:
|
|
62
|
-
for line in f:
|
|
63
|
-
valid_key = line.strip().lower()
|
|
64
|
-
if valid_key == API_key:
|
|
65
|
-
return True
|
|
66
|
-
|
|
67
|
-
return False
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
def check_posted_data(request):
|
|
71
|
-
|
|
72
|
-
files = request.files
|
|
73
|
-
params = request.args
|
|
74
|
-
|
|
75
|
-
# Verify that the content uploaded is not too big
|
|
76
|
-
#
|
|
77
|
-
# request.content_length is the length of the total payload
|
|
78
|
-
content_length = request.content_length
|
|
79
|
-
if not content_length:
|
|
80
|
-
return _make_error_object(411, 'No image(s) were sent, or content length cannot be determined.')
|
|
81
|
-
if content_length > config.MAX_CONTENT_LENGTH_IN_MB * 1024 * 1024:
|
|
82
|
-
return _make_error_object(413, ('Payload size {:.2f} MB exceeds the maximum allowed of {} MB. '
|
|
83
|
-
'Please upload fewer or more compressed images.').format(
|
|
84
|
-
content_length / (1024 * 1024), config.MAX_CONTENT_LENGTH_IN_MB))
|
|
85
|
-
|
|
86
|
-
render_boxes = True if params.get('render', '').lower() == 'true' else False
|
|
87
|
-
|
|
88
|
-
if 'min_confidence' in params:
|
|
89
|
-
return_confidence_threshold = float(params['min_confidence'])
|
|
90
|
-
print('runserver, post_detect_sync, user specified detection confidence: ', return_confidence_threshold)
|
|
91
|
-
if return_confidence_threshold < 0.0 or return_confidence_threshold > 1.0:
|
|
92
|
-
return _make_error_object(400, 'Detection confidence threshold {} is invalid, should be between 0.0 and 1.0.'.format(
|
|
93
|
-
return_confidence_threshold))
|
|
94
|
-
else:
|
|
95
|
-
return_confidence_threshold = config.DEFAULT_CONFIDENCE_THRESHOLD
|
|
96
|
-
|
|
97
|
-
if 'min_rendering_confidence' in params:
|
|
98
|
-
rendering_confidence_threshold = float(params['min_rendering_confidence'])
|
|
99
|
-
print('runserver, post_detect_sync, user specified rendering confidence: ', rendering_confidence_threshold)
|
|
100
|
-
if rendering_confidence_threshold < 0.0 or rendering_confidence_threshold > 1.0:
|
|
101
|
-
return _make_error_object(400, 'Rendering confidence threshold {} is invalid, should be between 0.0 and 1.0.'.format(
|
|
102
|
-
rendering_confidence_threshold))
|
|
103
|
-
else:
|
|
104
|
-
rendering_confidence_threshold = config.DEFAULT_RENDERING_CONFIDENCE_THRESHOLD
|
|
105
|
-
|
|
106
|
-
# Verify that the number of images is acceptable
|
|
107
|
-
num_images = sum([1 if file.content_type in config.IMAGE_CONTENT_TYPES else 0 for file in files.values()])
|
|
108
|
-
print('runserver, post_detect_sync, number of images received: ', num_images)
|
|
109
|
-
|
|
110
|
-
if num_images > config.MAX_IMAGES_ACCEPTED:
|
|
111
|
-
return _make_error_object(413, 'Too many images. Maximum number of images that can be processed in one call is {}.'.format(config.MAX_IMAGES_ACCEPTED))
|
|
112
|
-
elif num_images == 0:
|
|
113
|
-
return _make_error_object(400, 'No image(s) of accepted types (image/jpeg, image/png, application/octet-stream) received.')
|
|
114
|
-
|
|
115
|
-
return {
|
|
116
|
-
'render_boxes': render_boxes,
|
|
117
|
-
'return_confidence_threshold': return_confidence_threshold,
|
|
118
|
-
'rendering_confidence_threshold': rendering_confidence_threshold
|
|
119
|
-
}
|
|
120
|
-
|
|
121
|
-
# ...def check_posted_data(request)
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
#%% Main loop
|
|
125
|
-
|
|
126
|
-
@app.route(config.API_PREFIX + '/detect', methods = ['POST'])
|
|
127
|
-
def detect_sync():
|
|
128
|
-
|
|
129
|
-
if not has_access(request):
|
|
130
|
-
print('Access denied, please provide a valid API key')
|
|
131
|
-
return _make_error_response(403, 'Access denied, please provide a valid API key')
|
|
132
|
-
|
|
133
|
-
# Check whether the request_processing_function had an error
|
|
134
|
-
post_data = check_posted_data(request)
|
|
135
|
-
if post_data.get('error_code', None) is not None:
|
|
136
|
-
return _make_error_response(post_data.get('error_code'), post_data.get('error_message'))
|
|
137
|
-
|
|
138
|
-
render_boxes = post_data.get('render_boxes')
|
|
139
|
-
return_confidence_threshold = post_data.get('return_confidence_threshold')
|
|
140
|
-
rendering_confidence_threshold = post_data.get('rendering_confidence_threshold')
|
|
141
|
-
|
|
142
|
-
redis_id = str(uuid.uuid4())
|
|
143
|
-
d = {'id': redis_id, 'render_boxes': render_boxes, 'return_confidence_threshold': return_confidence_threshold}
|
|
144
|
-
temp_direc = os.path.join(config.TEMP_FOLDER, redis_id)
|
|
145
|
-
|
|
146
|
-
try:
|
|
147
|
-
|
|
148
|
-
try:
|
|
149
|
-
# Write images to temporary files
|
|
150
|
-
#
|
|
151
|
-
# TODO: read from memory rather than using intermediate files
|
|
152
|
-
os.makedirs(temp_direc,exist_ok=True)
|
|
153
|
-
for name, file in request.files.items():
|
|
154
|
-
if file.content_type in config.IMAGE_CONTENT_TYPES:
|
|
155
|
-
filename = request.files[name].filename
|
|
156
|
-
image_path = os.path.join(temp_direc, filename)
|
|
157
|
-
print('Saving image {} to {}'.format(name,image_path))
|
|
158
|
-
file.save(image_path)
|
|
159
|
-
assert os.path.isfile(image_path),'Error creating file {}'.format(image_path)
|
|
160
|
-
|
|
161
|
-
except Exception as e:
|
|
162
|
-
return _make_error_object(500, 'Error saving images: ' + str(e))
|
|
163
|
-
|
|
164
|
-
# Submit the image(s) for processing by api_backend.py, who is waiting on this queue
|
|
165
|
-
db.rpush(config.REDIS_QUEUE_NAME, json.dumps(d))
|
|
166
|
-
|
|
167
|
-
while True:
|
|
168
|
-
|
|
169
|
-
# TODO: convert to a blocking read and eliminate the sleep() statement in this loop
|
|
170
|
-
result = db.get(redis_id)
|
|
171
|
-
|
|
172
|
-
if result:
|
|
173
|
-
|
|
174
|
-
result = json.loads(result.decode())
|
|
175
|
-
print('Processing result {}'.format(str(result)))
|
|
176
|
-
|
|
177
|
-
if result['status'] == 200:
|
|
178
|
-
detections = result['detections']
|
|
179
|
-
db.delete(redis_id)
|
|
180
|
-
|
|
181
|
-
else:
|
|
182
|
-
db.delete(redis_id)
|
|
183
|
-
print('Detection error: ' + str(result))
|
|
184
|
-
return _make_error_response(500, 'Detection error: ' + str(result))
|
|
185
|
-
|
|
186
|
-
try:
|
|
187
|
-
print('detect_sync: postprocessing and sending images back...')
|
|
188
|
-
fields = {
|
|
189
|
-
'detection_result': ('detection_result', json.dumps(detections), 'application/json'),
|
|
190
|
-
}
|
|
191
|
-
|
|
192
|
-
if render_boxes and result['status'] == 200:
|
|
193
|
-
|
|
194
|
-
print('Rendering images')
|
|
195
|
-
|
|
196
|
-
for image_name, detections in detections.items():
|
|
197
|
-
|
|
198
|
-
#image = Image.open(os.path.join(temp_direc, image_name))
|
|
199
|
-
image = open(f'{temp_direc}/{image_name}', "rb")
|
|
200
|
-
image = vis_utils.load_image(image)
|
|
201
|
-
width, height = image.size
|
|
202
|
-
|
|
203
|
-
_detections = []
|
|
204
|
-
for d in detections:
|
|
205
|
-
y1,x1,y2,x2 = d[0:4]
|
|
206
|
-
width = x2 - x1
|
|
207
|
-
height = y2 - y1
|
|
208
|
-
bbox = [x1,y1,width,height]
|
|
209
|
-
_detections.append({'bbox': bbox, 'conf': d[4], 'category': d[5]})
|
|
210
|
-
|
|
211
|
-
vis_utils.render_detection_bounding_boxes(_detections, image,
|
|
212
|
-
confidence_threshold=rendering_confidence_threshold)
|
|
213
|
-
|
|
214
|
-
output_img_stream = BytesIO()
|
|
215
|
-
image.save(output_img_stream, format='jpeg')
|
|
216
|
-
output_img_stream.seek(0)
|
|
217
|
-
fields[image_name] = (image_name, output_img_stream, 'image/jpeg')
|
|
218
|
-
print('Done rendering images')
|
|
219
|
-
|
|
220
|
-
m = MultipartEncoder(fields=fields)
|
|
221
|
-
return Response(m.to_string(), mimetype=m.content_type)
|
|
222
|
-
|
|
223
|
-
except Exception as e:
|
|
224
|
-
|
|
225
|
-
print(traceback.format_exc())
|
|
226
|
-
print('Error returning result or rendering the detection boxes: ' + str(e))
|
|
227
|
-
|
|
228
|
-
finally:
|
|
229
|
-
|
|
230
|
-
try:
|
|
231
|
-
print('Removing temporary files')
|
|
232
|
-
shutil.rmtree(temp_direc)
|
|
233
|
-
except Exception as e:
|
|
234
|
-
print('Error removing temporary folder {}: {}'.format(temp_direc,str(e)))
|
|
235
|
-
|
|
236
|
-
else:
|
|
237
|
-
time.sleep(0.005)
|
|
238
|
-
|
|
239
|
-
# ...if we do/don't have a request available on the queue
|
|
240
|
-
|
|
241
|
-
# ...while(True)
|
|
242
|
-
|
|
243
|
-
except Exception as e:
|
|
244
|
-
|
|
245
|
-
print(traceback.format_exc())
|
|
246
|
-
return _make_error_object(500, 'Error processing images: ' + str(e))
|
|
247
|
-
|
|
248
|
-
# ...def detect_sync()
|
|
249
|
-
|
|
250
|
-
|
|
251
|
-
#%% Command-line driver
|
|
252
|
-
|
|
253
|
-
if __name__ == '__main__':
|
|
254
|
-
|
|
255
|
-
parser = argparse.ArgumentParser(description='api frontend')
|
|
256
|
-
|
|
257
|
-
# use --non-docker if you are testing without Docker
|
|
258
|
-
#
|
|
259
|
-
# python api_frontend.py --non-docker
|
|
260
|
-
parser.add_argument('--non-docker', action="store_true", default=False)
|
|
261
|
-
args = parser.parse_args()
|
|
262
|
-
|
|
263
|
-
if args.non_docker:
|
|
264
|
-
app.run(host='0.0.0.0', port=5050)
|
|
265
|
-
else:
|
|
266
|
-
app.run()
|
|
@@ -1,35 +0,0 @@
|
|
|
1
|
-
## Camera trap real-time API configuration
|
|
2
|
-
|
|
3
|
-
REDIS_HOST = 'localhost'
|
|
4
|
-
|
|
5
|
-
REDIS_PORT = 6379
|
|
6
|
-
|
|
7
|
-
# Full path to the temporary folder for image storage, only meaningful
|
|
8
|
-
# within the Docker container
|
|
9
|
-
TEMP_FOLDER = '/app/temp'
|
|
10
|
-
|
|
11
|
-
REDIS_QUEUE_NAME = 'camera-trap-queue'
|
|
12
|
-
|
|
13
|
-
# Upper limit on total content length (all images and parameters)
|
|
14
|
-
MAX_CONTENT_LENGTH_IN_MB = 5 * 8 # 5MB per image * number of images allowed
|
|
15
|
-
|
|
16
|
-
MAX_IMAGES_ACCEPTED = 8
|
|
17
|
-
|
|
18
|
-
IMAGE_CONTENT_TYPES = ['image/png', 'application/octet-stream', 'image/jpeg']
|
|
19
|
-
|
|
20
|
-
DETECTOR_MODEL_PATH = '/app/animal_detection_api/model/md_v5a.0.0.pt'
|
|
21
|
-
|
|
22
|
-
DETECTOR_MODEL_VERSION = 'v5a.0.0'
|
|
23
|
-
|
|
24
|
-
# Minimum confidence threshold for detections
|
|
25
|
-
DEFAULT_CONFIDENCE_THRESHOLD = 0.01
|
|
26
|
-
|
|
27
|
-
# Minimum confidence threshold for showing a bounding box on the output image
|
|
28
|
-
DEFAULT_RENDERING_CONFIDENCE_THRESHOLD = 0.2
|
|
29
|
-
|
|
30
|
-
API_PREFIX = '/v1/camera-trap/sync'
|
|
31
|
-
|
|
32
|
-
API_KEYS_FILE = 'allowed_keys.txt'
|
|
33
|
-
|
|
34
|
-
# Use this when testing without Docker
|
|
35
|
-
DETECTOR_MODEL_PATH_DEBUG = 'model/md_v5a.0.0.pt'
|