megadetector 5.0.11__py3-none-any.whl → 5.0.12__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of megadetector might be problematic. Click here for more details.

Files changed (201) hide show
  1. megadetector/api/__init__.py +0 -0
  2. megadetector/api/batch_processing/__init__.py +0 -0
  3. megadetector/api/batch_processing/api_core/__init__.py +0 -0
  4. megadetector/api/batch_processing/api_core/batch_service/__init__.py +0 -0
  5. megadetector/api/batch_processing/api_core/batch_service/score.py +439 -0
  6. megadetector/api/batch_processing/api_core/server.py +294 -0
  7. megadetector/api/batch_processing/api_core/server_api_config.py +98 -0
  8. megadetector/api/batch_processing/api_core/server_app_config.py +55 -0
  9. megadetector/api/batch_processing/api_core/server_batch_job_manager.py +220 -0
  10. megadetector/api/batch_processing/api_core/server_job_status_table.py +152 -0
  11. megadetector/api/batch_processing/api_core/server_orchestration.py +360 -0
  12. megadetector/api/batch_processing/api_core/server_utils.py +92 -0
  13. megadetector/api/batch_processing/api_core_support/__init__.py +0 -0
  14. megadetector/api/batch_processing/api_core_support/aggregate_results_manually.py +46 -0
  15. megadetector/api/batch_processing/api_support/__init__.py +0 -0
  16. megadetector/api/batch_processing/api_support/summarize_daily_activity.py +152 -0
  17. megadetector/api/batch_processing/data_preparation/__init__.py +0 -0
  18. megadetector/api/batch_processing/integration/digiKam/setup.py +6 -0
  19. megadetector/api/batch_processing/integration/digiKam/xmp_integration.py +465 -0
  20. megadetector/api/batch_processing/integration/eMammal/test_scripts/config_template.py +5 -0
  21. megadetector/api/batch_processing/integration/eMammal/test_scripts/push_annotations_to_emammal.py +126 -0
  22. megadetector/api/batch_processing/integration/eMammal/test_scripts/select_images_for_testing.py +55 -0
  23. megadetector/api/synchronous/__init__.py +0 -0
  24. megadetector/api/synchronous/api_core/animal_detection_api/__init__.py +0 -0
  25. megadetector/api/synchronous/api_core/animal_detection_api/api_backend.py +152 -0
  26. megadetector/api/synchronous/api_core/animal_detection_api/api_frontend.py +266 -0
  27. megadetector/api/synchronous/api_core/animal_detection_api/config.py +35 -0
  28. megadetector/api/synchronous/api_core/tests/__init__.py +0 -0
  29. megadetector/api/synchronous/api_core/tests/load_test.py +110 -0
  30. megadetector/classification/__init__.py +0 -0
  31. megadetector/classification/aggregate_classifier_probs.py +108 -0
  32. megadetector/classification/analyze_failed_images.py +227 -0
  33. megadetector/classification/cache_batchapi_outputs.py +198 -0
  34. megadetector/classification/create_classification_dataset.py +627 -0
  35. megadetector/classification/crop_detections.py +516 -0
  36. megadetector/classification/csv_to_json.py +226 -0
  37. megadetector/classification/detect_and_crop.py +855 -0
  38. megadetector/classification/efficientnet/__init__.py +9 -0
  39. megadetector/classification/efficientnet/model.py +415 -0
  40. megadetector/classification/efficientnet/utils.py +610 -0
  41. megadetector/classification/evaluate_model.py +520 -0
  42. megadetector/classification/identify_mislabeled_candidates.py +152 -0
  43. megadetector/classification/json_to_azcopy_list.py +63 -0
  44. megadetector/classification/json_validator.py +699 -0
  45. megadetector/classification/map_classification_categories.py +276 -0
  46. megadetector/classification/merge_classification_detection_output.py +506 -0
  47. megadetector/classification/prepare_classification_script.py +194 -0
  48. megadetector/classification/prepare_classification_script_mc.py +228 -0
  49. megadetector/classification/run_classifier.py +287 -0
  50. megadetector/classification/save_mislabeled.py +110 -0
  51. megadetector/classification/train_classifier.py +827 -0
  52. megadetector/classification/train_classifier_tf.py +725 -0
  53. megadetector/classification/train_utils.py +323 -0
  54. megadetector/data_management/__init__.py +0 -0
  55. megadetector/data_management/annotations/__init__.py +0 -0
  56. megadetector/data_management/annotations/annotation_constants.py +34 -0
  57. megadetector/data_management/camtrap_dp_to_coco.py +239 -0
  58. megadetector/data_management/cct_json_utils.py +395 -0
  59. megadetector/data_management/cct_to_md.py +176 -0
  60. megadetector/data_management/cct_to_wi.py +289 -0
  61. megadetector/data_management/coco_to_labelme.py +272 -0
  62. megadetector/data_management/coco_to_yolo.py +662 -0
  63. megadetector/data_management/databases/__init__.py +0 -0
  64. megadetector/data_management/databases/add_width_and_height_to_db.py +33 -0
  65. megadetector/data_management/databases/combine_coco_camera_traps_files.py +206 -0
  66. megadetector/data_management/databases/integrity_check_json_db.py +477 -0
  67. megadetector/data_management/databases/subset_json_db.py +115 -0
  68. megadetector/data_management/generate_crops_from_cct.py +149 -0
  69. megadetector/data_management/get_image_sizes.py +189 -0
  70. megadetector/data_management/importers/add_nacti_sizes.py +52 -0
  71. megadetector/data_management/importers/add_timestamps_to_icct.py +79 -0
  72. megadetector/data_management/importers/animl_results_to_md_results.py +158 -0
  73. megadetector/data_management/importers/auckland_doc_test_to_json.py +373 -0
  74. megadetector/data_management/importers/auckland_doc_to_json.py +201 -0
  75. megadetector/data_management/importers/awc_to_json.py +191 -0
  76. megadetector/data_management/importers/bellevue_to_json.py +273 -0
  77. megadetector/data_management/importers/cacophony-thermal-importer.py +796 -0
  78. megadetector/data_management/importers/carrizo_shrubfree_2018.py +269 -0
  79. megadetector/data_management/importers/carrizo_trail_cam_2017.py +289 -0
  80. megadetector/data_management/importers/cct_field_adjustments.py +58 -0
  81. megadetector/data_management/importers/channel_islands_to_cct.py +913 -0
  82. megadetector/data_management/importers/eMammal/copy_and_unzip_emammal.py +180 -0
  83. megadetector/data_management/importers/eMammal/eMammal_helpers.py +249 -0
  84. megadetector/data_management/importers/eMammal/make_eMammal_json.py +223 -0
  85. megadetector/data_management/importers/ena24_to_json.py +276 -0
  86. megadetector/data_management/importers/filenames_to_json.py +386 -0
  87. megadetector/data_management/importers/helena_to_cct.py +283 -0
  88. megadetector/data_management/importers/idaho-camera-traps.py +1407 -0
  89. megadetector/data_management/importers/idfg_iwildcam_lila_prep.py +294 -0
  90. megadetector/data_management/importers/jb_csv_to_json.py +150 -0
  91. megadetector/data_management/importers/mcgill_to_json.py +250 -0
  92. megadetector/data_management/importers/missouri_to_json.py +490 -0
  93. megadetector/data_management/importers/nacti_fieldname_adjustments.py +79 -0
  94. megadetector/data_management/importers/noaa_seals_2019.py +181 -0
  95. megadetector/data_management/importers/pc_to_json.py +365 -0
  96. megadetector/data_management/importers/plot_wni_giraffes.py +123 -0
  97. megadetector/data_management/importers/prepare-noaa-fish-data-for-lila.py +359 -0
  98. megadetector/data_management/importers/prepare_zsl_imerit.py +131 -0
  99. megadetector/data_management/importers/rspb_to_json.py +356 -0
  100. megadetector/data_management/importers/save_the_elephants_survey_A.py +320 -0
  101. megadetector/data_management/importers/save_the_elephants_survey_B.py +329 -0
  102. megadetector/data_management/importers/snapshot_safari_importer.py +758 -0
  103. megadetector/data_management/importers/snapshot_safari_importer_reprise.py +665 -0
  104. megadetector/data_management/importers/snapshot_serengeti_lila.py +1067 -0
  105. megadetector/data_management/importers/snapshotserengeti/make_full_SS_json.py +150 -0
  106. megadetector/data_management/importers/snapshotserengeti/make_per_season_SS_json.py +153 -0
  107. megadetector/data_management/importers/sulross_get_exif.py +65 -0
  108. megadetector/data_management/importers/timelapse_csv_set_to_json.py +490 -0
  109. megadetector/data_management/importers/ubc_to_json.py +399 -0
  110. megadetector/data_management/importers/umn_to_json.py +507 -0
  111. megadetector/data_management/importers/wellington_to_json.py +263 -0
  112. megadetector/data_management/importers/wi_to_json.py +442 -0
  113. megadetector/data_management/importers/zamba_results_to_md_results.py +181 -0
  114. megadetector/data_management/labelme_to_coco.py +547 -0
  115. megadetector/data_management/labelme_to_yolo.py +272 -0
  116. megadetector/data_management/lila/__init__.py +0 -0
  117. megadetector/data_management/lila/add_locations_to_island_camera_traps.py +97 -0
  118. megadetector/data_management/lila/add_locations_to_nacti.py +147 -0
  119. megadetector/data_management/lila/create_lila_blank_set.py +558 -0
  120. megadetector/data_management/lila/create_lila_test_set.py +152 -0
  121. megadetector/data_management/lila/create_links_to_md_results_files.py +106 -0
  122. megadetector/data_management/lila/download_lila_subset.py +178 -0
  123. megadetector/data_management/lila/generate_lila_per_image_labels.py +516 -0
  124. megadetector/data_management/lila/get_lila_annotation_counts.py +170 -0
  125. megadetector/data_management/lila/get_lila_image_counts.py +112 -0
  126. megadetector/data_management/lila/lila_common.py +300 -0
  127. megadetector/data_management/lila/test_lila_metadata_urls.py +132 -0
  128. megadetector/data_management/ocr_tools.py +874 -0
  129. megadetector/data_management/read_exif.py +681 -0
  130. megadetector/data_management/remap_coco_categories.py +84 -0
  131. megadetector/data_management/remove_exif.py +66 -0
  132. megadetector/data_management/resize_coco_dataset.py +189 -0
  133. megadetector/data_management/wi_download_csv_to_coco.py +246 -0
  134. megadetector/data_management/yolo_output_to_md_output.py +441 -0
  135. megadetector/data_management/yolo_to_coco.py +676 -0
  136. megadetector/detection/__init__.py +0 -0
  137. megadetector/detection/detector_training/__init__.py +0 -0
  138. megadetector/detection/detector_training/model_main_tf2.py +114 -0
  139. megadetector/detection/process_video.py +702 -0
  140. megadetector/detection/pytorch_detector.py +341 -0
  141. megadetector/detection/run_detector.py +779 -0
  142. megadetector/detection/run_detector_batch.py +1219 -0
  143. megadetector/detection/run_inference_with_yolov5_val.py +917 -0
  144. megadetector/detection/run_tiled_inference.py +934 -0
  145. megadetector/detection/tf_detector.py +189 -0
  146. megadetector/detection/video_utils.py +606 -0
  147. megadetector/postprocessing/__init__.py +0 -0
  148. megadetector/postprocessing/add_max_conf.py +64 -0
  149. megadetector/postprocessing/categorize_detections_by_size.py +163 -0
  150. megadetector/postprocessing/combine_api_outputs.py +249 -0
  151. megadetector/postprocessing/compare_batch_results.py +958 -0
  152. megadetector/postprocessing/convert_output_format.py +396 -0
  153. megadetector/postprocessing/load_api_results.py +195 -0
  154. megadetector/postprocessing/md_to_coco.py +310 -0
  155. megadetector/postprocessing/md_to_labelme.py +330 -0
  156. megadetector/postprocessing/merge_detections.py +401 -0
  157. megadetector/postprocessing/postprocess_batch_results.py +1902 -0
  158. megadetector/postprocessing/remap_detection_categories.py +170 -0
  159. megadetector/postprocessing/render_detection_confusion_matrix.py +660 -0
  160. megadetector/postprocessing/repeat_detection_elimination/find_repeat_detections.py +211 -0
  161. megadetector/postprocessing/repeat_detection_elimination/remove_repeat_detections.py +83 -0
  162. megadetector/postprocessing/repeat_detection_elimination/repeat_detections_core.py +1631 -0
  163. megadetector/postprocessing/separate_detections_into_folders.py +730 -0
  164. megadetector/postprocessing/subset_json_detector_output.py +696 -0
  165. megadetector/postprocessing/top_folders_to_bottom.py +223 -0
  166. megadetector/taxonomy_mapping/__init__.py +0 -0
  167. megadetector/taxonomy_mapping/map_lila_taxonomy_to_wi_taxonomy.py +491 -0
  168. megadetector/taxonomy_mapping/map_new_lila_datasets.py +150 -0
  169. megadetector/taxonomy_mapping/prepare_lila_taxonomy_release.py +142 -0
  170. megadetector/taxonomy_mapping/preview_lila_taxonomy.py +590 -0
  171. megadetector/taxonomy_mapping/retrieve_sample_image.py +71 -0
  172. megadetector/taxonomy_mapping/simple_image_download.py +219 -0
  173. megadetector/taxonomy_mapping/species_lookup.py +834 -0
  174. megadetector/taxonomy_mapping/taxonomy_csv_checker.py +159 -0
  175. megadetector/taxonomy_mapping/taxonomy_graph.py +346 -0
  176. megadetector/taxonomy_mapping/validate_lila_category_mappings.py +83 -0
  177. megadetector/utils/__init__.py +0 -0
  178. megadetector/utils/azure_utils.py +178 -0
  179. megadetector/utils/ct_utils.py +612 -0
  180. megadetector/utils/directory_listing.py +246 -0
  181. megadetector/utils/md_tests.py +968 -0
  182. megadetector/utils/path_utils.py +1044 -0
  183. megadetector/utils/process_utils.py +157 -0
  184. megadetector/utils/sas_blob_utils.py +509 -0
  185. megadetector/utils/split_locations_into_train_val.py +228 -0
  186. megadetector/utils/string_utils.py +92 -0
  187. megadetector/utils/url_utils.py +323 -0
  188. megadetector/utils/write_html_image_list.py +225 -0
  189. megadetector/visualization/__init__.py +0 -0
  190. megadetector/visualization/plot_utils.py +293 -0
  191. megadetector/visualization/render_images_with_thumbnails.py +275 -0
  192. megadetector/visualization/visualization_utils.py +1536 -0
  193. megadetector/visualization/visualize_db.py +550 -0
  194. megadetector/visualization/visualize_detector_output.py +405 -0
  195. {megadetector-5.0.11.dist-info → megadetector-5.0.12.dist-info}/METADATA +1 -1
  196. megadetector-5.0.12.dist-info/RECORD +199 -0
  197. megadetector-5.0.12.dist-info/top_level.txt +1 -0
  198. megadetector-5.0.11.dist-info/RECORD +0 -5
  199. megadetector-5.0.11.dist-info/top_level.txt +0 -1
  200. {megadetector-5.0.11.dist-info → megadetector-5.0.12.dist-info}/LICENSE +0 -0
  201. {megadetector-5.0.11.dist-info → megadetector-5.0.12.dist-info}/WHEEL +0 -0
@@ -0,0 +1,399 @@
1
+ """
2
+
3
+ ubc_to_json.py
4
+
5
+ Convert the .csv file provided for the UBC data set to a
6
+ COCO-camera-traps .json file
7
+
8
+ Images were provided in eight folders, each of which contained a .csv
9
+ file with annotations. Those annotations came in two slightly different
10
+ formats, the two formats corresponding to folders starting with "SC_" and
11
+ otherwise.
12
+
13
+ """
14
+
15
+ #%% Constants and environment
16
+
17
+ import pandas as pd
18
+ import os
19
+ import json
20
+ import uuid
21
+ import numpy as np
22
+ import shutil
23
+
24
+ from tqdm import tqdm
25
+ from PIL import Image
26
+
27
+ from megadetector.visualization import visualize_db
28
+ from megadetector.data_management.databases import integrity_check_json_db
29
+ from megadetector.utils.path_utils import find_images
30
+
31
+ input_base = r'e:\ubc'
32
+ assert(os.path.isdir(input_base))
33
+
34
+ output_base = r'f:\data_staging\ubc'
35
+ output_json_file = os.path.join(output_base,'ubc.json')
36
+ file_list_file = os.path.join(output_base,'all_files.txt')
37
+
38
+ os.makedirs(output_base,exist_ok=True)
39
+
40
+ # Map Excel column names - which vary a little across spreadsheets - to a common set of names
41
+ mapped_fields = {"Survey.Name" : "survey_name",
42
+ "project_id": "survey_name",
43
+ "Camera.Name": "camera_name",
44
+ "station_id": "camera_name",
45
+ "Media.Filename": "filename",
46
+ "orig_file": "filename",
47
+ "timestamp_pst": "datetime",
48
+ "Date.Time": "datetime",
49
+ "Species": "species",
50
+ "latin_name": "species",
51
+ "common.name": "common_name",
52
+ "common_names": "common_name",
53
+ "Sighting.Quantity": "species_count"
54
+ }
55
+
56
+ category_mappings = {
57
+ 'bird_spp.':'unknown_bird',
58
+ 'dog_dog':'dog',
59
+ 'hiker_hiker':'hiker',
60
+ 'quad_quad':'quad',
61
+ 'skier_skier':'skier',
62
+ 'snowshoer_snowshoer':'showshoer',
63
+ 'quad_quad':'quad'
64
+ }
65
+
66
+ target_fields = ['species_count','group_count','behaviour']
67
+
68
+
69
+ #%% Enumerate images
70
+
71
+ # Load from file if we've already enumerated
72
+ if os.path.isfile(file_list_file):
73
+ with open(file_list_file,'r') as f:
74
+ files = f.readlines()
75
+ files = [s.strip() for s in files]
76
+ image_full_paths = files
77
+ print('Loaded {} images from {}'.format(len(image_full_paths),file_list_file))
78
+ else:
79
+ image_full_paths = find_images(input_base, recursive=True)
80
+ with open(file_list_file,'w') as f:
81
+ for line in image_full_paths:
82
+ f.write(line + '\n')
83
+ print('Enumerated {} images from {}'.format(len(image_full_paths),input_base))
84
+
85
+ image_full_paths_set = set(image_full_paths)
86
+ image_relative_paths = [os.path.relpath(fn,input_base) for fn in image_full_paths]
87
+ image_relative_paths_set = set(image_relative_paths)
88
+
89
+
90
+ #%% Create CCT dictionaries
91
+
92
+ annotations = []
93
+ image_ids_to_images = {}
94
+ category_name_to_category = {}
95
+
96
+ # Force the empty category to be ID 0
97
+ empty_category = {}
98
+ empty_category['name'] = 'empty'
99
+ empty_category['id'] = 0
100
+ empty_category['common_name'] = 'empty'
101
+ category_name_to_category['empty'] = empty_category
102
+ next_category_id = 1
103
+
104
+ latin_to_common = {}
105
+
106
+ folders = os.listdir(input_base)
107
+
108
+ # To simplify debugging of the loop below
109
+ i_folder = 0; folder = folders[i_folder];
110
+
111
+
112
+ ##%% Create CCT dictionaries (loop)
113
+
114
+ invalid_images = []
115
+
116
+ for i_folder,folder in enumerate(folders):
117
+
118
+ ##%%
119
+
120
+ print('\nProcessing folder {} of {}: {}'.format(i_folder,len(folders),folder))
121
+
122
+ filenames_to_rows = {}
123
+ filenames_with_multiple_annotations = []
124
+ missing_images = []
125
+ image_directory = os.path.join(input_base, folder)
126
+ files = os.listdir(image_directory)
127
+ files = list(filter(lambda f: f.endswith('.csv'), files))
128
+ input_metadata_file = os.path.join(input_base, folder, files[0])
129
+ assert(os.path.isfile(input_metadata_file))
130
+
131
+ # Read source data for this folder
132
+ input_metadata = pd.read_csv(input_metadata_file)
133
+
134
+ # Rename columns
135
+ input_metadata.rename(columns=mapped_fields, inplace=True)
136
+ print('Read {} columns and {} rows from metadata file'.format(len(input_metadata.columns), len(input_metadata)))
137
+
138
+ if folder.startswith("SC_"):
139
+ # Folder name is the first two characters of the filename
140
+ #
141
+ # Create relative path names from the filename itself
142
+ input_metadata['image_relative_path'] = input_metadata['filename'].apply(
143
+ lambda x: os.path.join(folder, x[0:2], x.replace(".JPG", ".jpg")))
144
+ else:
145
+ # Folder name is the camera name
146
+ #
147
+ # Create relative path names from camera name and filename
148
+ input_metadata['image_relative_path'] = input_metadata[['camera_name', 'filename']].apply(
149
+ lambda x: os.path.join(folder, x[0], x[1]), axis = 1)
150
+
151
+ # Which of our images are in the spreadsheet?
152
+ # i_row = 0; fn = input_metadata['image_relative_path'][i_row]
153
+ for i_row, image_relative_path in enumerate(input_metadata['image_relative_path']):
154
+
155
+ if image_relative_path in filenames_to_rows:
156
+ filenames_with_multiple_annotations.append(image_relative_path)
157
+ filenames_to_rows[image_relative_path].append(i_row)
158
+ else:
159
+ filenames_to_rows[image_relative_path] = [i_row]
160
+ image_full_path = os.path.join(input_base, image_relative_path)
161
+
162
+ if not image_full_path in image_full_paths_set:
163
+ assert image_relative_path not in image_relative_paths_set
164
+ missing_images.append(image_full_path)
165
+
166
+ print('Finished verifying image existence for {} files in {} rows\nFound {} filenames with multiple labels, {} missing images'.format(
167
+ len(input_metadata), len(filenames_to_rows),
168
+ len(filenames_with_multiple_annotations), len(missing_images)))
169
+
170
+
171
+ ##%% Check for images that aren't included in the metadata file
172
+
173
+ # Find all the images in this folder
174
+ image_relative_paths_this_folder = [s for s in image_relative_paths if s.startswith(folder)]
175
+
176
+ # Which of these aren't in the spreadsheet?
177
+ annotated_files_this_folder = list(filenames_to_rows.keys())
178
+ annotated_files_this_folder_set = set(annotated_files_this_folder)
179
+ unannotated_images = [s for s in image_relative_paths_this_folder if s not in annotated_files_this_folder_set]
180
+
181
+ print('Found {} unannotated images (of {}) in this folder'.format(
182
+ len(unannotated_images),len(image_relative_paths_this_folder)))
183
+
184
+
185
+ ##%% Create entries in CCT dictionaries
186
+
187
+ image_relative_path = list(filenames_to_rows.keys())[0]
188
+
189
+ for image_relative_path in list(filenames_to_rows.keys()):
190
+
191
+ # Only process images we have on disk
192
+ if image_relative_path not in image_relative_paths_set:
193
+ continue
194
+
195
+ image_full_path = os.path.join(input_base,image_relative_path)
196
+
197
+ # This is redundant, but doing this for clarity, at basically no performance
198
+ # cost since we need to *read* the images below to check validity.
199
+ assert os.path.isfile(image_full_path)
200
+
201
+ img_id = image_relative_path.replace('\\','/').replace('/','_').replace(' ','_')
202
+ row_indices = filenames_to_rows[image_relative_path]
203
+
204
+ # i_row = row_indices[0]
205
+ for i_row in row_indices:
206
+
207
+ row = input_metadata.iloc[i_row]
208
+ assert(row['image_relative_path'] == image_relative_path)
209
+ timestamp = row['datetime']
210
+ location = row['survey_name'] + '_' + row['camera_name']
211
+
212
+ if img_id in image_ids_to_images:
213
+ im = image_ids_to_images[img_id]
214
+ assert im['file_name'] == image_relative_path
215
+ assert im['location'] == location
216
+ else:
217
+ im = {}
218
+
219
+ try:
220
+ pil_image = Image.open(image_full_path)
221
+ width, height = pil_image.size
222
+ im['width'] = width
223
+ im['height'] = height
224
+ except:
225
+ # These generally represent zero-byte images in this data set, don't try
226
+ # to find the very small handful that might be other kinds of failures we
227
+ # might want to keep around.
228
+ # print('Error opening image {}'.format(image_relative_path))
229
+ invalid_images.append(image_relative_path)
230
+ continue
231
+
232
+ im['id'] = img_id
233
+ im['file_name'] = image_relative_path
234
+ im['datetime'] = timestamp
235
+ im['location'] = location
236
+
237
+ image_ids_to_images[img_id] = im
238
+
239
+ species = row['species'].lower().strip().replace(' ','_')
240
+
241
+ if (isinstance(species,float) or \
242
+ (isinstance(species,str) and (len(species) == 0))):
243
+ category_name = 'empty'
244
+ else:
245
+ category_name = species
246
+ del species
247
+
248
+ category_name = category_name.strip().lower()
249
+
250
+ common_name = row['common_name']
251
+ if isinstance(common_name,float) and np.isnan(common_name):
252
+ common_name = ''
253
+ else:
254
+ common_name = str(common_name).lower().strip().replace(', ',',').replace(' ','_')
255
+
256
+ for k,v in category_mappings.items():
257
+ common_name = common_name.replace(k,v)
258
+ category_name = category_name.replace(k,v)
259
+ common_name = common_name.replace('.','').replace('spp','species')
260
+ category_name = category_name.replace('.','').replace('spp','species')
261
+
262
+ if category_name == 'passerine_species' and common_name != '' and common_name != 'passerine_species':
263
+ category_name = common_name
264
+
265
+ # If we've seen this category before...
266
+ if category_name in category_name_to_category:
267
+
268
+ category = category_name_to_category[category_name]
269
+ category_id = category['id']
270
+
271
+ # ...make sure it used the same latin --> common mapping
272
+ #
273
+ # If the previous instance had no mapping, use the new one.
274
+ if category['common_name'] == '':
275
+ category['common_name'] = common_name
276
+ else:
277
+ # assert common_name == category['common_name']
278
+ if common_name != category['common_name']:
279
+ print('Warning: common name {} used for species {}, previously {}'.format(
280
+ common_name,category_name,category['common_name']))
281
+
282
+ else:
283
+
284
+ category_id = next_category_id
285
+ category = {}
286
+ category['id'] = category_id
287
+ category['name'] = category_name
288
+ category['common_name'] = common_name
289
+ category_name_to_category[category_name] = category
290
+ next_category_id += 1
291
+
292
+ # Create an annotation
293
+ ann = {}
294
+ ann['id'] = str(uuid.uuid1())
295
+ ann['image_id'] = im['id']
296
+ ann['category_id'] = category_id
297
+
298
+ for target_field in target_fields:
299
+ if target_field in input_metadata.columns:
300
+ val = row[target_field]
301
+ if isinstance(val,float) and np.isnan(val):
302
+ val = ''
303
+ else:
304
+ val = str(val).strip()
305
+ ann[target_field] = val
306
+
307
+ annotations.append(ann)
308
+
309
+ # ...for each annotation we found for this image
310
+
311
+ # ...for each image
312
+
313
+ # ...for each dataset
314
+
315
+ images = list(image_ids_to_images.values())
316
+ categories = list(category_name_to_category.values())
317
+
318
+ # Print all of our species mappings
319
+ for c in categories:
320
+ print(c['name'].ljust(30) + c['common_name'])
321
+
322
+ print('Finished creating CCT dictionaries, loaded {} images of {} total on disk ({} invalid)'.format(
323
+ len(images), len(image_relative_paths_set), len(invalid_images)))
324
+
325
+
326
+ #%% Copy images for which we actually have annotations to a new folder, lowercase everything
327
+
328
+ # im = images[0]
329
+ for im in tqdm(images):
330
+ relative_filename = im['file_name']
331
+ input_filename = os.path.join(input_base,relative_filename)
332
+ output_filename = os.path.join(output_base,relative_filename).lower()
333
+ os.makedirs(os.path.dirname(output_filename),exist_ok=True)
334
+
335
+ shutil.copy(input_filename, output_filename)
336
+ im['file_name'] = im['file_name'].lower()
337
+ im['id'] = im['id'].lower()
338
+
339
+
340
+ #%% Create info struct
341
+
342
+ info = {}
343
+ info['year'] = 2020
344
+ info['version'] = 1
345
+ info['description'] = 'UBC Camera Traps'
346
+ info['contributor'] = 'UBC'
347
+
348
+
349
+ #%% Convert image IDs to lowercase in annotations, tag as sequence level
350
+
351
+ # While there isn't any sequence information, the nature of false positives
352
+ # here leads me to believe the images were labeled at the sequence level, so
353
+ # we should trust labels more when positives are verified. Overall false
354
+ # positive rate looks to be between 1% and 5%.
355
+
356
+ for ann in annotations:
357
+ ann['image_id'] = ann['image_id'].lower()
358
+ ann['sequence_level_annotation'] = True
359
+
360
+
361
+ #%% Write output
362
+
363
+ json_data = {}
364
+ json_data['images'] = images
365
+ json_data['annotations'] = annotations
366
+ json_data['categories'] = categories
367
+ json_data['info'] = info
368
+ json.dump(json_data, open(output_json_file, 'w'), indent=2)
369
+
370
+ print('Finished writing .json file with {} images, {} annotations, and {} categories'.format(
371
+ len(images), len(annotations), len(categories)))
372
+
373
+
374
+ #%% Validate output
375
+
376
+ options = integrity_check_json_db.IntegrityCheckOptions()
377
+ options.baseDir = output_base
378
+ options.bCheckImageSizes = False
379
+ options.bCheckImageExistence = False
380
+ options.bFindUnusedImages = True
381
+
382
+ sortedCategories, data, errors = integrity_check_json_db.integrity_check_json_db(
383
+ output_json_file, options)
384
+
385
+
386
+ #%% Preview labels
387
+
388
+ viz_options = visualize_db.DbVizOptions()
389
+ viz_options.num_to_visualize = 2000
390
+ viz_options.trim_to_images_with_bboxes = False
391
+ viz_options.add_search_links = True
392
+ viz_options.sort_by_filename = False
393
+ viz_options.parallelize_rendering = True
394
+ html_output_file, image_db = visualize_db.visualize_db(db_path=output_json_file,
395
+ output_dir=os.path.join(
396
+ output_base, 'preview'),
397
+ image_base_dir=output_base,
398
+ options=viz_options)
399
+ os.startfile(html_output_file)