vismatch 1.1.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- vismatch/TEMPLATE.py +101 -0
- vismatch/__init__.py +475 -0
- vismatch/assets/example_pairs/false_positive/chartres.jpg +0 -0
- vismatch/assets/example_pairs/false_positive/notre_dame.jpg +0 -0
- vismatch/assets/example_pairs/fresco/fsm.jpg +0 -0
- vismatch/assets/example_pairs/fresco/sist_chapel.jpg +0 -0
- vismatch/assets/example_pairs/indoor/gcs_close.jpg +0 -0
- vismatch/assets/example_pairs/indoor/gcs_far.jpg +0 -0
- vismatch/assets/example_pairs/outdoor/montmartre_close.jpg +0 -0
- vismatch/assets/example_pairs/outdoor/montmartre_far.jpg +0 -0
- vismatch/assets/example_pairs/sat2iss/photo_from_iss.jpg +0 -0
- vismatch/assets/example_pairs/sat2iss/satellite_img.jpg +0 -0
- vismatch/assets/example_pairs/sphereglue/barbershop-00000000.jpg +0 -0
- vismatch/assets/example_pairs/sphereglue/barbershop-00000001.jpg +0 -0
- vismatch/assets/example_pairs/thermal/thermal.jpg +0 -0
- vismatch/assets/example_pairs/thermal/visible.jpg +0 -0
- vismatch/assets/example_test/original.jpg +0 -0
- vismatch/assets/example_test/warped.jpg +0 -0
- vismatch/base_matcher.py +242 -0
- vismatch/im_models/__init__.py +0 -0
- vismatch/im_models/aff_steerers.py +143 -0
- vismatch/im_models/aspanformer.py +74 -0
- vismatch/im_models/dedode.py +150 -0
- vismatch/im_models/duster.py +104 -0
- vismatch/im_models/edm.py +64 -0
- vismatch/im_models/efficient_loftr.py +60 -0
- vismatch/im_models/gim.py +187 -0
- vismatch/im_models/handcrafted.py +81 -0
- vismatch/im_models/keypt2subpx.py +154 -0
- vismatch/im_models/kornia.py +72 -0
- vismatch/im_models/liftfeat.py +44 -0
- vismatch/im_models/lightglue.py +75 -0
- vismatch/im_models/lisrd.py +98 -0
- vismatch/im_models/loftr.py +23 -0
- vismatch/im_models/master.py +107 -0
- vismatch/im_models/matchanything.py +221 -0
- vismatch/im_models/matchformer.py +61 -0
- vismatch/im_models/matching_toolbox.py +238 -0
- vismatch/im_models/minima.py +164 -0
- vismatch/im_models/omniglue.py +91 -0
- vismatch/im_models/rdd.py +250 -0
- vismatch/im_models/ripe.py +55 -0
- vismatch/im_models/roma.py +92 -0
- vismatch/im_models/romav2.py +62 -0
- vismatch/im_models/se2loftr.py +71 -0
- vismatch/im_models/silk.py +405 -0
- vismatch/im_models/sphereglue.py +97 -0
- vismatch/im_models/steerers.py +140 -0
- vismatch/im_models/topicfm.py +93 -0
- vismatch/im_models/ufm.py +57 -0
- vismatch/im_models/xfeat.py +78 -0
- vismatch/im_models/xfeat_steerers.py +151 -0
- vismatch/im_models/xoftr.py +71 -0
- vismatch/third_party/DeDoDe/DeDoDe/__init__.py +2 -0
- vismatch/third_party/DeDoDe/DeDoDe/benchmarks/__init__.py +4 -0
- vismatch/third_party/DeDoDe/DeDoDe/benchmarks/mega_pose_est.py +114 -0
- vismatch/third_party/DeDoDe/DeDoDe/benchmarks/mega_pose_est_mnn.py +119 -0
- vismatch/third_party/DeDoDe/DeDoDe/benchmarks/nll_benchmark.py +57 -0
- vismatch/third_party/DeDoDe/DeDoDe/benchmarks/num_inliers.py +76 -0
- vismatch/third_party/DeDoDe/DeDoDe/checkpoint.py +59 -0
- vismatch/third_party/DeDoDe/DeDoDe/datasets/__init__.py +0 -0
- vismatch/third_party/DeDoDe/DeDoDe/datasets/megadepth.py +269 -0
- vismatch/third_party/DeDoDe/DeDoDe/decoder.py +90 -0
- vismatch/third_party/DeDoDe/DeDoDe/descriptors/__init__.py +0 -0
- vismatch/third_party/DeDoDe/DeDoDe/descriptors/dedode_descriptor.py +50 -0
- vismatch/third_party/DeDoDe/DeDoDe/descriptors/descriptor_loss.py +68 -0
- vismatch/third_party/DeDoDe/DeDoDe/detectors/__init__.py +0 -0
- vismatch/third_party/DeDoDe/DeDoDe/detectors/dedode_detector.py +76 -0
- vismatch/third_party/DeDoDe/DeDoDe/detectors/keypoint_loss.py +185 -0
- vismatch/third_party/DeDoDe/DeDoDe/encoder.py +87 -0
- vismatch/third_party/DeDoDe/DeDoDe/matchers/__init__.py +0 -0
- vismatch/third_party/DeDoDe/DeDoDe/matchers/dual_softmax_matcher.py +38 -0
- vismatch/third_party/DeDoDe/DeDoDe/model_zoo/__init__.py +3 -0
- vismatch/third_party/DeDoDe/DeDoDe/model_zoo/dedode_models.py +249 -0
- vismatch/third_party/DeDoDe/DeDoDe/train.py +76 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/__init__.py +8 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/dinov2.py +359 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/__init__.py +12 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/attention.py +81 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/block.py +252 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/dino_head.py +59 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/drop_path.py +35 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/layer_scale.py +28 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/mlp.py +41 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/patch_embed.py +89 -0
- vismatch/third_party/DeDoDe/DeDoDe/transformer/layers/swiglu_ffn.py +63 -0
- vismatch/third_party/DeDoDe/DeDoDe/utils.py +717 -0
- vismatch/third_party/DeDoDe/data_prep/prep_keypoints.py +103 -0
- vismatch/third_party/DeDoDe/demo/demo_kpts.py +24 -0
- vismatch/third_party/DeDoDe/demo/demo_match.py +46 -0
- vismatch/third_party/DeDoDe/demo/demo_match_dedode_G.py +45 -0
- vismatch/third_party/DeDoDe/demo/demo_scoremap.py +23 -0
- vismatch/third_party/DeDoDe/experiments/dedode_descriptor-B.py +135 -0
- vismatch/third_party/DeDoDe/experiments/dedode_descriptor-G.py +145 -0
- vismatch/third_party/DeDoDe/experiments/dedode_detector.py +126 -0
- vismatch/third_party/DeDoDe/experiments/eval/eval_dedode_descriptor-B.py +38 -0
- vismatch/third_party/DeDoDe/experiments/eval/eval_dedode_descriptor-G.py +38 -0
- vismatch/third_party/DeDoDe/setup.py +11 -0
- vismatch/third_party/EDM/configs/data/__init__.py +0 -0
- vismatch/third_party/EDM/configs/data/base.py +37 -0
- vismatch/third_party/EDM/configs/data/megadepth_test_1500.py +23 -0
- vismatch/third_party/EDM/configs/data/megadepth_trainval_832.py +32 -0
- vismatch/third_party/EDM/configs/data/scannet_test_1500.py +24 -0
- vismatch/third_party/EDM/configs/data/scannet_trainval.py +31 -0
- vismatch/third_party/EDM/configs/edm/indoor/edm_base.py +15 -0
- vismatch/third_party/EDM/configs/edm/outdoor/edm_base.py +17 -0
- vismatch/third_party/EDM/deploy/export_onnx.py +69 -0
- vismatch/third_party/EDM/deploy/run_onnx.py +138 -0
- vismatch/third_party/EDM/runtime_single_pair.py +73 -0
- vismatch/third_party/EDM/src/__init__.py +0 -0
- vismatch/third_party/EDM/src/config/default.py +184 -0
- vismatch/third_party/EDM/src/datasets/megadepth.py +164 -0
- vismatch/third_party/EDM/src/datasets/sampler.py +95 -0
- vismatch/third_party/EDM/src/datasets/scannet.py +147 -0
- vismatch/third_party/EDM/src/edm/__init__.py +2 -0
- vismatch/third_party/EDM/src/edm/backbone/resnet.py +116 -0
- vismatch/third_party/EDM/src/edm/edm.py +204 -0
- vismatch/third_party/EDM/src/edm/head/coarse_matching.py +158 -0
- vismatch/third_party/EDM/src/edm/head/fine_matching.py +383 -0
- vismatch/third_party/EDM/src/edm/neck/__init__.py +1 -0
- vismatch/third_party/EDM/src/edm/neck/loftr_module/__init__.py +1 -0
- vismatch/third_party/EDM/src/edm/neck/loftr_module/transformer.py +418 -0
- vismatch/third_party/EDM/src/edm/neck/neck.py +156 -0
- vismatch/third_party/EDM/src/edm/utils/geometry.py +58 -0
- vismatch/third_party/EDM/src/edm/utils/supervision.py +255 -0
- vismatch/third_party/EDM/src/lightning/data.py +450 -0
- vismatch/third_party/EDM/src/lightning/lightning_edm.py +379 -0
- vismatch/third_party/EDM/src/losses/edm_loss.py +206 -0
- vismatch/third_party/EDM/src/optimizers/__init__.py +57 -0
- vismatch/third_party/EDM/src/utils/augment.py +65 -0
- vismatch/third_party/EDM/src/utils/comm.py +271 -0
- vismatch/third_party/EDM/src/utils/dataloader.py +24 -0
- vismatch/third_party/EDM/src/utils/dataset.py +192 -0
- vismatch/third_party/EDM/src/utils/metrics.py +299 -0
- vismatch/third_party/EDM/src/utils/misc.py +113 -0
- vismatch/third_party/EDM/src/utils/plotting.py +186 -0
- vismatch/third_party/EDM/src/utils/profiler.py +40 -0
- vismatch/third_party/EDM/src/utils/warppers.py +428 -0
- vismatch/third_party/EDM/src/utils/warppers_utils.py +172 -0
- vismatch/third_party/EDM/test.py +132 -0
- vismatch/third_party/EDM/train.py +156 -0
- vismatch/third_party/EfficientLoFTR/configs/data/__init__.py +0 -0
- vismatch/third_party/EfficientLoFTR/configs/data/base.py +35 -0
- vismatch/third_party/EfficientLoFTR/configs/data/megadepth_test_1500.py +13 -0
- vismatch/third_party/EfficientLoFTR/configs/data/megadepth_trainval_832.py +24 -0
- vismatch/third_party/EfficientLoFTR/configs/data/scannet_test_1500.py +16 -0
- vismatch/third_party/EfficientLoFTR/configs/loftr/eloftr_full.py +36 -0
- vismatch/third_party/EfficientLoFTR/configs/loftr/eloftr_optimized.py +37 -0
- vismatch/third_party/EfficientLoFTR/src/__init__.py +0 -0
- vismatch/third_party/EfficientLoFTR/src/config/default.py +182 -0
- vismatch/third_party/EfficientLoFTR/src/datasets/megadepth.py +133 -0
- vismatch/third_party/EfficientLoFTR/src/datasets/sampler.py +77 -0
- vismatch/third_party/EfficientLoFTR/src/datasets/scannet.py +129 -0
- vismatch/third_party/EfficientLoFTR/src/lightning/data.py +357 -0
- vismatch/third_party/EfficientLoFTR/src/lightning/lightning_loftr.py +272 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/__init__.py +4 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/backbone/__init__.py +11 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/backbone/backbone.py +37 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/backbone/repvgg.py +224 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/loftr.py +124 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/loftr_module/__init__.py +2 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/loftr_module/fine_preprocess.py +112 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/loftr_module/linear_attention.py +103 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/loftr_module/transformer.py +164 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/utils/coarse_matching.py +241 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/utils/fine_matching.py +156 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/utils/full_config.py +50 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/utils/geometry.py +54 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/utils/opt_config.py +50 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/utils/position_encoding.py +50 -0
- vismatch/third_party/EfficientLoFTR/src/loftr/utils/supervision.py +275 -0
- vismatch/third_party/EfficientLoFTR/src/losses/loftr_loss.py +229 -0
- vismatch/third_party/EfficientLoFTR/src/optimizers/__init__.py +42 -0
- vismatch/third_party/EfficientLoFTR/src/utils/augment.py +55 -0
- vismatch/third_party/EfficientLoFTR/src/utils/comm.py +265 -0
- vismatch/third_party/EfficientLoFTR/src/utils/dataloader.py +23 -0
- vismatch/third_party/EfficientLoFTR/src/utils/dataset.py +186 -0
- vismatch/third_party/EfficientLoFTR/src/utils/metrics.py +264 -0
- vismatch/third_party/EfficientLoFTR/src/utils/misc.py +106 -0
- vismatch/third_party/EfficientLoFTR/src/utils/plotting.py +154 -0
- vismatch/third_party/EfficientLoFTR/src/utils/profiler.py +39 -0
- vismatch/third_party/EfficientLoFTR/src/utils/warppers.py +426 -0
- vismatch/third_party/EfficientLoFTR/src/utils/warppers_utils.py +171 -0
- vismatch/third_party/EfficientLoFTR/test.py +143 -0
- vismatch/third_party/EfficientLoFTR/train.py +154 -0
- vismatch/third_party/LISRD/lisrd/__init__.py +0 -0
- vismatch/third_party/LISRD/lisrd/datasets/__init__.py +7 -0
- vismatch/third_party/LISRD/lisrd/datasets/base_dataset.py +38 -0
- vismatch/third_party/LISRD/lisrd/datasets/coco.py +148 -0
- vismatch/third_party/LISRD/lisrd/datasets/flashes.py +170 -0
- vismatch/third_party/LISRD/lisrd/datasets/hpatches.py +135 -0
- vismatch/third_party/LISRD/lisrd/datasets/mixed_dataset.py +53 -0
- vismatch/third_party/LISRD/lisrd/datasets/rdnim.py +117 -0
- vismatch/third_party/LISRD/lisrd/datasets/utils/data_augmentation.py +168 -0
- vismatch/third_party/LISRD/lisrd/datasets/utils/data_reader.py +48 -0
- vismatch/third_party/LISRD/lisrd/datasets/utils/homographies.py +215 -0
- vismatch/third_party/LISRD/lisrd/datasets/vidit.py +152 -0
- vismatch/third_party/LISRD/lisrd/evaluation/__init__.py +0 -0
- vismatch/third_party/LISRD/lisrd/evaluation/descriptor_evaluation.py +142 -0
- vismatch/third_party/LISRD/lisrd/experiment.py +129 -0
- vismatch/third_party/LISRD/lisrd/export_features.py +148 -0
- vismatch/third_party/LISRD/lisrd/models/__init__.py +7 -0
- vismatch/third_party/LISRD/lisrd/models/backbones/__init__.py +0 -0
- vismatch/third_party/LISRD/lisrd/models/backbones/net_vlad.py +62 -0
- vismatch/third_party/LISRD/lisrd/models/backbones/vgg.py +46 -0
- vismatch/third_party/LISRD/lisrd/models/base_model.py +336 -0
- vismatch/third_party/LISRD/lisrd/models/keypoint_detectors.py +34 -0
- vismatch/third_party/LISRD/lisrd/models/lisrd.py +328 -0
- vismatch/third_party/LISRD/lisrd/models/lisrd_sift.py +289 -0
- vismatch/third_party/LISRD/lisrd/third_party/super_point_magic_leap/demo_superpoint.py +734 -0
- vismatch/third_party/LISRD/lisrd/utils/geometry_utils.py +123 -0
- vismatch/third_party/LISRD/lisrd/utils/losses.py +191 -0
- vismatch/third_party/LISRD/lisrd/utils/metrics.py +66 -0
- vismatch/third_party/LISRD/lisrd/utils/pytorch_utils.py +14 -0
- vismatch/third_party/LISRD/lisrd/utils/stdout_capturing.py +81 -0
- vismatch/third_party/LISRD/notebooks/utils.py +103 -0
- vismatch/third_party/LISRD/setup.py +4 -0
- vismatch/third_party/LiftFeat/dataset/__init__.py +0 -0
- vismatch/third_party/LiftFeat/dataset/coco_augmentor.py +298 -0
- vismatch/third_party/LiftFeat/dataset/coco_wrapper.py +175 -0
- vismatch/third_party/LiftFeat/dataset/dataset_utils.py +183 -0
- vismatch/third_party/LiftFeat/dataset/megadepth.py +177 -0
- vismatch/third_party/LiftFeat/dataset/megadepth_wrapper.py +167 -0
- vismatch/third_party/LiftFeat/demo.py +116 -0
- vismatch/third_party/LiftFeat/evaluation/HPatch_evaluation.py +182 -0
- vismatch/third_party/LiftFeat/evaluation/MegaDepth1500_evaluation.py +105 -0
- vismatch/third_party/LiftFeat/evaluation/eval_utils.py +127 -0
- vismatch/third_party/LiftFeat/loss/loss.py +291 -0
- vismatch/third_party/LiftFeat/models/interpolator.py +34 -0
- vismatch/third_party/LiftFeat/models/liftfeat_wrapper.py +172 -0
- vismatch/third_party/LiftFeat/models/model.py +419 -0
- vismatch/third_party/LiftFeat/tools/demo_match_video.py +145 -0
- vismatch/third_party/LiftFeat/tools/demo_vo.py +163 -0
- vismatch/third_party/LiftFeat/train.py +369 -0
- vismatch/third_party/LiftFeat/utils/VisualOdometry.py +339 -0
- vismatch/third_party/LiftFeat/utils/__init__.py +0 -0
- vismatch/third_party/LiftFeat/utils/alike_wrapper.py +45 -0
- vismatch/third_party/LiftFeat/utils/config.py +16 -0
- vismatch/third_party/LiftFeat/utils/depth_anything_wrapper.py +150 -0
- vismatch/third_party/LiftFeat/utils/featurebooster.py +247 -0
- vismatch/third_party/LiftFeat/utils/post_process.py +21 -0
- vismatch/third_party/LightGlue/benchmark.py +255 -0
- vismatch/third_party/LightGlue/lightglue/__init__.py +7 -0
- vismatch/third_party/LightGlue/lightglue/aliked.py +760 -0
- vismatch/third_party/LightGlue/lightglue/disk.py +55 -0
- vismatch/third_party/LightGlue/lightglue/dog_hardnet.py +41 -0
- vismatch/third_party/LightGlue/lightglue/lightglue.py +662 -0
- vismatch/third_party/LightGlue/lightglue/sift.py +216 -0
- vismatch/third_party/LightGlue/lightglue/superpoint.py +227 -0
- vismatch/third_party/LightGlue/lightglue/utils.py +165 -0
- vismatch/third_party/LightGlue/lightglue/viz2d.py +203 -0
- vismatch/third_party/MINIMA/demo.py +201 -0
- vismatch/third_party/MINIMA/src/__init__.py +0 -0
- vismatch/third_party/MINIMA/src/config/default.py +203 -0
- vismatch/third_party/MINIMA/src/config/default_for_megadepth_dense.py +203 -0
- vismatch/third_party/MINIMA/src/config/default_for_megadepth_sparse.py +203 -0
- vismatch/third_party/MINIMA/src/utils/__init__.py +0 -0
- vismatch/third_party/MINIMA/src/utils/culculate_auc.py +28 -0
- vismatch/third_party/MINIMA/src/utils/data_io.py +156 -0
- vismatch/third_party/MINIMA/src/utils/data_io_loftr.py +152 -0
- vismatch/third_party/MINIMA/src/utils/data_io_roma.py +186 -0
- vismatch/third_party/MINIMA/src/utils/data_io_sp_lg.py +158 -0
- vismatch/third_party/MINIMA/src/utils/load_model.py +164 -0
- vismatch/third_party/MINIMA/src/utils/metrics.py +214 -0
- vismatch/third_party/MINIMA/src/utils/misc.py +101 -0
- vismatch/third_party/MINIMA/src/utils/plotting.py +291 -0
- vismatch/third_party/MINIMA/src/utils/sample_h.py +142 -0
- vismatch/third_party/MINIMA/test_relative_homo_depth.py +683 -0
- vismatch/third_party/MINIMA/test_relative_homo_event.py +722 -0
- vismatch/third_party/MINIMA/test_relative_homo_mmim.py +669 -0
- vismatch/third_party/MINIMA/test_relative_pose_infrared.py +500 -0
- vismatch/third_party/MINIMA/test_relative_pose_mega_1500.py +487 -0
- vismatch/third_party/MINIMA/test_relative_pose_mega_1500_syn.py +516 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/benchmark.py +255 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/__init__.py +7 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/aliked.py +758 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/disk.py +55 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/dog_hardnet.py +41 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/lightglue.py +655 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/sift.py +216 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/superpoint.py +227 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/utils.py +165 -0
- vismatch/third_party/MINIMA/third_party/LightGlue/lightglue/viz2d.py +184 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/data/__init__.py +0 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/data/base.py +35 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/data/megadepth_test_1500.py +11 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/data/megadepth_trainval_640.py +22 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/data/megadepth_trainval_840.py +22 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/data/scannet_test_1500.py +11 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/data/scannet_trainval.py +17 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ds.py +6 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ds_dense.py +8 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ot.py +6 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ot_dense.py +8 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/loftr_ds.py +5 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/loftr_ds_dense.py +7 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/loftr_ot.py +5 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/loftr_ot_dense.py +7 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/scannet/loftr_ds_eval.py +16 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/indoor/scannet/loftr_ds_eval_new.py +18 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ds.py +16 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ds_dense.py +17 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ot.py +16 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ot_dense.py +17 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/loftr_ds.py +15 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/loftr_ds_dense.py +16 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/loftr_ot.py +15 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/configs/loftr/outdoor/loftr_ot_dense.py +16 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/demo/demo_loftr.py +240 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/__init__.py +0 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/config/default.py +171 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/datasets/megadepth.py +127 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/datasets/sampler.py +77 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/datasets/scannet.py +114 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/lightning/data.py +320 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/lightning/lightning_loftr.py +249 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/__init__.py +2 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/backbone/__init__.py +11 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/backbone/resnet_fpn.py +199 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/loftr.py +81 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/loftr_module/__init__.py +2 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/loftr_module/fine_preprocess.py +59 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/loftr_module/linear_attention.py +81 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/loftr_module/transformer.py +101 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/utils/coarse_matching.py +261 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/utils/cvpr_ds_config.py +50 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/utils/fine_matching.py +74 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/utils/geometry.py +54 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/utils/position_encoding.py +42 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/loftr/utils/supervision.py +151 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/losses/loftr_loss.py +192 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/optimizers/__init__.py +42 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/augment.py +55 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/comm.py +265 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/dataloader.py +23 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/dataset.py +185 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/metrics.py +193 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/misc.py +101 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/plotting.py +154 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/src/utils/profiler.py +39 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/test.py +68 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/third_party/SuperGluePretrainedNetwork/demo_superglue.py +259 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/third_party/SuperGluePretrainedNetwork/match_pairs.py +425 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/third_party/SuperGluePretrainedNetwork/models/__init__.py +0 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/third_party/SuperGluePretrainedNetwork/models/matching.py +84 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/third_party/SuperGluePretrainedNetwork/models/superglue.py +283 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/third_party/SuperGluePretrainedNetwork/models/superpoint.py +202 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/third_party/SuperGluePretrainedNetwork/models/utils.py +555 -0
- vismatch/third_party/MINIMA/third_party/LoFTR/train.py +123 -0
- vismatch/third_party/MINIMA/third_party/RoMa/demo/demo_3D_effect.py +47 -0
- vismatch/third_party/MINIMA/third_party/RoMa/demo/demo_fundamental.py +34 -0
- vismatch/third_party/MINIMA/third_party/RoMa/demo/demo_match.py +50 -0
- vismatch/third_party/MINIMA/third_party/RoMa/demo/demo_match_opencv_sift.py +43 -0
- vismatch/third_party/MINIMA/third_party/RoMa/demo/demo_match_tiny.py +77 -0
- vismatch/third_party/MINIMA/third_party/RoMa/experiments/eval_roma_outdoor.py +57 -0
- vismatch/third_party/MINIMA/third_party/RoMa/experiments/eval_tiny_roma_v1_outdoor.py +84 -0
- vismatch/third_party/MINIMA/third_party/RoMa/experiments/roma_indoor.py +320 -0
- vismatch/third_party/MINIMA/third_party/RoMa/experiments/train_roma_outdoor.py +307 -0
- vismatch/third_party/MINIMA/third_party/RoMa/experiments/train_tiny_roma_v1_outdoor.py +498 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/__init__.py +8 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/benchmarks/__init__.py +6 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/benchmarks/hpatches_sequences_homog_benchmark.py +113 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/benchmarks/megadepth_dense_benchmark.py +106 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/benchmarks/megadepth_pose_estimation_benchmark.py +118 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/benchmarks/megadepth_pose_estimation_benchmark_poselib.py +119 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/benchmarks/scannet_benchmark.py +143 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/checkpointing/__init__.py +1 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/checkpointing/checkpoint.py +60 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/datasets/__init__.py +2 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/datasets/megadepth.py +232 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/datasets/scannet.py +160 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/losses/__init__.py +1 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/losses/robust_loss.py +161 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/losses/robust_loss_tiny_roma.py +160 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/__init__.py +1 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/encoders.py +122 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/matcher.py +766 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/model_zoo/__init__.py +73 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/model_zoo/roma_models.py +170 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/tiny.py +304 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/__init__.py +48 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/dinov2.py +359 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/__init__.py +12 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/attention.py +81 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/block.py +252 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/dino_head.py +59 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/drop_path.py +35 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/layer_scale.py +28 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/mlp.py +41 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/patch_embed.py +89 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/models/transformer/layers/swiglu_ffn.py +63 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/train/__init__.py +1 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/train/train.py +102 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/utils/__init__.py +16 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/utils/kde.py +13 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/utils/local_correlation.py +48 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/utils/transforms.py +118 -0
- vismatch/third_party/MINIMA/third_party/RoMa/romatch/utils/utils.py +662 -0
- vismatch/third_party/MINIMA/third_party/RoMa/setup.py +9 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/configs/data/__init__.py +0 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/configs/data/base.py +35 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/configs/data/megadepth_trainval_840.py +22 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/configs/data/megadepth_vistir_trainval_640.py +23 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/configs/data/pretrain.py +8 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/configs/xoftr/outdoor/visible_thermal.py +17 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/configs/xoftr/pretrain/pretrain.py +12 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/pretrain.py +125 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/__init__.py +0 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/config/default.py +203 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/datasets/megadepth.py +143 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/datasets/pretrain_dataset.py +156 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/datasets/sampler.py +77 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/datasets/scannet.py +114 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/datasets/vistir.py +109 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/lightning/data.py +346 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/lightning/data_pretrain.py +125 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/lightning/lightning_xoftr.py +334 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/lightning/lightning_xoftr_pretrain.py +171 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/losses/xoftr_loss.py +170 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/losses/xoftr_loss_pretrain.py +37 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/optimizers/__init__.py +42 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/augment.py +113 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/comm.py +265 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/data_io.py +144 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/dataloader.py +23 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/dataset.py +279 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/metrics.py +211 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/misc.py +101 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/plotting.py +227 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/pretrain_utils.py +83 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/utils/profiler.py +39 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/__init__.py +2 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/backbone/__init__.py +1 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/backbone/resnet.py +95 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/utils/geometry.py +107 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/utils/position_encoding.py +36 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/utils/supervision.py +290 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr.py +94 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr_module/__init__.py +4 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr_module/coarse_matching.py +305 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr_module/fine_matching.py +170 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr_module/fine_process.py +321 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr_module/linear_attention.py +81 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr_module/transformer.py +101 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/src/xoftr/xoftr_pretrain.py +209 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/test.py +68 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/test_relative_pose.py +330 -0
- vismatch/third_party/MINIMA/third_party/XoFTR/train.py +126 -0
- vismatch/third_party/MatchAnything/app.py +27 -0
- vismatch/third_party/MatchAnything/imcui/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/api/__init__.py +47 -0
- vismatch/third_party/MatchAnything/imcui/api/client.py +232 -0
- vismatch/third_party/MatchAnything/imcui/api/core.py +308 -0
- vismatch/third_party/MatchAnything/imcui/api/server.py +170 -0
- vismatch/third_party/MatchAnything/imcui/hloc/__init__.py +65 -0
- vismatch/third_party/MatchAnything/imcui/hloc/colmap_from_nvm.py +216 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extract_features.py +607 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/alike.py +61 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/aliked.py +32 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/cosplace.py +44 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/d2net.py +60 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/darkfeat.py +44 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/dedode.py +86 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/dir.py +78 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/disk.py +35 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/dog.py +135 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/eigenplaces.py +57 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/example.py +56 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/fire.py +72 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/fire_local.py +84 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/lanet.py +63 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/netvlad.py +146 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/openibl.py +26 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/r2d2.py +73 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/rekd.py +60 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/rord.py +59 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/sfd2.py +44 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/sift.py +216 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/superpoint.py +51 -0
- vismatch/third_party/MatchAnything/imcui/hloc/extractors/xfeat.py +33 -0
- vismatch/third_party/MatchAnything/imcui/hloc/localize_inloc.py +179 -0
- vismatch/third_party/MatchAnything/imcui/hloc/localize_sfm.py +243 -0
- vismatch/third_party/MatchAnything/imcui/hloc/match_dense.py +1158 -0
- vismatch/third_party/MatchAnything/imcui/hloc/match_features.py +459 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/__init__.py +3 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/adalam.py +68 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/aspanformer.py +66 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/cotr.py +77 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/dkm.py +53 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/dual_softmax.py +71 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/duster.py +109 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/eloftr.py +97 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/gim.py +200 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/gluestick.py +99 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/imp.py +50 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/lightglue.py +67 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/loftr.py +58 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/mast3r.py +96 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/matchanything.py +191 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/mickey.py +50 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/nearest_neighbor.py +66 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/omniglue.py +80 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/roma.py +80 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/sgmnet.py +106 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/sold2.py +144 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/superglue.py +33 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/topicfm.py +60 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/xfeat_dense.py +54 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/xfeat_lightglue.py +48 -0
- vismatch/third_party/MatchAnything/imcui/hloc/matchers/xoftr.py +90 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pairs_from_covisibility.py +60 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pairs_from_exhaustive.py +64 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pairs_from_poses.py +68 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pairs_from_retrieval.py +133 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/4Seasons/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/4Seasons/localize.py +89 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/4Seasons/prepare_reference.py +51 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/4Seasons/utils.py +231 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/7Scenes/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/7Scenes/create_gt_sfm.py +134 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/7Scenes/pipeline.py +139 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/7Scenes/utils.py +34 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Aachen/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Aachen/pipeline.py +109 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Aachen_v1_1/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Aachen_v1_1/pipeline.py +104 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Aachen_v1_1/pipeline_loftr.py +104 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/CMU/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/CMU/pipeline.py +133 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Cambridge/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Cambridge/pipeline.py +140 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/Cambridge/utils.py +145 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/RobotCar/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/RobotCar/colmap_from_nvm.py +176 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/RobotCar/pipeline.py +143 -0
- vismatch/third_party/MatchAnything/imcui/hloc/pipelines/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/hloc/reconstruction.py +194 -0
- vismatch/third_party/MatchAnything/imcui/hloc/triangulation.py +311 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/__init__.py +12 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/base_model.py +56 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/database.py +412 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/geometry.py +16 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/io.py +77 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/parsers.py +59 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/read_write_model.py +588 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/viz.py +146 -0
- vismatch/third_party/MatchAnything/imcui/hloc/utils/viz_3d.py +203 -0
- vismatch/third_party/MatchAnything/imcui/hloc/visualization.py +178 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/configs/models/eloftr_model.py +128 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/configs/models/roma_model.py +27 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/notebooks/notebooks_utils/__init__.py +1 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/notebooks/notebooks_utils/plotting.py +344 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/config/default.py +344 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/datasets/common_data_pair.py +214 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/lightning/lightning_loftr.py +343 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/__init__.py +1 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/backbone/__init__.py +61 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/backbone/repvgg.py +319 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/backbone/resnet_fpn.py +1094 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/backbone/s2dnet.py +131 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/loftr.py +273 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/loftr_module/__init__.py +2 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/loftr_module/fine_preprocess.py +350 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/loftr_module/linear_attention.py +217 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/loftr_module/transformer.py +1768 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/loftr_module/transformer_utils.py +76 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/utils/coarse_matching.py +266 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/utils/fine_matching.py +493 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/utils/geometry.py +298 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/utils/position_encoding.py +131 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/loftr/utils/supervision.py +475 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/optimizers/__init__.py +50 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/augment.py +55 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/colmap/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/colmap/database.py +417 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/colmap/eval_helper.py +232 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/colmap/read_write_model.py +509 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/colmap.py +530 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/comm.py +265 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/dataloader.py +23 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/dataset.py +518 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/easydict.py +148 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/geometry.py +366 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/homography_utils.py +366 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/metrics.py +445 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/misc.py +101 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/plotting.py +248 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/profiler.py +39 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/ray_utils.py +134 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/sample_homo.py +58 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/src/utils/utils.py +600 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/demo/demo_3D_effect.py +46 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/demo/demo_fundamental.py +32 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/demo/demo_fundamental_model_warpper.py +34 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/demo/demo_match.py +50 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/demo/demo_match_opencv_sift.py +43 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/demo_single_pair.py +329 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/experiments/roma_indoor.py +320 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/experiments/roma_outdoor.py +327 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/notebooks/notebooks_utils/__init__.py +1 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/notebooks/notebooks_utils/plotting.py +331 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/__init__.py +8 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/benchmarks/__init__.py +4 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/benchmarks/hpatches_sequences_homog_benchmark.py +113 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/benchmarks/megadepth_dense_benchmark.py +106 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/benchmarks/megadepth_pose_estimation_benchmark.py +140 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/benchmarks/scannet_benchmark.py +143 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/checkpointing/__init__.py +1 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/checkpointing/checkpoint.py +60 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/datasets/__init__.py +2 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/datasets/megadepth.py +230 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/datasets/scannet.py +160 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/losses/__init__.py +1 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/losses/robust_loss.py +157 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/matchanything_roma_model.py +104 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/__init__.py +1 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/blocks.py +241 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/criterion.py +37 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/croco.py +253 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/croco_downstream.py +122 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/curope/__init__.py +4 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/curope/curope2d.py +40 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/curope/setup.py +34 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/dpt_block.py +450 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/head_downstream.py +58 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/masking.py +25 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/croco/pos_embed.py +159 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/__init__.py +2 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/cloud_opt/__init__.py +29 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/cloud_opt/base_opt.py +375 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/cloud_opt/commons.py +90 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/cloud_opt/init_im_poses.py +312 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/cloud_opt/optimizer.py +230 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/cloud_opt/pair_viewer.py +125 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/__init__.py +42 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/base/__init__.py +2 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/base/base_stereo_view_dataset.py +220 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/base/batched_sampler.py +74 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/base/easy_dataset.py +157 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/co3d.py +146 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/utils/__init__.py +2 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/utils/cropping.py +119 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/datasets/utils/transforms.py +11 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/heads/__init__.py +19 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/heads/dpt_head.py +114 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/heads/linear_head.py +41 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/heads/postprocess.py +58 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/image_pairs.py +83 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/inference.py +165 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/losses.py +297 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/model.py +167 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/optim_factory.py +14 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/patch_embed.py +70 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/post_process.py +60 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/utils/__init__.py +2 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/utils/device.py +76 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/utils/geometry.py +361 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/utils/image.py +104 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/utils/misc.py +121 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/utils/path_to_croco.py +19 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/dust3r/viz.py +320 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/encoders.py +137 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/matcher.py +937 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/model_zoo/__init__.py +53 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/model_zoo/roma_models.py +162 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/__init__.py +47 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/dinov2.py +359 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/__init__.py +12 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/attention.py +81 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/block.py +252 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/dino_head.py +59 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/drop_path.py +35 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/layer_scale.py +28 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/mlp.py +41 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/patch_embed.py +89 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/models/transformer/layers/swiglu_ffn.py +63 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/roma_adpat_model.py +32 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/train/__init__.py +1 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/train/train.py +102 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/utils/__init__.py +18 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/utils/kde.py +8 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/utils/local_correlation.py +44 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/utils/transforms.py +118 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/roma/utils/utils.py +661 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/third_party/ROMA/setup.py +9 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/tools/__init__.py +0 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/tools/evaluate_datasets.py +239 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/tools/tools_utils/data_io.py +94 -0
- vismatch/third_party/MatchAnything/imcui/third_party/MatchAnything/tools/tools_utils/plot.py +77 -0
- vismatch/third_party/MatchAnything/imcui/ui/__init__.py +5 -0
- vismatch/third_party/MatchAnything/imcui/ui/app_class.py +824 -0
- vismatch/third_party/MatchAnything/imcui/ui/sfm.py +164 -0
- vismatch/third_party/MatchAnything/imcui/ui/utils.py +1085 -0
- vismatch/third_party/MatchAnything/imcui/ui/viz.py +511 -0
- vismatch/third_party/MatchAnything/tests/test_basic.py +111 -0
- vismatch/third_party/MatchFormer/config/data/__init__.py +0 -0
- vismatch/third_party/MatchFormer/config/data/base.py +35 -0
- vismatch/third_party/MatchFormer/config/data/megadepth_test_1500.py +11 -0
- vismatch/third_party/MatchFormer/config/data/scannet_test_1500.py +11 -0
- vismatch/third_party/MatchFormer/config/defaultmf.py +88 -0
- vismatch/third_party/MatchFormer/model/backbone/__init__.py +17 -0
- vismatch/third_party/MatchFormer/model/backbone/coarse_matching.py +228 -0
- vismatch/third_party/MatchFormer/model/backbone/fine_matching.py +74 -0
- vismatch/third_party/MatchFormer/model/backbone/fine_preprocess.py +59 -0
- vismatch/third_party/MatchFormer/model/backbone/match_LA_large.py +254 -0
- vismatch/third_party/MatchFormer/model/backbone/match_LA_lite.py +254 -0
- vismatch/third_party/MatchFormer/model/backbone/match_SEA_large.py +291 -0
- vismatch/third_party/MatchFormer/model/backbone/match_SEA_lite.py +291 -0
- vismatch/third_party/MatchFormer/model/data.py +320 -0
- vismatch/third_party/MatchFormer/model/datasets/dataset.py +231 -0
- vismatch/third_party/MatchFormer/model/datasets/megadepth.py +126 -0
- vismatch/third_party/MatchFormer/model/datasets/sampler.py +77 -0
- vismatch/third_party/MatchFormer/model/datasets/scannet.py +113 -0
- vismatch/third_party/MatchFormer/model/lightning_loftr.py +102 -0
- vismatch/third_party/MatchFormer/model/matchformer.py +54 -0
- vismatch/third_party/MatchFormer/model/utils/augment.py +55 -0
- vismatch/third_party/MatchFormer/model/utils/comm.py +265 -0
- vismatch/third_party/MatchFormer/model/utils/dataloader.py +23 -0
- vismatch/third_party/MatchFormer/model/utils/metrics.py +193 -0
- vismatch/third_party/MatchFormer/model/utils/misc.py +101 -0
- vismatch/third_party/MatchFormer/model/utils/profiler.py +39 -0
- vismatch/third_party/MatchFormer/test.py +55 -0
- vismatch/third_party/RIPE/app.py +272 -0
- vismatch/third_party/RIPE/demo.py +51 -0
- vismatch/third_party/RIPE/ripe/__init__.py +1 -0
- vismatch/third_party/RIPE/ripe/benchmarks/imw_2020.py +320 -0
- vismatch/third_party/RIPE/ripe/data/__init__.py +0 -0
- vismatch/third_party/RIPE/ripe/data/data_transforms.py +204 -0
- vismatch/third_party/RIPE/ripe/data/datasets/__init__.py +0 -0
- vismatch/third_party/RIPE/ripe/data/datasets/acdc.py +154 -0
- vismatch/third_party/RIPE/ripe/data/datasets/dataset_combinator.py +88 -0
- vismatch/third_party/RIPE/ripe/data/datasets/disk_imw.py +160 -0
- vismatch/third_party/RIPE/ripe/data/datasets/disk_megadepth.py +157 -0
- vismatch/third_party/RIPE/ripe/data/datasets/tokyo247.py +134 -0
- vismatch/third_party/RIPE/ripe/data/datasets/tokyo_query_v3.py +78 -0
- vismatch/third_party/RIPE/ripe/losses/__init__.py +0 -0
- vismatch/third_party/RIPE/ripe/losses/contrastive_loss.py +88 -0
- vismatch/third_party/RIPE/ripe/matcher/__init__.py +0 -0
- vismatch/third_party/RIPE/ripe/matcher/concurrent_matcher.py +97 -0
- vismatch/third_party/RIPE/ripe/matcher/pose_estimator_poselib.py +31 -0
- vismatch/third_party/RIPE/ripe/model_zoo/__init__.py +1 -0
- vismatch/third_party/RIPE/ripe/model_zoo/vgg_hyper.py +39 -0
- vismatch/third_party/RIPE/ripe/models/__init__.py +0 -0
- vismatch/third_party/RIPE/ripe/models/backbones/__init__.py +0 -0
- vismatch/third_party/RIPE/ripe/models/backbones/backbone_base.py +61 -0
- vismatch/third_party/RIPE/ripe/models/backbones/vgg.py +99 -0
- vismatch/third_party/RIPE/ripe/models/backbones/vgg_utils.py +143 -0
- vismatch/third_party/RIPE/ripe/models/ripe.py +303 -0
- vismatch/third_party/RIPE/ripe/models/upsampler/hypercolumn_features.py +54 -0
- vismatch/third_party/RIPE/ripe/models/upsampler/interpolate_sparse2d.py +37 -0
- vismatch/third_party/RIPE/ripe/scheduler/__init__.py +0 -0
- vismatch/third_party/RIPE/ripe/scheduler/constant.py +6 -0
- vismatch/third_party/RIPE/ripe/scheduler/expDecay.py +26 -0
- vismatch/third_party/RIPE/ripe/scheduler/linearLR.py +37 -0
- vismatch/third_party/RIPE/ripe/scheduler/linear_with_plateaus.py +44 -0
- vismatch/third_party/RIPE/ripe/train.py +410 -0
- vismatch/third_party/RIPE/ripe/utils/__init__.py +2 -0
- vismatch/third_party/RIPE/ripe/utils/image_utils.py +62 -0
- vismatch/third_party/RIPE/ripe/utils/pose_error.py +62 -0
- vismatch/third_party/RIPE/ripe/utils/pylogger.py +32 -0
- vismatch/third_party/RIPE/ripe/utils/utils.py +192 -0
- vismatch/third_party/RIPE/ripe/utils/wandb_utils.py +16 -0
- vismatch/third_party/RoMa/demo/demo_3D_effect.py +47 -0
- vismatch/third_party/RoMa/demo/demo_fundamental.py +34 -0
- vismatch/third_party/RoMa/demo/demo_match.py +50 -0
- vismatch/third_party/RoMa/demo/demo_match_opencv_sift.py +43 -0
- vismatch/third_party/RoMa/demo/demo_match_tiny.py +77 -0
- vismatch/third_party/RoMa/experiments/eval_roma_outdoor.py +57 -0
- vismatch/third_party/RoMa/experiments/eval_tiny_roma_v1_outdoor.py +84 -0
- vismatch/third_party/RoMa/experiments/roma_indoor.py +320 -0
- vismatch/third_party/RoMa/experiments/train_roma_outdoor.py +307 -0
- vismatch/third_party/RoMa/experiments/train_tiny_roma_v1_outdoor.py +498 -0
- vismatch/third_party/RoMa/romatch/__init__.py +8 -0
- vismatch/third_party/RoMa/romatch/benchmarks/__init__.py +6 -0
- vismatch/third_party/RoMa/romatch/benchmarks/hpatches_sequences_homog_benchmark.py +113 -0
- vismatch/third_party/RoMa/romatch/benchmarks/megadepth_dense_benchmark.py +106 -0
- vismatch/third_party/RoMa/romatch/benchmarks/megadepth_pose_estimation_benchmark.py +118 -0
- vismatch/third_party/RoMa/romatch/benchmarks/megadepth_pose_estimation_benchmark_poselib.py +119 -0
- vismatch/third_party/RoMa/romatch/benchmarks/scannet_benchmark.py +143 -0
- vismatch/third_party/RoMa/romatch/checkpointing/__init__.py +1 -0
- vismatch/third_party/RoMa/romatch/checkpointing/checkpoint.py +60 -0
- vismatch/third_party/RoMa/romatch/datasets/__init__.py +2 -0
- vismatch/third_party/RoMa/romatch/datasets/megadepth.py +232 -0
- vismatch/third_party/RoMa/romatch/datasets/scannet.py +160 -0
- vismatch/third_party/RoMa/romatch/losses/__init__.py +1 -0
- vismatch/third_party/RoMa/romatch/losses/robust_loss.py +161 -0
- vismatch/third_party/RoMa/romatch/losses/robust_loss_tiny_roma.py +160 -0
- vismatch/third_party/RoMa/romatch/models/__init__.py +1 -0
- vismatch/third_party/RoMa/romatch/models/encoders.py +122 -0
- vismatch/third_party/RoMa/romatch/models/matcher.py +748 -0
- vismatch/third_party/RoMa/romatch/models/model_zoo/__init__.py +73 -0
- vismatch/third_party/RoMa/romatch/models/model_zoo/roma_models.py +170 -0
- vismatch/third_party/RoMa/romatch/models/tiny.py +304 -0
- vismatch/third_party/RoMa/romatch/models/transformer/__init__.py +48 -0
- vismatch/third_party/RoMa/romatch/models/transformer/dinov2.py +359 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/__init__.py +12 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/attention.py +81 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/block.py +252 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/dino_head.py +59 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/drop_path.py +35 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/layer_scale.py +28 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/mlp.py +41 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/patch_embed.py +89 -0
- vismatch/third_party/RoMa/romatch/models/transformer/layers/swiglu_ffn.py +63 -0
- vismatch/third_party/RoMa/romatch/train/__init__.py +1 -0
- vismatch/third_party/RoMa/romatch/train/train.py +102 -0
- vismatch/third_party/RoMa/romatch/utils/__init__.py +16 -0
- vismatch/third_party/RoMa/romatch/utils/kde.py +13 -0
- vismatch/third_party/RoMa/romatch/utils/local_correlation.py +48 -0
- vismatch/third_party/RoMa/romatch/utils/transforms.py +118 -0
- vismatch/third_party/RoMa/romatch/utils/utils.py +654 -0
- vismatch/third_party/RoMa/setup.py +9 -0
- vismatch/third_party/RoMaV2/demo/demo_covariance.py +52 -0
- vismatch/third_party/RoMaV2/demo/demo_match.py +55 -0
- vismatch/third_party/RoMaV2/src/romav2/__init__.py +8 -0
- vismatch/third_party/RoMaV2/src/romav2/benchmarks/__init__.py +4 -0
- vismatch/third_party/RoMaV2/src/romav2/benchmarks/mega1500.py +115 -0
- vismatch/third_party/RoMaV2/src/romav2/benchmarks/satast.py +463 -0
- vismatch/third_party/RoMaV2/src/romav2/benchmarks/scannet1500.py +125 -0
- vismatch/third_party/RoMaV2/src/romav2/benchmarks/wxbs.py +104 -0
- vismatch/third_party/RoMaV2/src/romav2/device.py +9 -0
- vismatch/third_party/RoMaV2/src/romav2/dpt.py +516 -0
- vismatch/third_party/RoMaV2/src/romav2/features.py +190 -0
- vismatch/third_party/RoMaV2/src/romav2/geometry.py +261 -0
- vismatch/third_party/RoMaV2/src/romav2/io.py +24 -0
- vismatch/third_party/RoMaV2/src/romav2/local_correlation.py +152 -0
- vismatch/third_party/RoMaV2/src/romav2/logging.py +97 -0
- vismatch/third_party/RoMaV2/src/romav2/matcher.py +207 -0
- vismatch/third_party/RoMaV2/src/romav2/normalizers.py +17 -0
- vismatch/third_party/RoMaV2/src/romav2/refiner.py +277 -0
- vismatch/third_party/RoMaV2/src/romav2/romav2.py +533 -0
- vismatch/third_party/RoMaV2/src/romav2/types.py +75 -0
- vismatch/third_party/RoMaV2/src/romav2/vis.py +36 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/__init__.py +304 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/attention.py +181 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/block.py +293 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/ffn_layers.py +83 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/layer_scale.py +29 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/patch_embed.py +94 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/rms_norm.py +24 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/rope.py +133 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/rope_mixed.py +111 -0
- vismatch/third_party/RoMaV2/src/romav2/vit/utils.py +48 -0
- vismatch/third_party/RoMaV2/tests/test_bidirectional.py +93 -0
- vismatch/third_party/RoMaV2/tests/test_fps.py +49 -0
- vismatch/third_party/RoMaV2/tests/test_mega1500.py +22 -0
- vismatch/third_party/RoMaV2/tests/test_scannet1500.py +21 -0
- vismatch/third_party/RoMaV2/tests/test_smoke.py +15 -0
- vismatch/third_party/Se2_LoFTR/configs/data/__init__.py +0 -0
- vismatch/third_party/Se2_LoFTR/configs/data/base.py +35 -0
- vismatch/third_party/Se2_LoFTR/configs/data/megadepth_test_1500.py +11 -0
- vismatch/third_party/Se2_LoFTR/configs/data/megadepth_trainval_640.py +22 -0
- vismatch/third_party/Se2_LoFTR/configs/data/megadepth_trainval_840.py +22 -0
- vismatch/third_party/Se2_LoFTR/configs/data/scannet_test_1500.py +11 -0
- vismatch/third_party/Se2_LoFTR/configs/data/scannet_trainval.py +17 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ds.py +6 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ds_dense.py +8 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ot.py +6 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/buggy_pos_enc/loftr_ot_dense.py +8 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/loftr_ds.py +5 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/loftr_ds_dense.py +7 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/loftr_ot.py +5 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/loftr_ot_dense.py +7 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/scannet/loftr_ds_eval.py +16 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/indoor/scannet/loftr_ds_eval_new.py +18 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ds.py +16 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ds_dense.py +17 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ot.py +16 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/buggy_pos_enc/loftr_ot_dense.py +17 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ds.py +17 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ds_dense.py +17 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ds_e2.py +20 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ds_e2_dense.py +23 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ds_e2_dense_8rot.py +23 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ds_e2_dense_big.py +22 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ot.py +17 -0
- vismatch/third_party/Se2_LoFTR/configs/loftr/outdoor/loftr_ot_dense.py +18 -0
- vismatch/third_party/Se2_LoFTR/demo/demo_loftr.py +240 -0
- vismatch/third_party/Se2_LoFTR/src/__init__.py +0 -0
- vismatch/third_party/Se2_LoFTR/src/config/default.py +173 -0
- vismatch/third_party/Se2_LoFTR/src/datasets/megadepth.py +127 -0
- vismatch/third_party/Se2_LoFTR/src/datasets/sampler.py +77 -0
- vismatch/third_party/Se2_LoFTR/src/datasets/scannet.py +114 -0
- vismatch/third_party/Se2_LoFTR/src/lightning/data.py +320 -0
- vismatch/third_party/Se2_LoFTR/src/lightning/lightning_loftr.py +249 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/__init__.py +2 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/backbone/__init__.py +17 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/backbone/resnet_e2.py +170 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/backbone/resnet_fpn.py +199 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/loftr.py +81 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/loftr_module/__init__.py +2 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/loftr_module/fine_preprocess.py +59 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/loftr_module/linear_attention.py +81 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/loftr_module/transformer.py +101 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/utils/coarse_matching.py +261 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/utils/cvpr_ds_config.py +50 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/utils/fine_matching.py +74 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/utils/geometry.py +54 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/utils/position_encoding.py +42 -0
- vismatch/third_party/Se2_LoFTR/src/loftr/utils/supervision.py +151 -0
- vismatch/third_party/Se2_LoFTR/src/losses/loftr_loss.py +192 -0
- vismatch/third_party/Se2_LoFTR/src/optimizers/__init__.py +42 -0
- vismatch/third_party/Se2_LoFTR/src/utils/augment.py +55 -0
- vismatch/third_party/Se2_LoFTR/src/utils/comm.py +265 -0
- vismatch/third_party/Se2_LoFTR/src/utils/dataloader.py +23 -0
- vismatch/third_party/Se2_LoFTR/src/utils/dataset.py +185 -0
- vismatch/third_party/Se2_LoFTR/src/utils/metrics.py +193 -0
- vismatch/third_party/Se2_LoFTR/src/utils/misc.py +104 -0
- vismatch/third_party/Se2_LoFTR/src/utils/plotting.py +154 -0
- vismatch/third_party/Se2_LoFTR/src/utils/profiler.py +39 -0
- vismatch/third_party/Se2_LoFTR/test.py +68 -0
- vismatch/third_party/Se2_LoFTR/train.py +123 -0
- vismatch/third_party/SphereGlue/demo_SphereGlue.py +141 -0
- vismatch/third_party/SphereGlue/model/sphereglue.py +230 -0
- vismatch/third_party/SphereGlue/utils/Utils.py +191 -0
- vismatch/third_party/SphereGlue/utils/demo_mydataset.py +119 -0
- vismatch/third_party/Steerers/rotation_steerers/matchers/dual_softmax_matcher.py +44 -0
- vismatch/third_party/Steerers/rotation_steerers/matchers/max_matches.py +205 -0
- vismatch/third_party/Steerers/rotation_steerers/matchers/max_similarity.py +115 -0
- vismatch/third_party/Steerers/rotation_steerers/steerers.py +37 -0
- vismatch/third_party/Steerers/setup.py +14 -0
- vismatch/third_party/TopicFM/configs/megadepth_test.py +17 -0
- vismatch/third_party/TopicFM/configs/megadepth_test_topicfmfast.py +17 -0
- vismatch/third_party/TopicFM/configs/megadepth_test_topicfmplus.py +20 -0
- vismatch/third_party/TopicFM/configs/megadepth_train.py +36 -0
- vismatch/third_party/TopicFM/configs/megadepth_train_topicfmfast.py +34 -0
- vismatch/third_party/TopicFM/configs/megadepth_train_topicfmplus.py +37 -0
- vismatch/third_party/TopicFM/configs/scannet_test.py +15 -0
- vismatch/third_party/TopicFM/configs/scannet_test_topicfmfast.py +15 -0
- vismatch/third_party/TopicFM/configs/scannet_test_topicfmplus.py +19 -0
- vismatch/third_party/TopicFM/src/__init__.py +11 -0
- vismatch/third_party/TopicFM/src/config/default.py +174 -0
- vismatch/third_party/TopicFM/src/datasets/aachen.py +29 -0
- vismatch/third_party/TopicFM/src/datasets/custom_dataloader.py +126 -0
- vismatch/third_party/TopicFM/src/datasets/inloc.py +29 -0
- vismatch/third_party/TopicFM/src/datasets/megadepth.py +170 -0
- vismatch/third_party/TopicFM/src/datasets/sampler.py +77 -0
- vismatch/third_party/TopicFM/src/datasets/scannet.py +115 -0
- vismatch/third_party/TopicFM/src/lightning_trainer/data.py +292 -0
- vismatch/third_party/TopicFM/src/lightning_trainer/trainer.py +244 -0
- vismatch/third_party/TopicFM/src/losses/loss.py +228 -0
- vismatch/third_party/TopicFM/src/models/__init__.py +1 -0
- vismatch/third_party/TopicFM/src/models/backbone/__init__.py +12 -0
- vismatch/third_party/TopicFM/src/models/backbone/convnext.py +165 -0
- vismatch/third_party/TopicFM/src/models/backbone/fpn.py +114 -0
- vismatch/third_party/TopicFM/src/models/modules/__init__.py +2 -0
- vismatch/third_party/TopicFM/src/models/modules/encoder.py +266 -0
- vismatch/third_party/TopicFM/src/models/modules/fine_preprocess.py +59 -0
- vismatch/third_party/TopicFM/src/models/modules/linear_attention.py +84 -0
- vismatch/third_party/TopicFM/src/models/topic_fm.py +100 -0
- vismatch/third_party/TopicFM/src/models/utils/coarse_matching.py +213 -0
- vismatch/third_party/TopicFM/src/models/utils/fine_matching.py +172 -0
- vismatch/third_party/TopicFM/src/models/utils/geometry.py +54 -0
- vismatch/third_party/TopicFM/src/models/utils/supervision.py +167 -0
- vismatch/third_party/TopicFM/src/optimizers/__init__.py +42 -0
- vismatch/third_party/TopicFM/src/utils/augment.py +55 -0
- vismatch/third_party/TopicFM/src/utils/comm.py +265 -0
- vismatch/third_party/TopicFM/src/utils/dataloader.py +23 -0
- vismatch/third_party/TopicFM/src/utils/dataset.py +206 -0
- vismatch/third_party/TopicFM/src/utils/metrics.py +193 -0
- vismatch/third_party/TopicFM/src/utils/misc.py +101 -0
- vismatch/third_party/TopicFM/src/utils/plotting.py +313 -0
- vismatch/third_party/TopicFM/src/utils/profiler.py +39 -0
- vismatch/third_party/TopicFM/test.py +70 -0
- vismatch/third_party/TopicFM/third_party/__init__.py +0 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/aspan/indoor/aspan_test.py +7 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/aspan/indoor/aspan_train.py +8 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/aspan/outdoor/aspan_test.py +18 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/aspan/outdoor/aspan_train.py +17 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/data/__init__.py +0 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/data/base.py +35 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/data/megadepth_test_1500.py +13 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/data/megadepth_trainval_832.py +22 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/data/scannet_test_1500.py +11 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/configs/data/scannet_trainval.py +17 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/demo/demo.py +63 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/demo/demo_utils.py +44 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/__init__.py +2 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/aspan_module/__init__.py +3 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/aspan_module/attention.py +198 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/aspan_module/fine_preprocess.py +59 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/aspan_module/loftr.py +112 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/aspan_module/transformer.py +244 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/aspanformer.py +133 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/backbone/__init__.py +11 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/backbone/resnet_fpn.py +199 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/utils/coarse_matching.py +331 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/utils/cvpr_ds_config.py +50 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/utils/fine_matching.py +74 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/utils/geometry.py +54 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/utils/position_encoding.py +61 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/ASpanFormer/utils/supervision.py +151 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/__init__.py +0 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/config/default.py +180 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/datasets/__init__.py +3 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/datasets/megadepth.py +127 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/datasets/sampler.py +77 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/datasets/scannet.py +113 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/lightning/data.py +326 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/lightning/lightning_aspanformer.py +276 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/losses/aspan_loss.py +231 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/optimizers/__init__.py +42 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/augment.py +55 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/comm.py +265 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/dataloader.py +23 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/dataset.py +222 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/metrics.py +260 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/misc.py +139 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/plotting.py +219 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/src/utils/profiler.py +39 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/test.py +69 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/tools/SensorData.py +125 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/tools/extract.py +47 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/tools/preprocess_scene.py +242 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/tools/reader.py +39 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/tools/undistort_mega.py +69 -0
- vismatch/third_party/TopicFM/third_party/aspanformer/train.py +134 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/data/__init__.py +0 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/data/base.py +35 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/data/megadepth_test_1500.py +11 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/data/megadepth_trainval_640.py +22 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/data/megadepth_trainval_840.py +22 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/data/scannet_test_1500.py +11 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/data/scannet_trainval.py +17 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/buggy_pos_enc/loftr_ds.py +6 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/buggy_pos_enc/loftr_ds_dense.py +8 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/buggy_pos_enc/loftr_ot.py +6 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/buggy_pos_enc/loftr_ot_dense.py +8 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/loftr_ds.py +5 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/loftr_ds_dense.py +7 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/loftr_ot.py +5 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/loftr_ot_dense.py +7 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/scannet/loftr_ds_eval.py +16 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/indoor/scannet/loftr_ds_eval_new.py +18 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/buggy_pos_enc/loftr_ds.py +16 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/buggy_pos_enc/loftr_ds_dense.py +17 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/buggy_pos_enc/loftr_ot.py +16 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/buggy_pos_enc/loftr_ot_dense.py +17 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/loftr_ds.py +15 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/loftr_ds_dense.py +16 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/loftr_ot.py +15 -0
- vismatch/third_party/TopicFM/third_party/loftr/configs/loftr/outdoor/loftr_ot_dense.py +16 -0
- vismatch/third_party/TopicFM/third_party/loftr/demo/demo_loftr.py +240 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/__init__.py +0 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/config/default.py +171 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/datasets/megadepth.py +127 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/datasets/sampler.py +77 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/datasets/scannet.py +114 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/lightning/data.py +320 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/lightning/lightning_loftr.py +249 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/__init__.py +2 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/backbone/__init__.py +11 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/backbone/resnet_fpn.py +199 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/loftr.py +81 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/loftr_module/__init__.py +2 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/loftr_module/fine_preprocess.py +59 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/loftr_module/linear_attention.py +81 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/loftr_module/transformer.py +101 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/utils/coarse_matching.py +261 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/utils/cvpr_ds_config.py +50 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/utils/fine_matching.py +74 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/utils/geometry.py +54 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/utils/position_encoding.py +42 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/loftr/utils/supervision.py +151 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/losses/loftr_loss.py +192 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/optimizers/__init__.py +42 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/augment.py +55 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/comm.py +265 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/dataloader.py +23 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/dataset.py +185 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/metrics.py +193 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/misc.py +101 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/plotting.py +154 -0
- vismatch/third_party/TopicFM/third_party/loftr/src/utils/profiler.py +39 -0
- vismatch/third_party/TopicFM/third_party/loftr/test.py +68 -0
- vismatch/third_party/TopicFM/third_party/loftr/third_party/SuperGluePretrainedNetwork/demo_superglue.py +259 -0
- vismatch/third_party/TopicFM/third_party/loftr/third_party/SuperGluePretrainedNetwork/match_pairs.py +425 -0
- vismatch/third_party/TopicFM/third_party/loftr/third_party/SuperGluePretrainedNetwork/models/__init__.py +0 -0
- vismatch/third_party/TopicFM/third_party/loftr/third_party/SuperGluePretrainedNetwork/models/matching.py +84 -0
- vismatch/third_party/TopicFM/third_party/loftr/third_party/SuperGluePretrainedNetwork/models/superglue.py +283 -0
- vismatch/third_party/TopicFM/third_party/loftr/third_party/SuperGluePretrainedNetwork/models/superpoint.py +202 -0
- vismatch/third_party/TopicFM/third_party/loftr/third_party/SuperGluePretrainedNetwork/models/utils.py +555 -0
- vismatch/third_party/TopicFM/third_party/loftr/train.py +123 -0
- vismatch/third_party/TopicFM/third_party/matchformer/config/data/__init__.py +0 -0
- vismatch/third_party/TopicFM/third_party/matchformer/config/data/base.py +35 -0
- vismatch/third_party/TopicFM/third_party/matchformer/config/data/megadepth_test_1500.py +11 -0
- vismatch/third_party/TopicFM/third_party/matchformer/config/data/scannet_test_1500.py +11 -0
- vismatch/third_party/TopicFM/third_party/matchformer/config/defaultmf.py +88 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/__init__.py +17 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/coarse_matching.py +228 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/fine_matching.py +74 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/fine_preprocess.py +59 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/match_LA_large.py +254 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/match_LA_lite.py +254 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/match_SEA_large.py +291 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/backbone/match_SEA_lite.py +291 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/data.py +320 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/datasets/dataset.py +231 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/datasets/megadepth.py +126 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/datasets/sampler.py +77 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/datasets/scannet.py +113 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/lightning_loftr.py +102 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/matchformer.py +54 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/utils/augment.py +55 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/utils/comm.py +265 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/utils/dataloader.py +23 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/utils/metrics.py +193 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/utils/misc.py +101 -0
- vismatch/third_party/TopicFM/third_party/matchformer/model/utils/profiler.py +39 -0
- vismatch/third_party/TopicFM/third_party/matchformer/test.py +55 -0
- vismatch/third_party/TopicFM/train.py +123 -0
- vismatch/third_party/TopicFM/visualization.py +123 -0
- vismatch/third_party/TopicFM/viz/__init__.py +1 -0
- vismatch/third_party/TopicFM/viz/configs/__init__.py +0 -0
- vismatch/third_party/TopicFM/viz/methods/__init__.py +0 -0
- vismatch/third_party/TopicFM/viz/methods/base.py +70 -0
- vismatch/third_party/TopicFM/viz/methods/topicfmv2.py +208 -0
- vismatch/third_party/UFM/UniCeption/examples/models/cosmos/autoencoding.py +48 -0
- vismatch/third_party/UFM/UniCeption/examples/models/dust3r/convert_dust3r_weights_to_uniception.py +331 -0
- vismatch/third_party/UFM/UniCeption/examples/models/dust3r/dust3r.py +261 -0
- vismatch/third_party/UFM/UniCeption/examples/models/dust3r/profile_dust3r.py +47 -0
- vismatch/third_party/UFM/UniCeption/scripts/check_dependencies.py +48 -0
- vismatch/third_party/UFM/UniCeption/scripts/download_checkpoints.py +50 -0
- vismatch/third_party/UFM/UniCeption/scripts/install_croco_rope.py +61 -0
- vismatch/third_party/UFM/UniCeption/scripts/prepare_offline_install.py +398 -0
- vismatch/third_party/UFM/UniCeption/scripts/validate_installation.py +212 -0
- vismatch/third_party/UFM/UniCeption/setup.py +185 -0
- vismatch/third_party/UFM/UniCeption/tests/models/encoders/conftest.py +26 -0
- vismatch/third_party/UFM/UniCeption/tests/models/encoders/test_encoders.py +202 -0
- vismatch/third_party/UFM/UniCeption/tests/models/encoders/viz_image_encoders.py +294 -0
- vismatch/third_party/UFM/UniCeption/tests/models/info_sharing/viz_mulit_view_cross_attn_transformers.py +337 -0
- vismatch/third_party/UFM/UniCeption/uniception/__init__.py +0 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/__init__.py +0 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/__init__.py +225 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/base.py +157 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/cosmos.py +137 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/croco.py +457 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/dense_rep_encoder.py +344 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/dinov2.py +333 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/global_rep_encoder.py +115 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/image_normalizations.py +35 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/list.py +10 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/patch_embedder.py +235 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/radio.py +367 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/encoders/utils.py +86 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/factory/__init__.py +3 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/factory/dust3r.py +332 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/info_sharing/__init__.py +39 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/info_sharing/alternating_attention_transformer.py +973 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/info_sharing/base.py +116 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/info_sharing/cross_attention_transformer.py +612 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/info_sharing/diff_cross_attention_transformer.py +588 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/info_sharing/global_attention_transformer.py +1154 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/__init__.py +0 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/__init__.py +14 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/image_cli.py +175 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/image_lib.py +123 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/modules/__init__.py +60 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/modules/distributions.py +41 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/modules/layers2d.py +326 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/modules/layers3d.py +965 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/modules/patching.py +310 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/modules/quantizers.py +510 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/modules/utils.py +115 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/networks/__init__.py +39 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/networks/configs.py +146 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/networks/continuous_image.py +86 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/networks/continuous_video.py +98 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/networks/discrete_image.py +113 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/networks/discrete_video.py +115 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/utils.py +402 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/video_cli.py +195 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/cosmos_tokenizer/video_lib.py +145 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/__init__.py +0 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/blocks.py +249 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/curope/__init__.py +4 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/curope/curope2d.py +39 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/curope/setup.py +33 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/dpt_block.py +530 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/patch_embed.py +127 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/libs/croco/pos_embed.py +155 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/__init__.py +18 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/adaptors.py +1765 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/base.py +210 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/cosmos.py +211 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/dpt.py +676 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/global_head.py +142 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/linear.py +95 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/mlp_feature.py +114 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/mlp_head.py +114 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/moge_conv.py +342 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/prediction_heads/pose_head.py +181 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/utils/__init__.py +0 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/utils/config.py +34 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/utils/intermediate_feature_return.py +85 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/utils/positional_encoding.py +23 -0
- vismatch/third_party/UFM/UniCeption/uniception/models/utils/transformer_blocks.py +1072 -0
- vismatch/third_party/UFM/UniCeption/uniception/utils/__init__.py +0 -0
- vismatch/third_party/UFM/UniCeption/uniception/utils/profile.py +13 -0
- vismatch/third_party/UFM/UniCeption/uniception/utils/viz.py +99 -0
- vismatch/third_party/UFM/example_inference.py +138 -0
- vismatch/third_party/UFM/gradio_demo.py +238 -0
- vismatch/third_party/UFM/setup.py +86 -0
- vismatch/third_party/UFM/uniflowmatch/__init__.py +16 -0
- vismatch/third_party/UFM/uniflowmatch/cli.py +217 -0
- vismatch/third_party/UFM/uniflowmatch/models/__init__.py +25 -0
- vismatch/third_party/UFM/uniflowmatch/models/base.py +334 -0
- vismatch/third_party/UFM/uniflowmatch/models/ufm.py +1323 -0
- vismatch/third_party/UFM/uniflowmatch/models/unet_encoder.py +90 -0
- vismatch/third_party/UFM/uniflowmatch/models/utils.py +16 -0
- vismatch/third_party/UFM/uniflowmatch/utils/__init__.py +63 -0
- vismatch/third_party/UFM/uniflowmatch/utils/flow_resizing.py +1091 -0
- vismatch/third_party/UFM/uniflowmatch/utils/geometry.py +612 -0
- vismatch/third_party/UFM/uniflowmatch/utils/viz.py +97 -0
- vismatch/third_party/XoFTR/configs/data/__init__.py +0 -0
- vismatch/third_party/XoFTR/configs/data/base.py +35 -0
- vismatch/third_party/XoFTR/configs/data/megadepth_trainval_840.py +22 -0
- vismatch/third_party/XoFTR/configs/data/megadepth_vistir_trainval_640.py +23 -0
- vismatch/third_party/XoFTR/configs/data/pretrain.py +8 -0
- vismatch/third_party/XoFTR/configs/xoftr/outdoor/visible_thermal.py +17 -0
- vismatch/third_party/XoFTR/configs/xoftr/pretrain/pretrain.py +12 -0
- vismatch/third_party/XoFTR/pretrain.py +125 -0
- vismatch/third_party/XoFTR/src/__init__.py +0 -0
- vismatch/third_party/XoFTR/src/config/default.py +203 -0
- vismatch/third_party/XoFTR/src/datasets/megadepth.py +143 -0
- vismatch/third_party/XoFTR/src/datasets/pretrain_dataset.py +156 -0
- vismatch/third_party/XoFTR/src/datasets/sampler.py +77 -0
- vismatch/third_party/XoFTR/src/datasets/scannet.py +114 -0
- vismatch/third_party/XoFTR/src/datasets/vistir.py +109 -0
- vismatch/third_party/XoFTR/src/lightning/data.py +346 -0
- vismatch/third_party/XoFTR/src/lightning/data_pretrain.py +125 -0
- vismatch/third_party/XoFTR/src/lightning/lightning_xoftr.py +334 -0
- vismatch/third_party/XoFTR/src/lightning/lightning_xoftr_pretrain.py +171 -0
- vismatch/third_party/XoFTR/src/losses/xoftr_loss.py +170 -0
- vismatch/third_party/XoFTR/src/losses/xoftr_loss_pretrain.py +37 -0
- vismatch/third_party/XoFTR/src/optimizers/__init__.py +42 -0
- vismatch/third_party/XoFTR/src/utils/augment.py +113 -0
- vismatch/third_party/XoFTR/src/utils/comm.py +265 -0
- vismatch/third_party/XoFTR/src/utils/data_io.py +144 -0
- vismatch/third_party/XoFTR/src/utils/dataloader.py +23 -0
- vismatch/third_party/XoFTR/src/utils/dataset.py +279 -0
- vismatch/third_party/XoFTR/src/utils/metrics.py +211 -0
- vismatch/third_party/XoFTR/src/utils/misc.py +101 -0
- vismatch/third_party/XoFTR/src/utils/plotting.py +227 -0
- vismatch/third_party/XoFTR/src/utils/pretrain_utils.py +83 -0
- vismatch/third_party/XoFTR/src/utils/profiler.py +39 -0
- vismatch/third_party/XoFTR/src/xoftr/__init__.py +2 -0
- vismatch/third_party/XoFTR/src/xoftr/backbone/__init__.py +1 -0
- vismatch/third_party/XoFTR/src/xoftr/backbone/resnet.py +95 -0
- vismatch/third_party/XoFTR/src/xoftr/utils/geometry.py +107 -0
- vismatch/third_party/XoFTR/src/xoftr/utils/position_encoding.py +36 -0
- vismatch/third_party/XoFTR/src/xoftr/utils/supervision.py +290 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr.py +94 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr_module/__init__.py +4 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr_module/coarse_matching.py +305 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr_module/fine_matching.py +170 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr_module/fine_process.py +321 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr_module/linear_attention.py +81 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr_module/transformer.py +101 -0
- vismatch/third_party/XoFTR/src/xoftr/xoftr_pretrain.py +209 -0
- vismatch/third_party/XoFTR/test.py +68 -0
- vismatch/third_party/XoFTR/test_relative_pose.py +330 -0
- vismatch/third_party/XoFTR/train.py +126 -0
- vismatch/third_party/accelerated_features/hubconf.py +15 -0
- vismatch/third_party/accelerated_features/minimal_example.py +49 -0
- vismatch/third_party/accelerated_features/modules/__init__.py +4 -0
- vismatch/third_party/accelerated_features/modules/dataset/__init__.py +5 -0
- vismatch/third_party/accelerated_features/modules/dataset/augmentation.py +314 -0
- vismatch/third_party/accelerated_features/modules/dataset/megadepth/__init__.py +7 -0
- vismatch/third_party/accelerated_features/modules/dataset/megadepth/megadepth.py +174 -0
- vismatch/third_party/accelerated_features/modules/dataset/megadepth/megadepth_warper.py +170 -0
- vismatch/third_party/accelerated_features/modules/dataset/megadepth/utils.py +160 -0
- vismatch/third_party/accelerated_features/modules/interpolator.py +33 -0
- vismatch/third_party/accelerated_features/modules/lighterglue.py +56 -0
- vismatch/third_party/accelerated_features/modules/model.py +154 -0
- vismatch/third_party/accelerated_features/modules/training/__init__.py +4 -0
- vismatch/third_party/accelerated_features/modules/training/losses.py +224 -0
- vismatch/third_party/accelerated_features/modules/training/train.py +311 -0
- vismatch/third_party/accelerated_features/modules/training/utils.py +200 -0
- vismatch/third_party/accelerated_features/modules/xfeat.py +402 -0
- vismatch/third_party/accelerated_features/realtime_demo.py +295 -0
- vismatch/third_party/accelerated_features/third_party/ALIKE/alike.py +143 -0
- vismatch/third_party/accelerated_features/third_party/ALIKE/alnet.py +164 -0
- vismatch/third_party/accelerated_features/third_party/ALIKE/demo.py +167 -0
- vismatch/third_party/accelerated_features/third_party/ALIKE/hseq/eval.py +162 -0
- vismatch/third_party/accelerated_features/third_party/ALIKE/hseq/extract.py +159 -0
- vismatch/third_party/accelerated_features/third_party/ALIKE/soft_detect.py +194 -0
- vismatch/third_party/accelerated_features/third_party/__init__.py +4 -0
- vismatch/third_party/accelerated_features/third_party/alike_wrapper.py +110 -0
- vismatch/third_party/affine-steerers/affine_steerers/__init__.py +7 -0
- vismatch/third_party/affine-steerers/affine_steerers/benchmarks/__init__.py +5 -0
- vismatch/third_party/affine-steerers/affine_steerers/benchmarks/hpatches.py +92 -0
- vismatch/third_party/affine-steerers/affine_steerers/benchmarks/hpatches_oracle_steer.py +108 -0
- vismatch/third_party/affine-steerers/affine_steerers/benchmarks/mega_pose_est.py +116 -0
- vismatch/third_party/affine-steerers/affine_steerers/benchmarks/mega_pose_est_mnn.py +162 -0
- vismatch/third_party/affine-steerers/affine_steerers/benchmarks/nll_benchmark.py +57 -0
- vismatch/third_party/affine-steerers/affine_steerers/benchmarks/num_inliers.py +76 -0
- vismatch/third_party/affine-steerers/affine_steerers/checkpoint.py +82 -0
- vismatch/third_party/affine-steerers/affine_steerers/datasets/__init__.py +0 -0
- vismatch/third_party/affine-steerers/affine_steerers/datasets/homog.py +284 -0
- vismatch/third_party/affine-steerers/affine_steerers/datasets/megadepth.py +408 -0
- vismatch/third_party/affine-steerers/affine_steerers/decoder.py +90 -0
- vismatch/third_party/affine-steerers/affine_steerers/descriptors/__init__.py +0 -0
- vismatch/third_party/affine-steerers/affine_steerers/descriptors/dedode_descriptor.py +77 -0
- vismatch/third_party/affine-steerers/affine_steerers/descriptors/descriptor_loss.py +358 -0
- vismatch/third_party/affine-steerers/affine_steerers/detectors/__init__.py +0 -0
- vismatch/third_party/affine-steerers/affine_steerers/detectors/dedode_detector.py +75 -0
- vismatch/third_party/affine-steerers/affine_steerers/detectors/keypoint_loss.py +215 -0
- vismatch/third_party/affine-steerers/affine_steerers/encoder.py +87 -0
- vismatch/third_party/affine-steerers/affine_steerers/matchers/__init__.py +0 -0
- vismatch/third_party/affine-steerers/affine_steerers/matchers/dual_softmax_matcher.py +816 -0
- vismatch/third_party/affine-steerers/affine_steerers/model_zoo/__init__.py +3 -0
- vismatch/third_party/affine-steerers/affine_steerers/model_zoo/dedode_models.py +298 -0
- vismatch/third_party/affine-steerers/affine_steerers/steerers.py +732 -0
- vismatch/third_party/affine-steerers/affine_steerers/train.py +90 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/__init__.py +8 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/dinov2.py +359 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/__init__.py +12 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/attention.py +81 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/block.py +252 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/dino_head.py +59 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/drop_path.py +35 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/layer_scale.py +28 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/mlp.py +41 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/patch_embed.py +89 -0
- vismatch/third_party/affine-steerers/affine_steerers/transformer/layers/swiglu_ffn.py +63 -0
- vismatch/third_party/affine-steerers/affine_steerers/utils.py +1422 -0
- vismatch/third_party/affine-steerers/experiments/aff_equi_B.py +182 -0
- vismatch/third_party/affine-steerers/experiments/aff_equi_G.py +193 -0
- vismatch/third_party/affine-steerers/experiments/aff_steer_B.py +213 -0
- vismatch/third_party/affine-steerers/experiments/aff_steer_G.py +223 -0
- vismatch/third_party/affine-steerers/experiments/aff_steer_pretrain_B.py +187 -0
- vismatch/third_party/affine-steerers/experiments/aff_steer_pretrain_G.py +198 -0
- vismatch/third_party/affine-steerers/setup.py +15 -0
- vismatch/third_party/aspanformer/configs/aspan/indoor/aspan_test.py +7 -0
- vismatch/third_party/aspanformer/configs/aspan/indoor/aspan_train.py +8 -0
- vismatch/third_party/aspanformer/configs/aspan/outdoor/aspan_test.py +19 -0
- vismatch/third_party/aspanformer/configs/aspan/outdoor/aspan_train.py +17 -0
- vismatch/third_party/aspanformer/configs/data/__init__.py +0 -0
- vismatch/third_party/aspanformer/configs/data/base.py +35 -0
- vismatch/third_party/aspanformer/configs/data/megadepth_test_1500.py +13 -0
- vismatch/third_party/aspanformer/configs/data/megadepth_trainval_832.py +22 -0
- vismatch/third_party/aspanformer/configs/data/scannet_test_1500.py +11 -0
- vismatch/third_party/aspanformer/configs/data/scannet_trainval.py +17 -0
- vismatch/third_party/aspanformer/demo/demo.py +63 -0
- vismatch/third_party/aspanformer/demo/demo_utils.py +44 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/__init__.py +2 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/aspan_module/__init__.py +3 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/aspan_module/attention.py +198 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/aspan_module/fine_preprocess.py +59 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/aspan_module/loftr.py +112 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/aspan_module/transformer.py +244 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/aspanformer.py +152 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/backbone/__init__.py +11 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/backbone/resnet_fpn.py +199 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/utils/coarse_matching.py +331 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/utils/cvpr_ds_config.py +50 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/utils/fine_matching.py +74 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/utils/geometry.py +54 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/utils/position_encoding.py +61 -0
- vismatch/third_party/aspanformer/src/ASpanFormer/utils/supervision.py +151 -0
- vismatch/third_party/aspanformer/src/__init__.py +0 -0
- vismatch/third_party/aspanformer/src/config/default.py +180 -0
- vismatch/third_party/aspanformer/src/datasets/__init__.py +3 -0
- vismatch/third_party/aspanformer/src/datasets/megadepth.py +127 -0
- vismatch/third_party/aspanformer/src/datasets/sampler.py +77 -0
- vismatch/third_party/aspanformer/src/datasets/scannet.py +113 -0
- vismatch/third_party/aspanformer/src/lightning/data.py +326 -0
- vismatch/third_party/aspanformer/src/lightning/lightning_aspanformer.py +276 -0
- vismatch/third_party/aspanformer/src/losses/aspan_loss.py +231 -0
- vismatch/third_party/aspanformer/src/optimizers/__init__.py +42 -0
- vismatch/third_party/aspanformer/src/utils/augment.py +55 -0
- vismatch/third_party/aspanformer/src/utils/comm.py +265 -0
- vismatch/third_party/aspanformer/src/utils/dataloader.py +23 -0
- vismatch/third_party/aspanformer/src/utils/dataset.py +222 -0
- vismatch/third_party/aspanformer/src/utils/metrics.py +260 -0
- vismatch/third_party/aspanformer/src/utils/misc.py +139 -0
- vismatch/third_party/aspanformer/src/utils/plotting.py +219 -0
- vismatch/third_party/aspanformer/src/utils/profiler.py +39 -0
- vismatch/third_party/aspanformer/test.py +69 -0
- vismatch/third_party/aspanformer/tools/SensorData.py +125 -0
- vismatch/third_party/aspanformer/tools/extract.py +47 -0
- vismatch/third_party/aspanformer/tools/preprocess_scene.py +242 -0
- vismatch/third_party/aspanformer/tools/reader.py +39 -0
- vismatch/third_party/aspanformer/tools/undistort_mega.py +69 -0
- vismatch/third_party/aspanformer/train.py +134 -0
- vismatch/third_party/duster/croco/datasets/__init__.py +0 -0
- vismatch/third_party/duster/croco/datasets/crops/extract_crops_from_images.py +159 -0
- vismatch/third_party/duster/croco/datasets/habitat_sim/__init__.py +0 -0
- vismatch/third_party/duster/croco/datasets/habitat_sim/generate_from_metadata.py +92 -0
- vismatch/third_party/duster/croco/datasets/habitat_sim/generate_from_metadata_files.py +27 -0
- vismatch/third_party/duster/croco/datasets/habitat_sim/generate_multiview_images.py +177 -0
- vismatch/third_party/duster/croco/datasets/habitat_sim/multiview_habitat_sim_generator.py +390 -0
- vismatch/third_party/duster/croco/datasets/habitat_sim/pack_metadata_files.py +69 -0
- vismatch/third_party/duster/croco/datasets/habitat_sim/paths.py +129 -0
- vismatch/third_party/duster/croco/datasets/pairs_dataset.py +109 -0
- vismatch/third_party/duster/croco/datasets/transforms.py +95 -0
- vismatch/third_party/duster/croco/demo.py +55 -0
- vismatch/third_party/duster/croco/models/blocks.py +241 -0
- vismatch/third_party/duster/croco/models/criterion.py +37 -0
- vismatch/third_party/duster/croco/models/croco.py +249 -0
- vismatch/third_party/duster/croco/models/croco_downstream.py +122 -0
- vismatch/third_party/duster/croco/models/curope/__init__.py +4 -0
- vismatch/third_party/duster/croco/models/curope/curope2d.py +40 -0
- vismatch/third_party/duster/croco/models/curope/setup.py +34 -0
- vismatch/third_party/duster/croco/models/dpt_block.py +450 -0
- vismatch/third_party/duster/croco/models/head_downstream.py +58 -0
- vismatch/third_party/duster/croco/models/masking.py +25 -0
- vismatch/third_party/duster/croco/models/pos_embed.py +157 -0
- vismatch/third_party/duster/croco/pretrain.py +254 -0
- vismatch/third_party/duster/croco/stereoflow/augmentor.py +290 -0
- vismatch/third_party/duster/croco/stereoflow/criterion.py +251 -0
- vismatch/third_party/duster/croco/stereoflow/datasets_flow.py +630 -0
- vismatch/third_party/duster/croco/stereoflow/datasets_stereo.py +674 -0
- vismatch/third_party/duster/croco/stereoflow/engine.py +280 -0
- vismatch/third_party/duster/croco/stereoflow/test.py +216 -0
- vismatch/third_party/duster/croco/stereoflow/train.py +253 -0
- vismatch/third_party/duster/croco/utils/misc.py +463 -0
- vismatch/third_party/duster/datasets_preprocess/habitat/find_scenes.py +78 -0
- vismatch/third_party/duster/datasets_preprocess/habitat/habitat_renderer/__init__.py +2 -0
- vismatch/third_party/duster/datasets_preprocess/habitat/habitat_renderer/habitat_sim_envmaps_renderer.py +170 -0
- vismatch/third_party/duster/datasets_preprocess/habitat/habitat_renderer/multiview_crop_generator.py +93 -0
- vismatch/third_party/duster/datasets_preprocess/habitat/habitat_renderer/projections.py +151 -0
- vismatch/third_party/duster/datasets_preprocess/habitat/habitat_renderer/projections_conversions.py +45 -0
- vismatch/third_party/duster/datasets_preprocess/habitat/preprocess_habitat.py +121 -0
- vismatch/third_party/duster/datasets_preprocess/path_to_root.py +13 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_arkitscenes.py +355 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_blendedMVS.py +149 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_co3d.py +295 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_megadepth.py +198 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_scannetpp.py +400 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_staticthings3d.py +130 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_waymo.py +257 -0
- vismatch/third_party/duster/datasets_preprocess/preprocess_wildrgbd.py +209 -0
- vismatch/third_party/duster/demo.py +45 -0
- vismatch/third_party/duster/dust3r/__init__.py +2 -0
- vismatch/third_party/duster/dust3r/cloud_opt/__init__.py +33 -0
- vismatch/third_party/duster/dust3r/cloud_opt/base_opt.py +405 -0
- vismatch/third_party/duster/dust3r/cloud_opt/commons.py +90 -0
- vismatch/third_party/duster/dust3r/cloud_opt/init_im_poses.py +316 -0
- vismatch/third_party/duster/dust3r/cloud_opt/modular_optimizer.py +145 -0
- vismatch/third_party/duster/dust3r/cloud_opt/optimizer.py +248 -0
- vismatch/third_party/duster/dust3r/cloud_opt/pair_viewer.py +127 -0
- vismatch/third_party/duster/dust3r/datasets/__init__.py +50 -0
- vismatch/third_party/duster/dust3r/datasets/arkitscenes.py +102 -0
- vismatch/third_party/duster/dust3r/datasets/base/__init__.py +2 -0
- vismatch/third_party/duster/dust3r/datasets/base/base_stereo_view_dataset.py +220 -0
- vismatch/third_party/duster/dust3r/datasets/base/batched_sampler.py +74 -0
- vismatch/third_party/duster/dust3r/datasets/base/easy_dataset.py +157 -0
- vismatch/third_party/duster/dust3r/datasets/blendedmvs.py +104 -0
- vismatch/third_party/duster/dust3r/datasets/co3d.py +165 -0
- vismatch/third_party/duster/dust3r/datasets/habitat.py +107 -0
- vismatch/third_party/duster/dust3r/datasets/megadepth.py +123 -0
- vismatch/third_party/duster/dust3r/datasets/scannetpp.py +96 -0
- vismatch/third_party/duster/dust3r/datasets/staticthings3d.py +96 -0
- vismatch/third_party/duster/dust3r/datasets/utils/__init__.py +2 -0
- vismatch/third_party/duster/dust3r/datasets/utils/cropping.py +124 -0
- vismatch/third_party/duster/dust3r/datasets/utils/transforms.py +11 -0
- vismatch/third_party/duster/dust3r/datasets/waymo.py +93 -0
- vismatch/third_party/duster/dust3r/datasets/wildrgbd.py +67 -0
- vismatch/third_party/duster/dust3r/demo.py +287 -0
- vismatch/third_party/duster/dust3r/heads/__init__.py +19 -0
- vismatch/third_party/duster/dust3r/heads/dpt_head.py +115 -0
- vismatch/third_party/duster/dust3r/heads/linear_head.py +41 -0
- vismatch/third_party/duster/dust3r/heads/postprocess.py +58 -0
- vismatch/third_party/duster/dust3r/image_pairs.py +104 -0
- vismatch/third_party/duster/dust3r/inference.py +150 -0
- vismatch/third_party/duster/dust3r/losses.py +299 -0
- vismatch/third_party/duster/dust3r/model.py +211 -0
- vismatch/third_party/duster/dust3r/optim_factory.py +14 -0
- vismatch/third_party/duster/dust3r/patch_embed.py +70 -0
- vismatch/third_party/duster/dust3r/post_process.py +60 -0
- vismatch/third_party/duster/dust3r/training.py +377 -0
- vismatch/third_party/duster/dust3r/utils/__init__.py +2 -0
- vismatch/third_party/duster/dust3r/utils/device.py +76 -0
- vismatch/third_party/duster/dust3r/utils/geometry.py +366 -0
- vismatch/third_party/duster/dust3r/utils/image.py +128 -0
- vismatch/third_party/duster/dust3r/utils/misc.py +121 -0
- vismatch/third_party/duster/dust3r/utils/parallel.py +79 -0
- vismatch/third_party/duster/dust3r/utils/path_to_croco.py +19 -0
- vismatch/third_party/duster/dust3r/viz.py +381 -0
- vismatch/third_party/duster/dust3r_visloc/__init__.py +2 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/__init__.py +6 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/aachen_day_night.py +24 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/base_colmap.py +282 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/base_dataset.py +19 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/cambridge_landmarks.py +19 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/inloc.py +167 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/sevenscenes.py +123 -0
- vismatch/third_party/duster/dust3r_visloc/datasets/utils.py +118 -0
- vismatch/third_party/duster/dust3r_visloc/evaluation.py +65 -0
- vismatch/third_party/duster/dust3r_visloc/localization.py +140 -0
- vismatch/third_party/duster/train.py +13 -0
- vismatch/third_party/duster/visloc.py +193 -0
- vismatch/third_party/gim/demo.py +479 -0
- vismatch/third_party/gim/dkm/__init__.py +4 -0
- vismatch/third_party/gim/dkm/benchmarks/__init__.py +4 -0
- vismatch/third_party/gim/dkm/benchmarks/hpatches_sequences_homog_benchmark.py +114 -0
- vismatch/third_party/gim/dkm/benchmarks/megadepth1500_benchmark.py +124 -0
- vismatch/third_party/gim/dkm/benchmarks/megadepth_dense_benchmark.py +86 -0
- vismatch/third_party/gim/dkm/benchmarks/scannet_benchmark.py +143 -0
- vismatch/third_party/gim/dkm/checkpointing/__init__.py +1 -0
- vismatch/third_party/gim/dkm/checkpointing/checkpoint.py +31 -0
- vismatch/third_party/gim/dkm/datasets/__init__.py +1 -0
- vismatch/third_party/gim/dkm/datasets/megadepth.py +177 -0
- vismatch/third_party/gim/dkm/datasets/scannet.py +151 -0
- vismatch/third_party/gim/dkm/losses/__init__.py +1 -0
- vismatch/third_party/gim/dkm/losses/depth_match_regression_loss.py +128 -0
- vismatch/third_party/gim/dkm/models/__init__.py +4 -0
- vismatch/third_party/gim/dkm/models/dkm.py +745 -0
- vismatch/third_party/gim/dkm/models/encoders.py +148 -0
- vismatch/third_party/gim/dkm/models/model_zoo/DKMv3.py +148 -0
- vismatch/third_party/gim/dkm/models/model_zoo/__init__.py +39 -0
- vismatch/third_party/gim/dkm/train/__init__.py +1 -0
- vismatch/third_party/gim/dkm/train/train.py +67 -0
- vismatch/third_party/gim/dkm/utils/__init__.py +13 -0
- vismatch/third_party/gim/dkm/utils/kde.py +26 -0
- vismatch/third_party/gim/dkm/utils/local_correlation.py +40 -0
- vismatch/third_party/gim/dkm/utils/transforms.py +104 -0
- vismatch/third_party/gim/dkm/utils/utils.py +341 -0
- vismatch/third_party/gim/gluefactory/__init__.py +17 -0
- vismatch/third_party/gim/gluefactory/datasets/__init__.py +25 -0
- vismatch/third_party/gim/gluefactory/datasets/augmentations.py +244 -0
- vismatch/third_party/gim/gluefactory/datasets/base_dataset.py +206 -0
- vismatch/third_party/gim/gluefactory/datasets/eth3d.py +254 -0
- vismatch/third_party/gim/gluefactory/datasets/homographies.py +311 -0
- vismatch/third_party/gim/gluefactory/datasets/hpatches.py +145 -0
- vismatch/third_party/gim/gluefactory/datasets/image_folder.py +59 -0
- vismatch/third_party/gim/gluefactory/datasets/image_pairs.py +100 -0
- vismatch/third_party/gim/gluefactory/datasets/megadepth.py +514 -0
- vismatch/third_party/gim/gluefactory/datasets/utils.py +131 -0
- vismatch/third_party/gim/gluefactory/eval/__init__.py +20 -0
- vismatch/third_party/gim/gluefactory/eval/eth3d.py +202 -0
- vismatch/third_party/gim/gluefactory/eval/eval_pipeline.py +109 -0
- vismatch/third_party/gim/gluefactory/eval/hpatches.py +203 -0
- vismatch/third_party/gim/gluefactory/eval/inspect.py +61 -0
- vismatch/third_party/gim/gluefactory/eval/io.py +109 -0
- vismatch/third_party/gim/gluefactory/eval/megadepth1500.py +189 -0
- vismatch/third_party/gim/gluefactory/eval/utils.py +272 -0
- vismatch/third_party/gim/gluefactory/geometry/depth.py +88 -0
- vismatch/third_party/gim/gluefactory/geometry/epipolar.py +155 -0
- vismatch/third_party/gim/gluefactory/geometry/gt_generation.py +558 -0
- vismatch/third_party/gim/gluefactory/geometry/homography.py +342 -0
- vismatch/third_party/gim/gluefactory/geometry/utils.py +167 -0
- vismatch/third_party/gim/gluefactory/geometry/wrappers.py +425 -0
- vismatch/third_party/gim/gluefactory/models/__init__.py +30 -0
- vismatch/third_party/gim/gluefactory/models/backbones/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/models/backbones/dinov2.py +30 -0
- vismatch/third_party/gim/gluefactory/models/base_model.py +157 -0
- vismatch/third_party/gim/gluefactory/models/cache_loader.py +139 -0
- vismatch/third_party/gim/gluefactory/models/extractors/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/models/extractors/aliked.py +786 -0
- vismatch/third_party/gim/gluefactory/models/extractors/disk_kornia.py +108 -0
- vismatch/third_party/gim/gluefactory/models/extractors/grid_extractor.py +60 -0
- vismatch/third_party/gim/gluefactory/models/extractors/keynet_affnet_hardnet.py +74 -0
- vismatch/third_party/gim/gluefactory/models/extractors/mixed.py +76 -0
- vismatch/third_party/gim/gluefactory/models/extractors/sift.py +234 -0
- vismatch/third_party/gim/gluefactory/models/extractors/sift_kornia.py +46 -0
- vismatch/third_party/gim/gluefactory/models/extractors/superpoint_open.py +210 -0
- vismatch/third_party/gim/gluefactory/models/lines/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/models/lines/deeplsd.py +106 -0
- vismatch/third_party/gim/gluefactory/models/lines/lsd.py +88 -0
- vismatch/third_party/gim/gluefactory/models/lines/wireframe.py +312 -0
- vismatch/third_party/gim/gluefactory/models/matchers/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/models/matchers/adalam.py +0 -0
- vismatch/third_party/gim/gluefactory/models/matchers/depth_matcher.py +82 -0
- vismatch/third_party/gim/gluefactory/models/matchers/gluestick.py +776 -0
- vismatch/third_party/gim/gluefactory/models/matchers/homography_matcher.py +66 -0
- vismatch/third_party/gim/gluefactory/models/matchers/kornia_loftr.py +66 -0
- vismatch/third_party/gim/gluefactory/models/matchers/lightglue.py +632 -0
- vismatch/third_party/gim/gluefactory/models/matchers/lightglue_pretrained.py +36 -0
- vismatch/third_party/gim/gluefactory/models/matchers/nearest_neighbor_matcher.py +97 -0
- vismatch/third_party/gim/gluefactory/models/triplet_pipeline.py +99 -0
- vismatch/third_party/gim/gluefactory/models/two_view_pipeline.py +114 -0
- vismatch/third_party/gim/gluefactory/models/utils/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/models/utils/losses.py +73 -0
- vismatch/third_party/gim/gluefactory/models/utils/metrics.py +50 -0
- vismatch/third_party/gim/gluefactory/models/utils/misc.py +70 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/__init__.py +15 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/base_estimator.py +33 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/homography/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/homography/homography_est.py +74 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/homography/opencv.py +53 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/homography/poselib.py +40 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/relative_pose/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/relative_pose/opencv.py +64 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/relative_pose/poselib.py +44 -0
- vismatch/third_party/gim/gluefactory/robust_estimators/relative_pose/pycolmap.py +52 -0
- vismatch/third_party/gim/gluefactory/scripts/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/scripts/export_local_features.py +127 -0
- vismatch/third_party/gim/gluefactory/scripts/export_megadepth.py +173 -0
- vismatch/third_party/gim/gluefactory/settings.py +6 -0
- vismatch/third_party/gim/gluefactory/superpoint.py +361 -0
- vismatch/third_party/gim/gluefactory/train.py +691 -0
- vismatch/third_party/gim/gluefactory/utils/__init__.py +0 -0
- vismatch/third_party/gim/gluefactory/utils/benchmark.py +33 -0
- vismatch/third_party/gim/gluefactory/utils/experiments.py +134 -0
- vismatch/third_party/gim/gluefactory/utils/export_predictions.py +81 -0
- vismatch/third_party/gim/gluefactory/utils/image.py +130 -0
- vismatch/third_party/gim/gluefactory/utils/misc.py +44 -0
- vismatch/third_party/gim/gluefactory/utils/patches.py +50 -0
- vismatch/third_party/gim/gluefactory/utils/stdout_capturing.py +134 -0
- vismatch/third_party/gim/gluefactory/utils/tensor.py +48 -0
- vismatch/third_party/gim/gluefactory/utils/tools.py +269 -0
- vismatch/third_party/gim/gluefactory/visualization/global_frame.py +289 -0
- vismatch/third_party/gim/gluefactory/visualization/tools.py +465 -0
- vismatch/third_party/gim/gluefactory/visualization/two_view_frame.py +158 -0
- vismatch/third_party/gim/gluefactory/visualization/visualize_batch.py +57 -0
- vismatch/third_party/gim/gluefactory/visualization/viz2d.py +486 -0
- vismatch/third_party/imatch-toolbox/configs/d2net.yml +26 -0
- vismatch/third_party/imatch-toolbox/configs/dogaffnethardnet.yml +10 -0
- vismatch/third_party/imatch-toolbox/configs/ncnet.yml +7 -0
- vismatch/third_party/imatch-toolbox/configs/patch2pix.yml +56 -0
- vismatch/third_party/imatch-toolbox/configs/patch2pix_superglue.yml +58 -0
- vismatch/third_party/imatch-toolbox/configs/r2d2.yml +31 -0
- vismatch/third_party/imatch-toolbox/configs/sift.yml +27 -0
- vismatch/third_party/imatch-toolbox/configs/superglue.yml +69 -0
- vismatch/third_party/imatch-toolbox/configs/superpoint.yml +21 -0
- vismatch/third_party/imatch-toolbox/environment.yml +14 -0
- vismatch/third_party/imatch-toolbox/immatch/__init__.py +8 -0
- vismatch/third_party/imatch-toolbox/immatch/eval_aachen.py +88 -0
- vismatch/third_party/imatch-toolbox/immatch/eval_hpatches.py +117 -0
- vismatch/third_party/imatch-toolbox/immatch/eval_inloc.py +45 -0
- vismatch/third_party/imatch-toolbox/immatch/eval_relapose.py +231 -0
- vismatch/third_party/imatch-toolbox/immatch/eval_robotcar.py +83 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/__init__.py +0 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/base.py +89 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/d2net.py +69 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/dogaffnethardnet.py +94 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/nn_matching.py +31 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/patch2pix.py +126 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/r2d2.py +64 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/sift.py +67 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/superglue.py +62 -0
- vismatch/third_party/imatch-toolbox/immatch/modules/superpoint.py +56 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/__init__.py +13 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/colmap/data_parsing.py +257 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/colmap/database.py +362 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/colmap/read_write_model.py +506 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/data_io.py +111 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/hpatches_helper.py +242 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/localize_sfm_helper.py +403 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/metrics.py +90 -0
- vismatch/third_party/imatch-toolbox/immatch/utils/model_helper.py +27 -0
- vismatch/third_party/imatch-toolbox/setup.py +36 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/extract_features.py +156 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/extract_kapture.py +248 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/lib/dataset.py +239 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/lib/exceptions.py +6 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/lib/loss.py +340 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/lib/model.py +121 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/lib/model_test.py +187 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/lib/pyramid.py +129 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/lib/utils.py +167 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/megadepth_utils/preprocess_scene.py +242 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/megadepth_utils/undistort_reconstructions.py +69 -0
- vismatch/third_party/imatch-toolbox/third_party/d2net/train.py +279 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/data_pairs/precompute_immatch_val_ovs.py +20 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/environment.yml +21 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/networks/modules.py +167 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/networks/ncn/conv4d.py +91 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/networks/ncn/extract_ncmatches.py +158 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/networks/ncn/model.py +333 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/networks/patch2pix.py +403 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/networks/resnet.py +191 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/networks/utils.py +111 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/train_patch2pix.py +374 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/colmap/data_loading.py +169 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/colmap/read_database.py +175 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/colmap/read_write_model.py +483 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/common/plotting.py +393 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/common/setup_helper.py +59 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/common/visdom_helper.py +95 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/datasets/__init__.py +1 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/datasets/data_parsing.py +145 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/datasets/dataset_megadepth.py +141 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/datasets/preprocess.py +184 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/eval/geometry.py +90 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/eval/measure.py +161 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/eval/model_helper.py +129 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/train/eval_epoch_immatch.py +99 -0
- vismatch/third_party/imatch-toolbox/third_party/patch2pix/utils/train/helper.py +196 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/datasets/__init__.py +33 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/datasets/aachen.py +146 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/datasets/dataset.py +77 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/datasets/imgfolder.py +23 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/datasets/pair_dataset.py +287 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/datasets/web_images.py +64 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/extract.py +183 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/extract_kapture.py +194 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/nets/ap_loss.py +67 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/nets/losses.py +56 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/nets/patchnet.py +134 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/nets/reliability_loss.py +59 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/nets/repeatability_loss.py +66 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/nets/sampler.py +390 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/tools/common.py +41 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/tools/dataloader.py +367 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/tools/trainer.py +76 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/tools/transforms.py +513 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/tools/transforms_tools.py +230 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/tools/viz.py +191 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/train.py +138 -0
- vismatch/third_party/imatch-toolbox/third_party/r2d2/viz_heatmaps.py +122 -0
- vismatch/third_party/imatch-toolbox/third_party/superglue/demo_superglue.py +259 -0
- vismatch/third_party/imatch-toolbox/third_party/superglue/match_pairs.py +425 -0
- vismatch/third_party/imatch-toolbox/third_party/superglue/models/__init__.py +0 -0
- vismatch/third_party/imatch-toolbox/third_party/superglue/models/matching.py +84 -0
- vismatch/third_party/imatch-toolbox/third_party/superglue/models/superglue.py +283 -0
- vismatch/third_party/imatch-toolbox/third_party/superglue/models/superpoint.py +202 -0
- vismatch/third_party/imatch-toolbox/third_party/superglue/models/utils.py +555 -0
- vismatch/third_party/keypt2subpx/dataprocess/aliked.py +163 -0
- vismatch/third_party/keypt2subpx/dataprocess/dedode.py +215 -0
- vismatch/third_party/keypt2subpx/dataprocess/splg.py +162 -0
- vismatch/third_party/keypt2subpx/dataprocess/spnn.py +157 -0
- vismatch/third_party/keypt2subpx/dataprocess/superpoint_densescore.py +357 -0
- vismatch/third_party/keypt2subpx/dataprocess/xfeat.py +187 -0
- vismatch/third_party/keypt2subpx/dataset.py +145 -0
- vismatch/third_party/keypt2subpx/hubconf.py +38 -0
- vismatch/third_party/keypt2subpx/logger.py +127 -0
- vismatch/third_party/keypt2subpx/model.py +183 -0
- vismatch/third_party/keypt2subpx/settings.py +108 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/__init__.py +17 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/__init__.py +25 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/augmentations.py +244 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/base_dataset.py +206 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/eth3d.py +254 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/homographies.py +311 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/hpatches.py +145 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/image_folder.py +59 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/image_pairs.py +100 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/megadepth.py +510 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/datasets/utils.py +131 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/__init__.py +20 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/eth3d.py +202 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/eval_pipeline.py +109 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/hpatches.py +203 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/inspect.py +61 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/io.py +109 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/megadepth1500.py +189 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/eval/utils.py +272 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/geometry/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/geometry/depth.py +88 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/geometry/epipolar.py +155 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/geometry/gt_generation.py +558 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/geometry/homography.py +342 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/geometry/utils.py +167 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/geometry/wrappers.py +425 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/__init__.py +30 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/backbones/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/backbones/dinov2.py +30 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/base_model.py +157 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/cache_loader.py +139 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/aliked.py +786 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/disk_kornia.py +108 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/grid_extractor.py +60 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/keynet_affnet_hardnet.py +74 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/mixed.py +76 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/sift.py +234 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/sift_kornia.py +46 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/extractors/superpoint_open.py +210 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/lines/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/lines/deeplsd.py +106 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/lines/lsd.py +88 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/lines/wireframe.py +312 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/adalam.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/depth_matcher.py +82 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/gluestick.py +776 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/homography_matcher.py +66 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/kornia_loftr.py +66 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/lightglue.py +612 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/lightglue_pretrained.py +36 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/matchers/nearest_neighbor_matcher.py +97 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/triplet_pipeline.py +99 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/two_view_pipeline.py +114 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/utils/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/utils/losses.py +73 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/utils/metrics.py +50 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/models/utils/misc.py +70 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/__init__.py +15 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/base_estimator.py +33 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/homography/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/homography/homography_est.py +74 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/homography/opencv.py +53 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/homography/poselib.py +40 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/relative_pose/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/relative_pose/opencv.py +64 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/relative_pose/poselib.py +44 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/robust_estimators/relative_pose/pycolmap.py +52 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/scripts/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/scripts/export_local_features.py +127 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/scripts/export_megadepth.py +173 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/settings.py +6 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/train.py +691 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/benchmark.py +33 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/experiments.py +134 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/export_predictions.py +81 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/image.py +130 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/misc.py +44 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/patches.py +50 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/stdout_capturing.py +134 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/tensor.py +48 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/utils/tools.py +269 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/visualization/global_frame.py +289 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/visualization/tools.py +465 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/visualization/two_view_frame.py +158 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/visualization/visualize_batch.py +57 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory/visualization/viz2d.py +486 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory_nonfree/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory_nonfree/superglue.py +342 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/gluefactory_nonfree/superpoint.py +356 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/tests/__init__.py +0 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/tests/test_eval_utils.py +88 -0
- vismatch/third_party/keypt2subpx/submodules/glue_factory/tests/test_integration.py +132 -0
- vismatch/third_party/keypt2subpx/summarize.py +44 -0
- vismatch/third_party/keypt2subpx/test.py +225 -0
- vismatch/third_party/keypt2subpx/train.py +180 -0
- vismatch/third_party/keypt2subpx/utils.py +150 -0
- vismatch/third_party/mast3r/demo.py +51 -0
- vismatch/third_party/mast3r/demo_dust3r_ga.py +99 -0
- vismatch/third_party/mast3r/demo_glomap.py +52 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/__init__.py +0 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/crops/extract_crops_from_images.py +159 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/habitat_sim/__init__.py +0 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/habitat_sim/generate_from_metadata.py +92 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/habitat_sim/generate_from_metadata_files.py +27 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/habitat_sim/generate_multiview_images.py +177 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/habitat_sim/multiview_habitat_sim_generator.py +390 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/habitat_sim/pack_metadata_files.py +69 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/habitat_sim/paths.py +129 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/pairs_dataset.py +109 -0
- vismatch/third_party/mast3r/dust3r/croco/datasets/transforms.py +95 -0
- vismatch/third_party/mast3r/dust3r/croco/demo.py +55 -0
- vismatch/third_party/mast3r/dust3r/croco/models/blocks.py +241 -0
- vismatch/third_party/mast3r/dust3r/croco/models/criterion.py +37 -0
- vismatch/third_party/mast3r/dust3r/croco/models/croco.py +249 -0
- vismatch/third_party/mast3r/dust3r/croco/models/croco_downstream.py +122 -0
- vismatch/third_party/mast3r/dust3r/croco/models/curope/__init__.py +4 -0
- vismatch/third_party/mast3r/dust3r/croco/models/curope/curope2d.py +40 -0
- vismatch/third_party/mast3r/dust3r/croco/models/curope/setup.py +34 -0
- vismatch/third_party/mast3r/dust3r/croco/models/dpt_block.py +450 -0
- vismatch/third_party/mast3r/dust3r/croco/models/head_downstream.py +58 -0
- vismatch/third_party/mast3r/dust3r/croco/models/masking.py +25 -0
- vismatch/third_party/mast3r/dust3r/croco/models/pos_embed.py +157 -0
- vismatch/third_party/mast3r/dust3r/croco/pretrain.py +254 -0
- vismatch/third_party/mast3r/dust3r/croco/stereoflow/augmentor.py +290 -0
- vismatch/third_party/mast3r/dust3r/croco/stereoflow/criterion.py +251 -0
- vismatch/third_party/mast3r/dust3r/croco/stereoflow/datasets_flow.py +630 -0
- vismatch/third_party/mast3r/dust3r/croco/stereoflow/datasets_stereo.py +674 -0
- vismatch/third_party/mast3r/dust3r/croco/stereoflow/engine.py +280 -0
- vismatch/third_party/mast3r/dust3r/croco/stereoflow/test.py +216 -0
- vismatch/third_party/mast3r/dust3r/croco/stereoflow/train.py +253 -0
- vismatch/third_party/mast3r/dust3r/croco/utils/misc.py +463 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/habitat/find_scenes.py +78 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/habitat/habitat_renderer/__init__.py +2 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/habitat/habitat_renderer/habitat_sim_envmaps_renderer.py +170 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/habitat/habitat_renderer/multiview_crop_generator.py +93 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/habitat/habitat_renderer/projections.py +151 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/habitat/habitat_renderer/projections_conversions.py +45 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/habitat/preprocess_habitat.py +121 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/path_to_root.py +13 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_arkitscenes.py +355 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_blendedMVS.py +149 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_co3d.py +295 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_megadepth.py +198 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_scannetpp.py +390 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_staticthings3d.py +130 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_waymo.py +257 -0
- vismatch/third_party/mast3r/dust3r/datasets_preprocess/preprocess_wildrgbd.py +209 -0
- vismatch/third_party/mast3r/dust3r/demo.py +45 -0
- vismatch/third_party/mast3r/dust3r/dust3r/__init__.py +2 -0
- vismatch/third_party/mast3r/dust3r/dust3r/cloud_opt/__init__.py +33 -0
- vismatch/third_party/mast3r/dust3r/dust3r/cloud_opt/base_opt.py +405 -0
- vismatch/third_party/mast3r/dust3r/dust3r/cloud_opt/commons.py +90 -0
- vismatch/third_party/mast3r/dust3r/dust3r/cloud_opt/init_im_poses.py +316 -0
- vismatch/third_party/mast3r/dust3r/dust3r/cloud_opt/modular_optimizer.py +145 -0
- vismatch/third_party/mast3r/dust3r/dust3r/cloud_opt/optimizer.py +248 -0
- vismatch/third_party/mast3r/dust3r/dust3r/cloud_opt/pair_viewer.py +127 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/__init__.py +50 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/arkitscenes.py +102 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/base/__init__.py +2 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/base/base_stereo_view_dataset.py +220 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/base/batched_sampler.py +74 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/base/easy_dataset.py +157 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/blendedmvs.py +104 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/co3d.py +165 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/habitat.py +107 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/megadepth.py +123 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/scannetpp.py +96 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/staticthings3d.py +96 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/utils/__init__.py +2 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/utils/cropping.py +124 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/utils/transforms.py +11 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/waymo.py +93 -0
- vismatch/third_party/mast3r/dust3r/dust3r/datasets/wildrgbd.py +67 -0
- vismatch/third_party/mast3r/dust3r/dust3r/demo.py +287 -0
- vismatch/third_party/mast3r/dust3r/dust3r/heads/__init__.py +19 -0
- vismatch/third_party/mast3r/dust3r/dust3r/heads/dpt_head.py +115 -0
- vismatch/third_party/mast3r/dust3r/dust3r/heads/linear_head.py +41 -0
- vismatch/third_party/mast3r/dust3r/dust3r/heads/postprocess.py +58 -0
- vismatch/third_party/mast3r/dust3r/dust3r/image_pairs.py +104 -0
- vismatch/third_party/mast3r/dust3r/dust3r/inference.py +150 -0
- vismatch/third_party/mast3r/dust3r/dust3r/losses.py +299 -0
- vismatch/third_party/mast3r/dust3r/dust3r/model.py +211 -0
- vismatch/third_party/mast3r/dust3r/dust3r/optim_factory.py +14 -0
- vismatch/third_party/mast3r/dust3r/dust3r/patch_embed.py +70 -0
- vismatch/third_party/mast3r/dust3r/dust3r/post_process.py +60 -0
- vismatch/third_party/mast3r/dust3r/dust3r/training.py +377 -0
- vismatch/third_party/mast3r/dust3r/dust3r/utils/__init__.py +2 -0
- vismatch/third_party/mast3r/dust3r/dust3r/utils/device.py +76 -0
- vismatch/third_party/mast3r/dust3r/dust3r/utils/geometry.py +366 -0
- vismatch/third_party/mast3r/dust3r/dust3r/utils/image.py +128 -0
- vismatch/third_party/mast3r/dust3r/dust3r/utils/misc.py +121 -0
- vismatch/third_party/mast3r/dust3r/dust3r/utils/parallel.py +79 -0
- vismatch/third_party/mast3r/dust3r/dust3r/utils/path_to_croco.py +19 -0
- vismatch/third_party/mast3r/dust3r/dust3r/viz.py +381 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/__init__.py +2 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/__init__.py +6 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/aachen_day_night.py +24 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/base_colmap.py +282 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/base_dataset.py +19 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/cambridge_landmarks.py +19 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/inloc.py +167 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/sevenscenes.py +123 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/datasets/utils.py +118 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/evaluation.py +65 -0
- vismatch/third_party/mast3r/dust3r/dust3r_visloc/localization.py +140 -0
- vismatch/third_party/mast3r/dust3r/train.py +13 -0
- vismatch/third_party/mast3r/dust3r/visloc.py +193 -0
- vismatch/third_party/mast3r/kapture_mast3r_mapping.py +127 -0
- vismatch/third_party/mast3r/make_pairs.py +105 -0
- vismatch/third_party/mast3r/mast3r/__init__.py +2 -0
- vismatch/third_party/mast3r/mast3r/catmlp_dpt_head.py +239 -0
- vismatch/third_party/mast3r/mast3r/cloud_opt/__init__.py +2 -0
- vismatch/third_party/mast3r/mast3r/cloud_opt/sparse_ga.py +1078 -0
- vismatch/third_party/mast3r/mast3r/cloud_opt/triangulation.py +80 -0
- vismatch/third_party/mast3r/mast3r/cloud_opt/tsdf_optimizer.py +273 -0
- vismatch/third_party/mast3r/mast3r/cloud_opt/utils/__init__.py +2 -0
- vismatch/third_party/mast3r/mast3r/cloud_opt/utils/losses.py +32 -0
- vismatch/third_party/mast3r/mast3r/cloud_opt/utils/schedules.py +17 -0
- vismatch/third_party/mast3r/mast3r/colmap/__init__.py +2 -0
- vismatch/third_party/mast3r/mast3r/colmap/database.py +383 -0
- vismatch/third_party/mast3r/mast3r/colmap/mapping.py +196 -0
- vismatch/third_party/mast3r/mast3r/datasets/__init__.py +62 -0
- vismatch/third_party/mast3r/mast3r/datasets/base/__init__.py +2 -0
- vismatch/third_party/mast3r/mast3r/datasets/base/mast3r_base_stereo_view_dataset.py +355 -0
- vismatch/third_party/mast3r/mast3r/datasets/utils/__init__.py +2 -0
- vismatch/third_party/mast3r/mast3r/datasets/utils/cropping.py +219 -0
- vismatch/third_party/mast3r/mast3r/demo.py +381 -0
- vismatch/third_party/mast3r/mast3r/demo_glomap.py +343 -0
- vismatch/third_party/mast3r/mast3r/fast_nn.py +223 -0
- vismatch/third_party/mast3r/mast3r/image_pairs.py +115 -0
- vismatch/third_party/mast3r/mast3r/losses.py +508 -0
- vismatch/third_party/mast3r/mast3r/model.py +213 -0
- vismatch/third_party/mast3r/mast3r/retrieval/graph.py +77 -0
- vismatch/third_party/mast3r/mast3r/retrieval/model.py +271 -0
- vismatch/third_party/mast3r/mast3r/retrieval/processor.py +129 -0
- vismatch/third_party/mast3r/mast3r/utils/__init__.py +2 -0
- vismatch/third_party/mast3r/mast3r/utils/coarse_to_fine.py +214 -0
- vismatch/third_party/mast3r/mast3r/utils/collate.py +62 -0
- vismatch/third_party/mast3r/mast3r/utils/misc.py +17 -0
- vismatch/third_party/mast3r/mast3r/utils/path_to_dust3r.py +19 -0
- vismatch/third_party/mast3r/train.py +48 -0
- vismatch/third_party/mast3r/visloc.py +538 -0
- vismatch/third_party/omniglue/__init__.py +19 -0
- vismatch/third_party/omniglue/demo.py +89 -0
- vismatch/third_party/omniglue/src/omniglue/__init__.py +17 -0
- vismatch/third_party/omniglue/src/omniglue/dino_extract.py +215 -0
- vismatch/third_party/omniglue/src/omniglue/omniglue_extract.py +159 -0
- vismatch/third_party/omniglue/src/omniglue/superpoint_extract.py +214 -0
- vismatch/third_party/omniglue/src/omniglue/utils.py +274 -0
- vismatch/third_party/omniglue/third_party/dinov2/__init__.py +0 -0
- vismatch/third_party/omniglue/third_party/dinov2/dino.py +411 -0
- vismatch/third_party/omniglue/third_party/dinov2/dino_utils.py +341 -0
- vismatch/third_party/rdd/RDD/RDD.py +262 -0
- vismatch/third_party/rdd/RDD/RDD_helper.py +181 -0
- vismatch/third_party/rdd/RDD/dataset/__init__.py +0 -0
- vismatch/third_party/rdd/RDD/dataset/megadepth/__init__.py +2 -0
- vismatch/third_party/rdd/RDD/dataset/megadepth/megadepth.py +313 -0
- vismatch/third_party/rdd/RDD/dataset/megadepth/megadepth_warper.py +75 -0
- vismatch/third_party/rdd/RDD/dataset/megadepth/utils.py +848 -0
- vismatch/third_party/rdd/RDD/matchers/__init__.py +3 -0
- vismatch/third_party/rdd/RDD/matchers/dense_matcher.py +137 -0
- vismatch/third_party/rdd/RDD/matchers/dual_softmax_matcher.py +31 -0
- vismatch/third_party/rdd/RDD/matchers/lightglue.py +667 -0
- vismatch/third_party/rdd/RDD/models/backbone.py +147 -0
- vismatch/third_party/rdd/RDD/models/deformable_transformer.py +270 -0
- vismatch/third_party/rdd/RDD/models/descriptor.py +116 -0
- vismatch/third_party/rdd/RDD/models/detector.py +141 -0
- vismatch/third_party/rdd/RDD/models/interpolator.py +33 -0
- vismatch/third_party/rdd/RDD/models/ops/functions/__init__.py +13 -0
- vismatch/third_party/rdd/RDD/models/ops/functions/ms_deform_attn_func.py +74 -0
- vismatch/third_party/rdd/RDD/models/ops/modules/__init__.py +12 -0
- vismatch/third_party/rdd/RDD/models/ops/modules/ms_deform_attn.py +125 -0
- vismatch/third_party/rdd/RDD/models/ops/setup.py +78 -0
- vismatch/third_party/rdd/RDD/models/ops/test.py +92 -0
- vismatch/third_party/rdd/RDD/models/position_encoding.py +48 -0
- vismatch/third_party/rdd/RDD/models/soft_detect.py +176 -0
- vismatch/third_party/rdd/RDD/utils/__init__.py +1 -0
- vismatch/third_party/rdd/RDD/utils/misc.py +531 -0
- vismatch/third_party/rdd/benchmarks/air_ground.py +250 -0
- vismatch/third_party/rdd/benchmarks/mega_1500.py +259 -0
- vismatch/third_party/rdd/benchmarks/mega_view.py +252 -0
- vismatch/third_party/rdd/benchmarks/scannet_1500.py +251 -0
- vismatch/third_party/rdd/benchmarks/utils.py +112 -0
- vismatch/third_party/rdd/configs/default.yaml +19 -0
- vismatch/third_party/rdd/sfm/extract_rdd.py +145 -0
- vismatch/third_party/rdd/sfm/match_rdd.py +259 -0
- vismatch/third_party/rdd/third_party/LightGlue/.github/workflows/code-quality.yml +24 -0
- vismatch/third_party/rdd/third_party/LightGlue/benchmark.py +255 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/__init__.py +7 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/aliked.py +760 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/disk.py +55 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/dog_hardnet.py +41 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/lightglue.py +662 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/sift.py +216 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/superpoint.py +227 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/utils.py +165 -0
- vismatch/third_party/rdd/third_party/LightGlue/lightglue/viz2d.py +203 -0
- vismatch/third_party/rdd/third_party/__init__.py +1 -0
- vismatch/third_party/rdd/third_party/aliked_wrapper.py +17 -0
- vismatch/third_party/rdd/training/losses/descriptor_loss.py +73 -0
- vismatch/third_party/rdd/training/losses/detector_loss.py +499 -0
- vismatch/third_party/rdd/training/train.py +473 -0
- vismatch/third_party/rdd/training/utils.py +246 -0
- vismatch/utils.py +390 -0
- vismatch/viz.py +222 -0
- vismatch-1.1.1.dist-info/METADATA +265 -0
- vismatch-1.1.1.dist-info/RECORD +2042 -0
- vismatch-1.1.1.dist-info/WHEEL +5 -0
- vismatch-1.1.1.dist-info/entry_points.txt +4 -0
- vismatch-1.1.1.dist-info/licenses/LICENSE +28 -0
- vismatch-1.1.1.dist-info/top_level.txt +4 -0
- vismatch_extract.py +103 -0
- vismatch_match.py +114 -0
- vismatch_test.py +186 -0
|
@@ -0,0 +1,1323 @@
|
|
|
1
|
+
"""
|
|
2
|
+
UFM (Unified Flow & Matching) model for dense correspondence prediction between two views.
|
|
3
|
+
"""
|
|
4
|
+
|
|
5
|
+
import os
|
|
6
|
+
from typing import Any, Dict, List, Optional, Tuple
|
|
7
|
+
|
|
8
|
+
import torch
|
|
9
|
+
from huggingface_hub import PyTorchModelHubMixin
|
|
10
|
+
from torch import nn
|
|
11
|
+
|
|
12
|
+
# Only enable flash attention backend
|
|
13
|
+
from uniception.models.encoders import ViTEncoderInput, feature_returner_encoder_factory
|
|
14
|
+
from uniception.models.info_sharing import INFO_SHARING_CLASSES, MultiViewTransformerInput
|
|
15
|
+
from uniception.models.prediction_heads.adaptors import (
|
|
16
|
+
ConfidenceAdaptor,
|
|
17
|
+
Covariance2DAdaptor,
|
|
18
|
+
FlowAdaptor,
|
|
19
|
+
FlowWithConfidenceAdaptor,
|
|
20
|
+
MaskAdaptor,
|
|
21
|
+
)
|
|
22
|
+
from uniception.models.prediction_heads.base import AdaptorMap, PredictionHeadInput, PredictionHeadLayeredInput
|
|
23
|
+
from uniception.models.prediction_heads.dpt import DPTFeature, DPTRegressionProcessor
|
|
24
|
+
from uniception.models.prediction_heads.mlp_feature import MLPFeature
|
|
25
|
+
from uniception.models.prediction_heads.moge_conv import MoGeConvFeature
|
|
26
|
+
|
|
27
|
+
from .base import (
|
|
28
|
+
UFMClassificationRefinementOutput,
|
|
29
|
+
UFMFlowFieldOutput,
|
|
30
|
+
UFMMaskFieldOutput,
|
|
31
|
+
UFMOutputInterface,
|
|
32
|
+
UniFlowMatchModelsBase,
|
|
33
|
+
)
|
|
34
|
+
from .unet_encoder import UNet
|
|
35
|
+
from .utils import get_meshgrid_torch
|
|
36
|
+
|
|
37
|
+
CLASSNAME_TO_ADAPTOR_CLASS = {
|
|
38
|
+
"FlowWithConfidenceAdaptor": FlowWithConfidenceAdaptor,
|
|
39
|
+
"FlowAdaptor": FlowAdaptor,
|
|
40
|
+
"MaskAdaptor": MaskAdaptor,
|
|
41
|
+
"Covariance2DAdaptor": Covariance2DAdaptor,
|
|
42
|
+
"ConfidenceAdaptor": ConfidenceAdaptor,
|
|
43
|
+
}
|
|
44
|
+
|
|
45
|
+
|
|
46
|
+
def is_symmetrized(gt1, gt2):
|
|
47
|
+
"""
|
|
48
|
+
Function to check if input pairs are symmetrized, i.e., (a, b) and (b, a) always exist in the input.
|
|
49
|
+
Reference: DUSt3R
|
|
50
|
+
|
|
51
|
+
Args:
|
|
52
|
+
gt1 (Dict[str, Any]): First ground truth data containing 'instance' key
|
|
53
|
+
gt2 (Dict[str, Any]): Second ground truth data containing 'instance' key
|
|
54
|
+
|
|
55
|
+
Returns:
|
|
56
|
+
bool: True if the input pairs are symmetrized, False otherwise
|
|
57
|
+
"""
|
|
58
|
+
x = gt1["instance"]
|
|
59
|
+
y = gt2["instance"]
|
|
60
|
+
if len(x) == len(y) and len(x) == 1:
|
|
61
|
+
return False # special case of batchsize 1
|
|
62
|
+
ok = True
|
|
63
|
+
for i in range(0, len(x), 2):
|
|
64
|
+
ok = ok and (x[i] == y[i + 1]) and (x[i + 1] == y[i])
|
|
65
|
+
|
|
66
|
+
return ok
|
|
67
|
+
|
|
68
|
+
|
|
69
|
+
def interleave(tensor1, tensor2):
|
|
70
|
+
"""
|
|
71
|
+
Interleave two tensors along the first dimension (used to avoid redundant encoding for symmetrized pairs).
|
|
72
|
+
|
|
73
|
+
Args:
|
|
74
|
+
tensor1 (torch.Tensor): First input tensor
|
|
75
|
+
tensor2 (torch.Tensor): Second input tensor
|
|
76
|
+
|
|
77
|
+
Returns:
|
|
78
|
+
Tuple[torch.Tensor, torch.Tensor]: Two interleaved tensors (tensor1, tensor2) and (tensor2, tensor1)
|
|
79
|
+
"""
|
|
80
|
+
res1 = torch.stack((tensor1, tensor2), dim=1).flatten(0, 1)
|
|
81
|
+
res2 = torch.stack((tensor2, tensor1), dim=1).flatten(0, 1)
|
|
82
|
+
return res1, res2
|
|
83
|
+
|
|
84
|
+
|
|
85
|
+
def modify_state_dict(original_state_dict, mappings):
|
|
86
|
+
"""
|
|
87
|
+
Modify state dict keys according to replacement mappings
|
|
88
|
+
|
|
89
|
+
Args:
|
|
90
|
+
original_state_dict: Loaded checkpoint state dict
|
|
91
|
+
mappings: Dictionary of {old_key_substr: new_key_substr_or_None}
|
|
92
|
+
|
|
93
|
+
Returns:
|
|
94
|
+
Modified state dictionary with updated keys
|
|
95
|
+
"""
|
|
96
|
+
new_state_dict = {}
|
|
97
|
+
|
|
98
|
+
for k, v in original_state_dict.items():
|
|
99
|
+
new_key = None
|
|
100
|
+
skip = False
|
|
101
|
+
|
|
102
|
+
# Check for all possible replacements
|
|
103
|
+
for replace_key, replace_value in mappings.items():
|
|
104
|
+
if replace_key in k:
|
|
105
|
+
if replace_value is None:
|
|
106
|
+
skip = True
|
|
107
|
+
break # Skip this key entirely
|
|
108
|
+
else:
|
|
109
|
+
new_key = k.replace(replace_key, replace_value)
|
|
110
|
+
break # Only apply first matching replacement
|
|
111
|
+
|
|
112
|
+
if skip:
|
|
113
|
+
continue
|
|
114
|
+
|
|
115
|
+
new_state_dict[new_key if new_key is not None else k] = v
|
|
116
|
+
|
|
117
|
+
return new_state_dict
|
|
118
|
+
|
|
119
|
+
|
|
120
|
+
class UniFlowMatch(UniFlowMatchModelsBase, PyTorchModelHubMixin):
|
|
121
|
+
"""
|
|
122
|
+
UniFlowMatch model for dense correspondence prediction between two views.
|
|
123
|
+
|
|
124
|
+
This is the base model that performs multi-view feature extraction, information sharing,
|
|
125
|
+
and correspondence prediction between image pairs. It combines a feature encoder with
|
|
126
|
+
an information sharing module and prediction heads to output dense flow fields and
|
|
127
|
+
optional covisibility masks.
|
|
128
|
+
"""
|
|
129
|
+
|
|
130
|
+
def __init__(
|
|
131
|
+
self,
|
|
132
|
+
# Encoder configurations
|
|
133
|
+
encoder_str: str,
|
|
134
|
+
encoder_kwargs: Dict[str, Any],
|
|
135
|
+
# Info sharing & output head structure configurations
|
|
136
|
+
info_sharing_and_head_structure: str = "dual+single", # only dual+single is supported
|
|
137
|
+
# Information sharing configurations
|
|
138
|
+
info_sharing_str: str = "global_attention",
|
|
139
|
+
info_sharing_kwargs: Dict[str, Any] = {},
|
|
140
|
+
# Skip-connections between encoder and info-sharing
|
|
141
|
+
encoder_skip_connection: Optional[List[int]] = None,
|
|
142
|
+
info_sharing_skip_connection: Optional[List[int]] = None,
|
|
143
|
+
# Prediction Heads & Adaptors
|
|
144
|
+
head_type: str = "dpt",
|
|
145
|
+
feature_head_kwargs: Dict[str, Any] = {},
|
|
146
|
+
adaptors_kwargs: Dict[str, Any] = {},
|
|
147
|
+
# Load Pretrained Weights
|
|
148
|
+
pretrained_checkpoint_path: Optional[str] = None,
|
|
149
|
+
# Inference Settings
|
|
150
|
+
inference_resolution: Optional[Tuple[int, int]] = (560, 420), # (W, H)
|
|
151
|
+
*args,
|
|
152
|
+
**kwargs,
|
|
153
|
+
):
|
|
154
|
+
"""
|
|
155
|
+
Initialize the UniFlowMatch Model.
|
|
156
|
+
|
|
157
|
+
Args:
|
|
158
|
+
encoder_str (str): Encoder identifier string
|
|
159
|
+
encoder_kwargs (Dict[str, Any]): Encoder configuration parameters
|
|
160
|
+
info_sharing_and_head_structure (str): Info sharing and head structure type.
|
|
161
|
+
Currently only "dual+single" is supported (dual view info sharing and single view prediction head)
|
|
162
|
+
info_sharing_str (str): Info sharing method identifier.
|
|
163
|
+
Supported: "global_attention_transformer", "global_attention"
|
|
164
|
+
info_sharing_kwargs (Dict[str, Any]): Info sharing module configuration parameters
|
|
165
|
+
encoder_skip_connection (Optional[List[int]]): Skip connection indices from encoder
|
|
166
|
+
info_sharing_skip_connection (Optional[List[int]]): Skip connection indices from info sharing
|
|
167
|
+
head_type (str): Prediction head type. Supported: "dpt", "moge_conv"
|
|
168
|
+
feature_head_kwargs (Dict[str, Any]): Feature head configuration parameters
|
|
169
|
+
adaptors_kwargs (Dict[str, Any]): Adaptors configuration parameters
|
|
170
|
+
pretrained_checkpoint_path (Optional[str]): Path to pretrained model checkpoint
|
|
171
|
+
inference_resolution (Optional[Tuple[int, int]]): Target resolution for inference (W, H)
|
|
172
|
+
*args: Additional arguments passed to parent classes
|
|
173
|
+
**kwargs: Additional keyword arguments passed to parent classes
|
|
174
|
+
"""
|
|
175
|
+
UniFlowMatchModelsBase.__init__(self, inference_resolution=inference_resolution, *args, **kwargs)
|
|
176
|
+
|
|
177
|
+
PyTorchModelHubMixin.__init__(self)
|
|
178
|
+
|
|
179
|
+
# assertion on architectures
|
|
180
|
+
assert info_sharing_and_head_structure == "dual+single", "Only dual+single is supported now"
|
|
181
|
+
|
|
182
|
+
# initialize the skip-connections
|
|
183
|
+
self.encoder_skip_connection = encoder_skip_connection
|
|
184
|
+
self.info_sharing_skip_connection = info_sharing_skip_connection
|
|
185
|
+
|
|
186
|
+
# initialize encoder
|
|
187
|
+
self.encoder: nn.Module = feature_returner_encoder_factory(encoder_str, **encoder_kwargs)
|
|
188
|
+
|
|
189
|
+
# initialize info-sharing module
|
|
190
|
+
assert head_type != "linear", "Linear head is not supported, because it have major disadvantage to DPTs"
|
|
191
|
+
self.head_type = head_type
|
|
192
|
+
|
|
193
|
+
self.info_sharing: nn.Module = INFO_SHARING_CLASSES[info_sharing_str][1](**info_sharing_kwargs)
|
|
194
|
+
|
|
195
|
+
self.head1: nn.Module = self._initialize_prediction_heads(head_type, feature_head_kwargs, adaptors_kwargs)
|
|
196
|
+
|
|
197
|
+
# load pretrained weights
|
|
198
|
+
if pretrained_checkpoint_path is not None:
|
|
199
|
+
ckpt = torch.load(pretrained_checkpoint_path, map_location="cpu")
|
|
200
|
+
|
|
201
|
+
if "state_dict" in ckpt:
|
|
202
|
+
# we are loading from training checkpoint directly.
|
|
203
|
+
model_state_dict = ckpt["state_dict"]
|
|
204
|
+
model_state_dict = {
|
|
205
|
+
k[6:]: v for k, v in model_state_dict.items() if k.startswith("model.")
|
|
206
|
+
} # remove "model." prefix
|
|
207
|
+
|
|
208
|
+
model_state_dict = modify_state_dict(
|
|
209
|
+
model_state_dict, {"feature_matching_proj": None, "encoder.model.mask_token": None}
|
|
210
|
+
)
|
|
211
|
+
|
|
212
|
+
self.load_state_dict(model_state_dict, strict=True)
|
|
213
|
+
else:
|
|
214
|
+
model_state_dict = ckpt["model"]
|
|
215
|
+
|
|
216
|
+
load_result = self.load_state_dict(model_state_dict, strict=False)
|
|
217
|
+
assert len(load_result.missing_keys) == 0, f"Missing keys: {load_result.missing_keys}"
|
|
218
|
+
|
|
219
|
+
@classmethod
|
|
220
|
+
def from_pretrained_ckpt(cls, pretrained_model_name_or_path, strict=True, **kw):
|
|
221
|
+
"""
|
|
222
|
+
Load a model from a pretrained checkpoint file.
|
|
223
|
+
|
|
224
|
+
Args:
|
|
225
|
+
pretrained_model_name_or_path (str): Path to the pretrained model checkpoint file
|
|
226
|
+
strict (bool): Whether to strictly enforce that the keys in state_dict match exactly
|
|
227
|
+
**kw: Additional keyword arguments
|
|
228
|
+
|
|
229
|
+
Returns:
|
|
230
|
+
UniFlowMatch: Loaded model instance
|
|
231
|
+
|
|
232
|
+
Raises:
|
|
233
|
+
ValueError: If the pretrained model file is not found
|
|
234
|
+
"""
|
|
235
|
+
if os.path.isfile(pretrained_model_name_or_path):
|
|
236
|
+
ckpt = torch.load(pretrained_model_name_or_path, map_location="cpu")
|
|
237
|
+
model = cls(**ckpt["model_args"])
|
|
238
|
+
model.load_state_dict(ckpt["model"], strict=strict)
|
|
239
|
+
return model
|
|
240
|
+
else:
|
|
241
|
+
raise ValueError(f"Pretrained model {pretrained_model_name_or_path} not found.")
|
|
242
|
+
|
|
243
|
+
def _initialize_prediction_heads(
|
|
244
|
+
self, head_type: str, feature_head_kwargs: Dict[str, Any], adaptors_kwargs: Dict[str, Any]
|
|
245
|
+
):
|
|
246
|
+
"""
|
|
247
|
+
Initialize prediction heads and adaptors.
|
|
248
|
+
|
|
249
|
+
Args:
|
|
250
|
+
head_type (str): Head type, supported: "dpt", "moge_conv"
|
|
251
|
+
feature_head_kwargs (Dict[str, Any]): Feature head configuration parameters
|
|
252
|
+
adaptors_kwargs (Dict[str, Any]): Adaptors configuration parameters
|
|
253
|
+
|
|
254
|
+
Returns:
|
|
255
|
+
nn.Module: Sequential module containing feature processor and adaptor map
|
|
256
|
+
|
|
257
|
+
Raises:
|
|
258
|
+
ValueError: If unsupported head type is specified
|
|
259
|
+
"""
|
|
260
|
+
feature_processor: nn.Module
|
|
261
|
+
if head_type == "dpt":
|
|
262
|
+
feature_processor = nn.Sequential(
|
|
263
|
+
DPTFeature(**feature_head_kwargs["dpt_feature"]),
|
|
264
|
+
DPTRegressionProcessor(**feature_head_kwargs["dpt_processor"]),
|
|
265
|
+
)
|
|
266
|
+
elif head_type == "moge_conv":
|
|
267
|
+
feature_processor = MoGeConvFeature(**feature_head_kwargs)
|
|
268
|
+
else:
|
|
269
|
+
raise ValueError(f"Head type {head_type} not supported.")
|
|
270
|
+
|
|
271
|
+
adaptors = self._initialize_adaptors(adaptors_kwargs)
|
|
272
|
+
|
|
273
|
+
return nn.Sequential(feature_processor, AdaptorMap(*adaptors.values()))
|
|
274
|
+
|
|
275
|
+
def _initialize_adaptors(self, adaptors_kwargs: Dict[str, Any]):
|
|
276
|
+
"""
|
|
277
|
+
Initialize a dictionary of adaptors.
|
|
278
|
+
|
|
279
|
+
Args:
|
|
280
|
+
adaptors_kwargs (Dict[str, Any]): Adaptors configuration parameters.
|
|
281
|
+
Each key is an adaptor name, and each value contains 'class' and 'kwargs' keys.
|
|
282
|
+
|
|
283
|
+
Returns:
|
|
284
|
+
Dict[str, nn.Module]: Dictionary mapping adaptor names to their corresponding adaptor modules
|
|
285
|
+
"""
|
|
286
|
+
return {
|
|
287
|
+
name: CLASSNAME_TO_ADAPTOR_CLASS[configs["class"]](**configs["kwargs"])
|
|
288
|
+
for name, configs in adaptors_kwargs.items()
|
|
289
|
+
}
|
|
290
|
+
|
|
291
|
+
def _encode_image_pairs(self, img1, img2, data_norm_type):
|
|
292
|
+
"""
|
|
293
|
+
Encode two different batches of images (each batch can have different image shape).
|
|
294
|
+
|
|
295
|
+
Args:
|
|
296
|
+
img1 (torch.Tensor): First batch of images with shape (B, C, H, W)
|
|
297
|
+
img2 (torch.Tensor): Second batch of images with shape (B, C, H, W)
|
|
298
|
+
data_norm_type (str): Data normalization type for the encoder
|
|
299
|
+
|
|
300
|
+
Returns:
|
|
301
|
+
Tuple[List[torch.Tensor], List[torch.Tensor]]: Two lists of encoded features,
|
|
302
|
+
one for each image batch
|
|
303
|
+
|
|
304
|
+
Raises:
|
|
305
|
+
NotImplementedError: If the images have different sizes
|
|
306
|
+
"""
|
|
307
|
+
if img1.shape[-2:] == img2.shape[-2:]:
|
|
308
|
+
encoder_input = ViTEncoderInput(image=torch.cat((img1, img2), dim=0), data_norm_type=data_norm_type)
|
|
309
|
+
encoder_output = self.encoder(encoder_input)
|
|
310
|
+
out_list, out2_list = [], []
|
|
311
|
+
|
|
312
|
+
for encoder_output_ in encoder_output:
|
|
313
|
+
out, out2 = encoder_output_.features.chunk(2, dim=0)
|
|
314
|
+
out_list.append(out)
|
|
315
|
+
out2_list.append(out2)
|
|
316
|
+
else:
|
|
317
|
+
raise NotImplementedError("Unequal Image sizes are not supported now")
|
|
318
|
+
|
|
319
|
+
return out_list, out2_list
|
|
320
|
+
|
|
321
|
+
def _encode_symmetrized(self, view1, view2, symmetrized=False):
|
|
322
|
+
"""
|
|
323
|
+
Encode image pairs accounting for symmetrization, i.e., (a, b) and (b, a) always exist in the input.
|
|
324
|
+
|
|
325
|
+
Args:
|
|
326
|
+
view1 (Dict[str, Any]): First view containing 'img' and 'data_norm_type' keys
|
|
327
|
+
view2 (Dict[str, Any]): Second view containing 'img' and 'data_norm_type' keys
|
|
328
|
+
symmetrized (bool): Whether the input pairs are symmetrized
|
|
329
|
+
|
|
330
|
+
Returns:
|
|
331
|
+
Tuple[List[torch.Tensor], List[torch.Tensor]]: Encoded features for both views
|
|
332
|
+
"""
|
|
333
|
+
img1 = view1["img"]
|
|
334
|
+
img2 = view2["img"]
|
|
335
|
+
|
|
336
|
+
feat1_list, feat2_list = [], []
|
|
337
|
+
|
|
338
|
+
if symmetrized:
|
|
339
|
+
# Computing half of forward pass'
|
|
340
|
+
# modified in conjunction with UFM for not copying the images again.
|
|
341
|
+
# used to be: feat1, feat2 = self._encode_image_pairs(img1[::2], img2[::2], data_norm_type=view1["data_norm_type"])
|
|
342
|
+
# be very carefult with this!!!
|
|
343
|
+
feat1_list_, feat2_list_ = self._encode_image_pairs(
|
|
344
|
+
img1[::2], img2[::2], data_norm_type=view1["data_norm_type"]
|
|
345
|
+
)
|
|
346
|
+
|
|
347
|
+
for feat1, feat2 in zip(feat1_list_, feat2_list_):
|
|
348
|
+
feat1, feat2 = interleave(feat1, feat2)
|
|
349
|
+
feat1_list.append(feat1)
|
|
350
|
+
feat2_list.append(feat2)
|
|
351
|
+
else:
|
|
352
|
+
feat1_list, feat2_list = self._encode_image_pairs(img1, img2, data_norm_type=view1["data_norm_type"])
|
|
353
|
+
|
|
354
|
+
return feat1_list, feat2_list
|
|
355
|
+
|
|
356
|
+
def forward(self, view1, view2) -> UFMOutputInterface:
|
|
357
|
+
"""
|
|
358
|
+
Forward interface of correspondence prediction networks.
|
|
359
|
+
|
|
360
|
+
Args:
|
|
361
|
+
view1 (Dict[str, Any]): Input view 1 containing:
|
|
362
|
+
- img (torch.Tensor): BCHW image tensor normalized according to encoder's data_norm_type
|
|
363
|
+
- instance (List[int]): List of instance indices, or id of the input image
|
|
364
|
+
- data_norm_type (str): Data normalization type, see uniception.models.encoders.IMAGE_NORMALIZATION_DICT
|
|
365
|
+
- symmetrized (bool): Whether input pairs are symmetrized
|
|
366
|
+
view2 (Dict[str, Any]): Input view 2 with same structure as view1
|
|
367
|
+
|
|
368
|
+
Returns:
|
|
369
|
+
- UFMOutputInterface (Dict[str, Any]): Output results
|
|
370
|
+
- flow [Required] (Dict[str, torch.Tensor]): Flow output
|
|
371
|
+
- [Required] flow_output (torch.Tensor): Flow output tensor, BCHW
|
|
372
|
+
- [Optional] flow_covariance
|
|
373
|
+
- [Optional] flow_covariance_inv
|
|
374
|
+
- [Optional] flow_covariance_log_det
|
|
375
|
+
- covisibility [Optional] (Dict[str, torch.Tensor]): Covisibility output
|
|
376
|
+
- [Optional] mask
|
|
377
|
+
- [Optional] logits
|
|
378
|
+
"""
|
|
379
|
+
|
|
380
|
+
# Get input shapes
|
|
381
|
+
_, _, height1, width1 = view1["img"].shape
|
|
382
|
+
_, _, height2, width2 = view2["img"].shape
|
|
383
|
+
shape1 = (int(height1), int(width1))
|
|
384
|
+
shape2 = (int(height2), int(width2))
|
|
385
|
+
|
|
386
|
+
# Encode the two images --> Each feat output: BCHW features (batch_size, feature_dim, feature_height, feature_width)
|
|
387
|
+
feat1_list, feat2_list = self._encode_symmetrized(view1, view2, view1["symmetrized"])
|
|
388
|
+
|
|
389
|
+
# Pass the features through the info_sharing
|
|
390
|
+
info_sharing_input = MultiViewTransformerInput(features=[feat1_list[-1], feat2_list[-1]])
|
|
391
|
+
|
|
392
|
+
final_info_sharing_multi_view_feat, intermediate_info_sharing_multi_view_feat = self.info_sharing(
|
|
393
|
+
info_sharing_input
|
|
394
|
+
)
|
|
395
|
+
|
|
396
|
+
info_sharing_outputs = {
|
|
397
|
+
"1": [
|
|
398
|
+
feat1_list[-1].float().contiguous(),
|
|
399
|
+
intermediate_info_sharing_multi_view_feat[0].features[0].float().contiguous(),
|
|
400
|
+
intermediate_info_sharing_multi_view_feat[1].features[0].float().contiguous(),
|
|
401
|
+
final_info_sharing_multi_view_feat.features[0].float().contiguous(),
|
|
402
|
+
],
|
|
403
|
+
"2": [
|
|
404
|
+
feat2_list[-1].float().contiguous(),
|
|
405
|
+
intermediate_info_sharing_multi_view_feat[0].features[1].float().contiguous(),
|
|
406
|
+
intermediate_info_sharing_multi_view_feat[1].features[1].float().contiguous(),
|
|
407
|
+
final_info_sharing_multi_view_feat.features[1].float().contiguous(),
|
|
408
|
+
],
|
|
409
|
+
}
|
|
410
|
+
|
|
411
|
+
result = UFMOutputInterface()
|
|
412
|
+
|
|
413
|
+
# The prediction need precision, so we disable any autocasting here
|
|
414
|
+
with torch.autocast("cuda", torch.float32):
|
|
415
|
+
# run the collected info_sharing features through the prediction heads
|
|
416
|
+
head_output1 = self._downstream_head(1, info_sharing_outputs, shape1)
|
|
417
|
+
|
|
418
|
+
if "flow" in head_output1:
|
|
419
|
+
# output is flow only
|
|
420
|
+
result.flow = UFMFlowFieldOutput(flow_output=head_output1["flow"].value)
|
|
421
|
+
|
|
422
|
+
if "flow_cov" in head_output1:
|
|
423
|
+
result.flow.flow_covariance = head_output1["flow_cov"].covariance
|
|
424
|
+
result.flow.flow_covariance_inv = head_output1["flow_cov"].inv_covariance
|
|
425
|
+
result.flow.flow_covariance_log_det = head_output1["flow_cov"].log_det
|
|
426
|
+
|
|
427
|
+
if "non_occluded_mask" in head_output1:
|
|
428
|
+
result.covisibility = UFMMaskFieldOutput(
|
|
429
|
+
mask=head_output1["non_occluded_mask"].mask,
|
|
430
|
+
logits=head_output1["non_occluded_mask"].logits,
|
|
431
|
+
)
|
|
432
|
+
|
|
433
|
+
return result
|
|
434
|
+
|
|
435
|
+
def _downstream_head(self, head_num, decout, img_shape):
|
|
436
|
+
"""
|
|
437
|
+
Run the respective prediction heads.
|
|
438
|
+
|
|
439
|
+
Args:
|
|
440
|
+
head_num (Union[int, str]): Head number identifier (1)
|
|
441
|
+
decout (Dict[str, List[torch.Tensor]]): Decoded features from info sharing
|
|
442
|
+
img_shape (Tuple[int, int]): Target image shape (H, W)
|
|
443
|
+
|
|
444
|
+
Returns:
|
|
445
|
+
Dict[str, Any]: Head output containing predictions from adaptors
|
|
446
|
+
"""
|
|
447
|
+
head = getattr(self, f"head{head_num}")
|
|
448
|
+
if self.head_type == "linear":
|
|
449
|
+
head_input = PredictionHeadInput(last_feature=decout[f"{head_num}"])
|
|
450
|
+
elif self.head_type in ["dpt", "moge_conv"]:
|
|
451
|
+
head_input = PredictionHeadLayeredInput(list_features=decout[f"{head_num}"], target_output_shape=img_shape)
|
|
452
|
+
|
|
453
|
+
return head(head_input)
|
|
454
|
+
|
|
455
|
+
def get_parameter_groups(self) -> Dict[str, torch.nn.ParameterList]:
|
|
456
|
+
"""
|
|
457
|
+
Get parameter groups for optimizer. This method guides the optimizer
|
|
458
|
+
to apply correct learning rate to different parts of the model.
|
|
459
|
+
|
|
460
|
+
Returns:
|
|
461
|
+
Dict[str, torch.nn.ParameterList]: Parameter groups for optimizer with keys:
|
|
462
|
+
- "encoder": Parameters from the feature encoder
|
|
463
|
+
- "info_sharing": Parameters from the information sharing module
|
|
464
|
+
- "output_head": Parameters from the prediction head
|
|
465
|
+
"""
|
|
466
|
+
|
|
467
|
+
return {
|
|
468
|
+
"encoder": torch.nn.ParameterList(self.encoder.parameters()),
|
|
469
|
+
"info_sharing": torch.nn.ParameterList(self.info_sharing.parameters()),
|
|
470
|
+
"output_head": torch.nn.ParameterList(self.head1.parameters()),
|
|
471
|
+
}
|
|
472
|
+
|
|
473
|
+
|
|
474
|
+
class UniFlowMatchConfidence(UniFlowMatch, PyTorchModelHubMixin):
|
|
475
|
+
"""
|
|
476
|
+
UniFlowMatch model with uncertainty estimation capabilities.
|
|
477
|
+
|
|
478
|
+
This extends the base UniFlowMatch model by adding an uncertainty estimation head
|
|
479
|
+
that can predict flow covariance, keypoint confidence, and other uncertainty-related
|
|
480
|
+
outputs alongside the main flow predictions.
|
|
481
|
+
"""
|
|
482
|
+
|
|
483
|
+
def __init__(
|
|
484
|
+
self,
|
|
485
|
+
# Encoder configurations
|
|
486
|
+
encoder_str: str,
|
|
487
|
+
encoder_kwargs: Dict[str, Any],
|
|
488
|
+
# Info sharing & output head structure configurations
|
|
489
|
+
info_sharing_and_head_structure: str = "dual+single", # only dual+single is supported
|
|
490
|
+
# Information sharing configurations
|
|
491
|
+
info_sharing_str: str = "global_attention",
|
|
492
|
+
info_sharing_kwargs: Dict[str, Any] = {},
|
|
493
|
+
# Prediction Heads & Adaptors
|
|
494
|
+
head_type: str = "dpt",
|
|
495
|
+
feature_head_kwargs: Dict[str, Any] = {},
|
|
496
|
+
adaptors_kwargs: Dict[str, Any] = {},
|
|
497
|
+
# Uncertainty Heads & Adaptors
|
|
498
|
+
detach_uncertainty_head: bool = True,
|
|
499
|
+
uncertainty_head_type: str = "dpt",
|
|
500
|
+
uncertainty_head_kwargs: Dict[str, Any] = {},
|
|
501
|
+
uncertainty_adaptors_kwargs: Dict[str, Any] = {},
|
|
502
|
+
# Load Pretrained Weights
|
|
503
|
+
pretrained_backbone_checkpoint_path: Optional[str] = None,
|
|
504
|
+
pretrained_checkpoint_path: Optional[str] = None,
|
|
505
|
+
# Inference Settings
|
|
506
|
+
inference_resolution: Optional[Tuple[int, int]] = (560, 420), # WH
|
|
507
|
+
*args,
|
|
508
|
+
**kwargs,
|
|
509
|
+
):
|
|
510
|
+
"""
|
|
511
|
+
Initialize the UniFlowMatchConfidence model.
|
|
512
|
+
|
|
513
|
+
Args:
|
|
514
|
+
encoder_str (str): Encoder identifier string
|
|
515
|
+
encoder_kwargs (Dict[str, Any]): Encoder configuration parameters
|
|
516
|
+
info_sharing_and_head_structure (str): Info sharing structure, only "dual+single" supported
|
|
517
|
+
info_sharing_str (str): Info sharing method identifier
|
|
518
|
+
info_sharing_kwargs (Dict[str, Any]): Info sharing configuration parameters
|
|
519
|
+
head_type (str): Main prediction head type
|
|
520
|
+
feature_head_kwargs (Dict[str, Any]): Main feature head configuration
|
|
521
|
+
adaptors_kwargs (Dict[str, Any]): Main adaptors configuration
|
|
522
|
+
detach_uncertainty_head (bool): Whether to detach gradients for uncertainty head
|
|
523
|
+
uncertainty_head_type (str): Uncertainty head type, currently only "dpt" supported
|
|
524
|
+
uncertainty_head_kwargs (Dict[str, Any]): Uncertainty head configuration
|
|
525
|
+
uncertainty_adaptors_kwargs (Dict[str, Any]): Uncertainty adaptors configuration
|
|
526
|
+
pretrained_backbone_checkpoint_path (Optional[str]): Path to pretrained backbone checkpoint
|
|
527
|
+
pretrained_checkpoint_path (Optional[str]): Path to pretrained full model checkpoint
|
|
528
|
+
inference_resolution (Optional[Tuple[int, int]]): Target resolution for inference (W, H)
|
|
529
|
+
*args: Additional arguments passed to parent classes
|
|
530
|
+
**kwargs: Additional keyword arguments passed to parent classes
|
|
531
|
+
"""
|
|
532
|
+
UniFlowMatch.__init__(
|
|
533
|
+
self,
|
|
534
|
+
encoder_str=encoder_str,
|
|
535
|
+
encoder_kwargs=encoder_kwargs,
|
|
536
|
+
info_sharing_and_head_structure=info_sharing_and_head_structure,
|
|
537
|
+
info_sharing_str=info_sharing_str,
|
|
538
|
+
info_sharing_kwargs=info_sharing_kwargs,
|
|
539
|
+
head_type=head_type,
|
|
540
|
+
feature_head_kwargs=feature_head_kwargs,
|
|
541
|
+
adaptors_kwargs=adaptors_kwargs,
|
|
542
|
+
pretrained_checkpoint_path=pretrained_backbone_checkpoint_path,
|
|
543
|
+
inference_resolution=inference_resolution,
|
|
544
|
+
*args,
|
|
545
|
+
**kwargs,
|
|
546
|
+
)
|
|
547
|
+
|
|
548
|
+
PyTorchModelHubMixin.__init__(self)
|
|
549
|
+
|
|
550
|
+
# initialize uncertainty heads
|
|
551
|
+
assert uncertainty_head_type == "dpt", "Only DPT is supported for uncertainty head now"
|
|
552
|
+
|
|
553
|
+
self.uncertainty_head = self._initialize_prediction_heads(
|
|
554
|
+
uncertainty_head_type, uncertainty_head_kwargs, uncertainty_adaptors_kwargs
|
|
555
|
+
)
|
|
556
|
+
self.uncertainty_adaptors = self._initialize_adaptors(uncertainty_adaptors_kwargs)
|
|
557
|
+
|
|
558
|
+
assert pretrained_checkpoint_path is None, "Pretrained weights are not supported for now"
|
|
559
|
+
|
|
560
|
+
self.detach_uncertainty_head = detach_uncertainty_head
|
|
561
|
+
|
|
562
|
+
def forward(self, view1, view2) -> UFMOutputInterface:
|
|
563
|
+
"""
|
|
564
|
+
Forward interface of correspondence prediction networks with uncertainty estimation.
|
|
565
|
+
|
|
566
|
+
Args:
|
|
567
|
+
view1 (Dict[str, Any]): Input view 1 containing:
|
|
568
|
+
- img (torch.Tensor): BCHW image tensor normalized according to encoder's data_norm_type
|
|
569
|
+
- instance (List[int]): List of instance indices, or id of the input image
|
|
570
|
+
- data_norm_type (str): Data normalization type
|
|
571
|
+
- symmetrized (bool): Whether input pairs are symmetrized
|
|
572
|
+
view2 (Dict[str, Any]): Input view 2 with same structure as view1
|
|
573
|
+
|
|
574
|
+
Returns:
|
|
575
|
+
- UFMOutputInterace (Dict[str, Any]): Output results
|
|
576
|
+
- flow [Required] (Dict[str, torch.Tensor]): Flow output
|
|
577
|
+
- [Required] flow_output (torch.Tensor): Flow output tensor, BCHW
|
|
578
|
+
- [Optional] flow_covariance
|
|
579
|
+
- [Optional] flow_covariance_inv
|
|
580
|
+
- [Optional] flow_covariance_log_det
|
|
581
|
+
- covisibility [Optional] (Dict[str, torch.Tensor]): Covisibiltiy output
|
|
582
|
+
- [Optional] mask
|
|
583
|
+
- [Optional] logits
|
|
584
|
+
"""
|
|
585
|
+
|
|
586
|
+
# Get input shapes
|
|
587
|
+
_, _, height1, width1 = view1["img"].shape
|
|
588
|
+
_, _, height2, width2 = view2["img"].shape
|
|
589
|
+
shape1 = (int(height1), int(width1))
|
|
590
|
+
shape2 = (int(height2), int(width2))
|
|
591
|
+
|
|
592
|
+
# Encode the two images --> Each feat output: BCHW features (batch_size, feature_dim, feature_height, feature_width)
|
|
593
|
+
feat1_list, feat2_list = self._encode_symmetrized(view1, view2, view1["symmetrized"])
|
|
594
|
+
|
|
595
|
+
# Pass the features through the info_sharing
|
|
596
|
+
info_sharing_input = MultiViewTransformerInput(features=[feat1_list[-1], feat2_list[-1]])
|
|
597
|
+
|
|
598
|
+
final_info_sharing_multi_view_feat, intermediate_info_sharing_multi_view_feat = self.info_sharing(
|
|
599
|
+
info_sharing_input
|
|
600
|
+
)
|
|
601
|
+
|
|
602
|
+
info_sharing_outputs = {
|
|
603
|
+
"1": [
|
|
604
|
+
feat1_list[-1].float().contiguous(),
|
|
605
|
+
intermediate_info_sharing_multi_view_feat[0].features[0].float().contiguous(),
|
|
606
|
+
intermediate_info_sharing_multi_view_feat[1].features[0].float().contiguous(),
|
|
607
|
+
final_info_sharing_multi_view_feat.features[0].float().contiguous(),
|
|
608
|
+
],
|
|
609
|
+
"2": [
|
|
610
|
+
feat2_list[-1].float().contiguous(),
|
|
611
|
+
intermediate_info_sharing_multi_view_feat[0].features[1].float().contiguous(),
|
|
612
|
+
intermediate_info_sharing_multi_view_feat[1].features[1].float().contiguous(),
|
|
613
|
+
final_info_sharing_multi_view_feat.features[1].float().contiguous(),
|
|
614
|
+
],
|
|
615
|
+
}
|
|
616
|
+
|
|
617
|
+
info_sharing_outputs_detached = {
|
|
618
|
+
"1": [
|
|
619
|
+
feat1_list[-1].detach().float().contiguous(),
|
|
620
|
+
intermediate_info_sharing_multi_view_feat[0].features[0].detach().float().contiguous(),
|
|
621
|
+
intermediate_info_sharing_multi_view_feat[1].features[0].detach().float().contiguous(),
|
|
622
|
+
final_info_sharing_multi_view_feat.features[0].detach().float().contiguous(),
|
|
623
|
+
],
|
|
624
|
+
"2": [
|
|
625
|
+
feat2_list[-1].detach().float().contiguous(),
|
|
626
|
+
intermediate_info_sharing_multi_view_feat[0].features[1].detach().float().contiguous(),
|
|
627
|
+
intermediate_info_sharing_multi_view_feat[1].features[1].detach().float().contiguous(),
|
|
628
|
+
final_info_sharing_multi_view_feat.features[1].detach().float().contiguous(),
|
|
629
|
+
],
|
|
630
|
+
}
|
|
631
|
+
|
|
632
|
+
result = UFMOutputInterface()
|
|
633
|
+
|
|
634
|
+
# The prediction need precision, so we disable any autocasting here
|
|
635
|
+
with torch.autocast("cuda", torch.float32):
|
|
636
|
+
# run the collected info_sharing features through the prediction heads
|
|
637
|
+
head_output1 = self._downstream_head(1, info_sharing_outputs, shape1)
|
|
638
|
+
head_output_uncertainty = self._downstream_head(
|
|
639
|
+
"uncertainty",
|
|
640
|
+
info_sharing_outputs_detached if self.detach_uncertainty_head else info_sharing_outputs,
|
|
641
|
+
shape1,
|
|
642
|
+
)
|
|
643
|
+
|
|
644
|
+
result.flow = UFMFlowFieldOutput(
|
|
645
|
+
flow_output=head_output1["flow"].value,
|
|
646
|
+
)
|
|
647
|
+
|
|
648
|
+
if "flow_cov" in head_output_uncertainty:
|
|
649
|
+
result.flow.flow_covariance = head_output_uncertainty["flow_cov"].covariance
|
|
650
|
+
result.flow.flow_covariance_inv = head_output_uncertainty["flow_cov"].inv_covariance
|
|
651
|
+
result.flow.flow_covariance_log_det = head_output_uncertainty["flow_cov"].log_det
|
|
652
|
+
|
|
653
|
+
if "keypoint_confidence" in head_output_uncertainty:
|
|
654
|
+
result.keypoint_confidence = head_output_uncertainty["keypoint_confidence"].value.squeeze(1)
|
|
655
|
+
|
|
656
|
+
if "non_occluded_mask" in head_output_uncertainty:
|
|
657
|
+
result.covisibility = UFMMaskFieldOutput(
|
|
658
|
+
mask=head_output_uncertainty["non_occluded_mask"].mask,
|
|
659
|
+
logits=head_output_uncertainty["non_occluded_mask"].logits,
|
|
660
|
+
)
|
|
661
|
+
|
|
662
|
+
return result
|
|
663
|
+
|
|
664
|
+
def get_parameter_groups(self) -> Dict[str, torch.nn.ParameterList]:
|
|
665
|
+
"""
|
|
666
|
+
Get parameter groups for optimizer. This method guides the optimizer
|
|
667
|
+
to apply correct learning rate to different parts of the model.
|
|
668
|
+
|
|
669
|
+
Returns:
|
|
670
|
+
Dict[str, torch.nn.ParameterList]: Parameter groups for optimizer with keys:
|
|
671
|
+
- "encoder": Parameters from the feature encoder
|
|
672
|
+
- "info_sharing": Parameters from the information sharing module
|
|
673
|
+
- "output_head": Parameters from the main prediction head
|
|
674
|
+
- "uncertainty_head": Parameters from the uncertainty estimation head
|
|
675
|
+
"""
|
|
676
|
+
|
|
677
|
+
return {
|
|
678
|
+
"encoder": torch.nn.ParameterList(self.encoder.parameters()),
|
|
679
|
+
"info_sharing": torch.nn.ParameterList(self.info_sharing.parameters()),
|
|
680
|
+
"output_head": torch.nn.ParameterList(self.head1.parameters()),
|
|
681
|
+
"uncertainty_head": torch.nn.ParameterList(self.uncertainty_head.parameters()),
|
|
682
|
+
}
|
|
683
|
+
|
|
684
|
+
def _downstream_head(self, head_num, decout, img_shape):
|
|
685
|
+
"""
|
|
686
|
+
Run the respective prediction heads (main or uncertainty).
|
|
687
|
+
|
|
688
|
+
Args:
|
|
689
|
+
head_num (Union[int, str]): Head number identifier (1) or "uncertainty" for uncertainty head
|
|
690
|
+
decout (Dict[str, List[torch.Tensor]]): Decoded features from info sharing
|
|
691
|
+
img_shape (Tuple[int, int]): Target image shape (H, W)
|
|
692
|
+
|
|
693
|
+
Returns:
|
|
694
|
+
Dict[str, Any]: Head output containing predictions from adaptors
|
|
695
|
+
"""
|
|
696
|
+
# if self.info_sharing_and_head_structure == "dual+single":
|
|
697
|
+
|
|
698
|
+
head = getattr(self, f"head{head_num}") if head_num != "uncertainty" else self.uncertainty_head
|
|
699
|
+
|
|
700
|
+
head_num = head_num if head_num != "uncertainty" else 1 # uncertainty head is always from branch 1
|
|
701
|
+
|
|
702
|
+
if self.head_type == "linear":
|
|
703
|
+
head_input = PredictionHeadInput(last_feature=decout[f"{head_num}"])
|
|
704
|
+
elif self.head_type in ["dpt", "moge_conv"]:
|
|
705
|
+
head_input = PredictionHeadLayeredInput(list_features=decout[f"{head_num}"], target_output_shape=img_shape)
|
|
706
|
+
|
|
707
|
+
return head(head_input)
|
|
708
|
+
|
|
709
|
+
|
|
710
|
+
class UniFlowMatchClassificationRefinement(UniFlowMatch, PyTorchModelHubMixin):
|
|
711
|
+
"""
|
|
712
|
+
The variant of UniFlowMatch with local classification-based refinement.
|
|
713
|
+
|
|
714
|
+
This model extends UniFlowMatch by adding a classification head that performs
|
|
715
|
+
local refinement of flow predictions using attention-based mechanisms. It can
|
|
716
|
+
also optionally include uncertainty estimation and U-Net features for enhanced
|
|
717
|
+
performance.
|
|
718
|
+
"""
|
|
719
|
+
|
|
720
|
+
def __init__(
|
|
721
|
+
self,
|
|
722
|
+
# Encoder configurations
|
|
723
|
+
encoder_str: str,
|
|
724
|
+
encoder_kwargs: Dict[str, Any],
|
|
725
|
+
# Info sharing & output head structure configurations
|
|
726
|
+
info_sharing_and_head_structure: str = "dual+single", # only dual+single is supported
|
|
727
|
+
# Information sharing configurations
|
|
728
|
+
info_sharing_str: str = "global_attention",
|
|
729
|
+
info_sharing_kwargs: Dict[str, Any] = {},
|
|
730
|
+
# Prediction Heads & Adaptors
|
|
731
|
+
head_type: str = "dpt",
|
|
732
|
+
feature_head_kwargs: Dict[str, Any] = {},
|
|
733
|
+
adaptors_kwargs: Dict[str, Any] = {},
|
|
734
|
+
# Uncertainty Heads & Adaptors
|
|
735
|
+
detach_uncertainty_head: bool = True,
|
|
736
|
+
uncertainty_head_type: str = "dpt",
|
|
737
|
+
uncertainty_head_kwargs: Dict[str, Any] = {},
|
|
738
|
+
uncertainty_adaptors_kwargs: Dict[str, Any] = {},
|
|
739
|
+
# Classification Heads & Adaptors
|
|
740
|
+
temperature: float = 4,
|
|
741
|
+
use_unet_feature: bool = False,
|
|
742
|
+
use_unet_batchnorm: bool = False,
|
|
743
|
+
classification_head_type: str = "patch_mlp",
|
|
744
|
+
classification_head_kwargs: Dict[str, Any] = {},
|
|
745
|
+
feature_combine_method: str = "conv",
|
|
746
|
+
# Refinement Range
|
|
747
|
+
refinement_range: int = 5,
|
|
748
|
+
# Load Pretrained Weights
|
|
749
|
+
pretrained_backbone_checkpoint_path: Optional[str] = None,
|
|
750
|
+
pretrained_checkpoint_path: Optional[str] = None,
|
|
751
|
+
# Inference Settings
|
|
752
|
+
inference_resolution: Optional[Tuple[int, int]] = (560, 420), # WH
|
|
753
|
+
*args,
|
|
754
|
+
**kwargs,
|
|
755
|
+
):
|
|
756
|
+
"""
|
|
757
|
+
Initialize the UniFlowMatchClassificationRefinement model.
|
|
758
|
+
|
|
759
|
+
Args:
|
|
760
|
+
encoder_str (str): Encoder identifier string
|
|
761
|
+
encoder_kwargs (Dict[str, Any]): Encoder configuration parameters
|
|
762
|
+
info_sharing_and_head_structure (str): Info sharing structure, only "dual+single" supported
|
|
763
|
+
info_sharing_str (str): Info sharing method identifier
|
|
764
|
+
info_sharing_kwargs (Dict[str, Any]): Info sharing configuration parameters
|
|
765
|
+
head_type (str): Main prediction head type
|
|
766
|
+
feature_head_kwargs (Dict[str, Any]): Main feature head configuration
|
|
767
|
+
adaptors_kwargs (Dict[str, Any]): Main adaptors configuration
|
|
768
|
+
detach_uncertainty_head (bool): Whether to detach gradients for uncertainty head
|
|
769
|
+
uncertainty_head_type (str): Uncertainty head type, currently only "dpt" supported
|
|
770
|
+
uncertainty_head_kwargs (Dict[str, Any]): Uncertainty head configuration
|
|
771
|
+
uncertainty_adaptors_kwargs (Dict[str, Any]): Uncertainty adaptors configuration
|
|
772
|
+
temperature (float): Temperature parameter for attention softmax in classification refinement
|
|
773
|
+
use_unet_feature (bool): Whether to use additional U-Net features for refinement
|
|
774
|
+
use_unet_batchnorm (bool): Whether to use batch normalization in U-Net, helps with convergence
|
|
775
|
+
classification_head_type (str): Classification head type, currently only "patch_mlp" supported
|
|
776
|
+
classification_head_kwargs (Dict[str, Any]): Classification head configuration
|
|
777
|
+
feature_combine_method (str): Method for combining features ("conv" or "modulate")
|
|
778
|
+
refinement_range (int): Local patch size for refinement (must be odd)
|
|
779
|
+
pretrained_backbone_checkpoint_path (Optional[str]): Path to pretrained backbone checkpoint
|
|
780
|
+
pretrained_checkpoint_path (Optional[str]): Path to pretrained full model checkpoint
|
|
781
|
+
inference_resolution (Optional[Tuple[int, int]]): Target resolution for inference (W, H)
|
|
782
|
+
*args: Additional arguments passed to parent classes
|
|
783
|
+
**kwargs: Additional keyword arguments passed to parent classes
|
|
784
|
+
"""
|
|
785
|
+
UniFlowMatch.__init__(
|
|
786
|
+
self,
|
|
787
|
+
encoder_str=encoder_str,
|
|
788
|
+
encoder_kwargs=encoder_kwargs,
|
|
789
|
+
info_sharing_and_head_structure=info_sharing_and_head_structure,
|
|
790
|
+
info_sharing_str=info_sharing_str,
|
|
791
|
+
info_sharing_kwargs=info_sharing_kwargs,
|
|
792
|
+
head_type=head_type,
|
|
793
|
+
feature_head_kwargs=feature_head_kwargs,
|
|
794
|
+
adaptors_kwargs=adaptors_kwargs,
|
|
795
|
+
pretrained_checkpoint_path=pretrained_backbone_checkpoint_path,
|
|
796
|
+
inference_resolution=inference_resolution,
|
|
797
|
+
*args,
|
|
798
|
+
**kwargs,
|
|
799
|
+
)
|
|
800
|
+
|
|
801
|
+
PyTorchModelHubMixin.__init__(self)
|
|
802
|
+
|
|
803
|
+
# initialize uncertainty heads
|
|
804
|
+
assert classification_head_type == "patch_mlp", "Only DPT is supported for uncertainty head now"
|
|
805
|
+
self.classification_head_type = classification_head_type
|
|
806
|
+
|
|
807
|
+
self.classification_head = self._initialize_classification_head(classification_head_kwargs)
|
|
808
|
+
|
|
809
|
+
self.refinement_range = refinement_range
|
|
810
|
+
self.temperature = temperature
|
|
811
|
+
|
|
812
|
+
assert pretrained_checkpoint_path is None, "Pretrained weights are not supported for now"
|
|
813
|
+
|
|
814
|
+
self.use_unet_feature = use_unet_feature
|
|
815
|
+
|
|
816
|
+
self.feature_combine_method = feature_combine_method
|
|
817
|
+
|
|
818
|
+
# Unet experiment
|
|
819
|
+
if self.use_unet_feature:
|
|
820
|
+
self.use_unet_batchnorm = use_unet_batchnorm
|
|
821
|
+
self.unet_feature = UNet(
|
|
822
|
+
in_channels=3, out_channels=16, features=[64, 128, 256, 512], use_batch_norm=self.use_unet_batchnorm
|
|
823
|
+
)
|
|
824
|
+
|
|
825
|
+
self.conv1 = nn.Conv2d(32, 32, kernel_size=1, stride=1, padding=0)
|
|
826
|
+
|
|
827
|
+
if self.feature_combine_method == "conv":
|
|
828
|
+
self.conv2 = nn.Conv2d(32, 16, kernel_size=1, stride=1, padding=0)
|
|
829
|
+
elif self.feature_combine_method == "modulate":
|
|
830
|
+
self.conv2 = nn.Conv2d(16, 16, kernel_size=1, stride=1, padding=0)
|
|
831
|
+
|
|
832
|
+
default_attention_bias = torch.zeros(self.refinement_range * self.refinement_range)
|
|
833
|
+
self.classification_bias = nn.Parameter(default_attention_bias)
|
|
834
|
+
|
|
835
|
+
# initialize uncertainty heads
|
|
836
|
+
if len(uncertainty_head_kwargs) > 0:
|
|
837
|
+
assert uncertainty_head_type == "dpt", "Only DPT is supported for uncertainty head now"
|
|
838
|
+
|
|
839
|
+
self.uncertainty_head = self._initialize_prediction_heads(
|
|
840
|
+
uncertainty_head_type, uncertainty_head_kwargs, uncertainty_adaptors_kwargs
|
|
841
|
+
)
|
|
842
|
+
self.uncertainty_adaptors = self._initialize_adaptors(uncertainty_adaptors_kwargs)
|
|
843
|
+
|
|
844
|
+
assert pretrained_checkpoint_path is None, "Pretrained weights are not supported for now"
|
|
845
|
+
|
|
846
|
+
self.detach_uncertainty_head = detach_uncertainty_head
|
|
847
|
+
|
|
848
|
+
def forward(self, view1, view2) -> UFMOutputInterface:
|
|
849
|
+
"""
|
|
850
|
+
Forward interface of correspondence prediction networks with classification-based refinement.
|
|
851
|
+
|
|
852
|
+
Args:
|
|
853
|
+
view1 (Dict[str, Any]): Input view 1 containing:
|
|
854
|
+
- img (torch.Tensor): BCHW image tensor normalized according to encoder's data_norm_type
|
|
855
|
+
- instance (List[int]): List of instance indices, or id of the input image
|
|
856
|
+
- data_norm_type (str): Data normalization type
|
|
857
|
+
- symmetrized (bool): Whether input pairs are symmetrized
|
|
858
|
+
view2 (Dict[str, Any]): Input view 2 with same structure as view1
|
|
859
|
+
|
|
860
|
+
Returns:
|
|
861
|
+
- Dict[str, Any]: Output results
|
|
862
|
+
- flow [Required] (Dict[str, torch.Tensor]): Flow output
|
|
863
|
+
- [Required] flow_output (torch.Tensor): Flow output tensor, BCHW
|
|
864
|
+
- [Optional] flow_covariance
|
|
865
|
+
- [Optional] flow_covariance_inv
|
|
866
|
+
- [Optional] flow_covariance_log_det
|
|
867
|
+
- covisibility [Optional] (Dict[str, torch.Tensor]): Covisibility output
|
|
868
|
+
- [Optional] mask
|
|
869
|
+
- [Optional] logits
|
|
870
|
+
- classification [Optional]: Probability and targets of the classification head
|
|
871
|
+
"""
|
|
872
|
+
|
|
873
|
+
# Get input shapes
|
|
874
|
+
_, _, height1, width1 = view1["img"].shape
|
|
875
|
+
_, _, height2, width2 = view2["img"].shape
|
|
876
|
+
shape1 = (int(height1), int(width1))
|
|
877
|
+
shape2 = (int(height2), int(width2))
|
|
878
|
+
|
|
879
|
+
# Encode the two images --> Each feat output: BCHW features (batch_size, feature_dim, feature_height, feature_width)
|
|
880
|
+
feat1_list, feat2_list = self._encode_symmetrized(view1, view2, view1["symmetrized"])
|
|
881
|
+
|
|
882
|
+
# Pass the features through the info_sharing
|
|
883
|
+
info_sharing_input = MultiViewTransformerInput(features=[feat1_list[-1], feat2_list[-1]])
|
|
884
|
+
|
|
885
|
+
final_info_sharing_multi_view_feat, intermediate_info_sharing_multi_view_feat = self.info_sharing(
|
|
886
|
+
info_sharing_input
|
|
887
|
+
)
|
|
888
|
+
|
|
889
|
+
info_sharing_outputs = {
|
|
890
|
+
"1": [
|
|
891
|
+
feat1_list[-1].float().contiguous(),
|
|
892
|
+
intermediate_info_sharing_multi_view_feat[0].features[0].float().contiguous(),
|
|
893
|
+
intermediate_info_sharing_multi_view_feat[1].features[0].float().contiguous(),
|
|
894
|
+
final_info_sharing_multi_view_feat.features[0].float().contiguous(),
|
|
895
|
+
],
|
|
896
|
+
"2": [
|
|
897
|
+
feat2_list[-1].float().contiguous(),
|
|
898
|
+
intermediate_info_sharing_multi_view_feat[0].features[1].float().contiguous(),
|
|
899
|
+
intermediate_info_sharing_multi_view_feat[1].features[1].float().contiguous(),
|
|
900
|
+
final_info_sharing_multi_view_feat.features[1].float().contiguous(),
|
|
901
|
+
],
|
|
902
|
+
}
|
|
903
|
+
|
|
904
|
+
info_sharing_outputs_detached = {
|
|
905
|
+
"1": [
|
|
906
|
+
feat1_list[-1].detach().float().contiguous(),
|
|
907
|
+
intermediate_info_sharing_multi_view_feat[0].features[0].detach().float().contiguous(),
|
|
908
|
+
intermediate_info_sharing_multi_view_feat[1].features[0].detach().float().contiguous(),
|
|
909
|
+
final_info_sharing_multi_view_feat.features[0].detach().float().contiguous(),
|
|
910
|
+
],
|
|
911
|
+
"2": [
|
|
912
|
+
feat2_list[-1].detach().float().contiguous(),
|
|
913
|
+
intermediate_info_sharing_multi_view_feat[0].features[1].detach().float().contiguous(),
|
|
914
|
+
intermediate_info_sharing_multi_view_feat[1].features[1].detach().float().contiguous(),
|
|
915
|
+
final_info_sharing_multi_view_feat.features[1].detach().float().contiguous(),
|
|
916
|
+
],
|
|
917
|
+
}
|
|
918
|
+
|
|
919
|
+
# optionally inference for U-Net Features
|
|
920
|
+
if self.use_unet_feature:
|
|
921
|
+
unet_feat1 = self.unet_feature(view1["img"])
|
|
922
|
+
unet_feat2 = self.unet_feature(view2["img"])
|
|
923
|
+
|
|
924
|
+
result = UFMOutputInterface()
|
|
925
|
+
# The prediction need precision, so we disable any autocasting here
|
|
926
|
+
with torch.autocast("cuda", torch.float32):
|
|
927
|
+
# run the collected info_sharing features through the prediction heads
|
|
928
|
+
head_output1 = self._downstream_head(1, info_sharing_outputs, shape1)
|
|
929
|
+
flow_prediction = head_output1["flow"].value
|
|
930
|
+
|
|
931
|
+
if hasattr(self, "uncertainty_head"):
|
|
932
|
+
# run the uncertainty head
|
|
933
|
+
head_output_uncertainty = self._downstream_head(
|
|
934
|
+
"uncertainty",
|
|
935
|
+
info_sharing_outputs_detached if self.detach_uncertainty_head else info_sharing_outputs,
|
|
936
|
+
shape1,
|
|
937
|
+
)
|
|
938
|
+
|
|
939
|
+
if "flow_cov" in head_output_uncertainty:
|
|
940
|
+
result.flow.flow_covariance = head_output_uncertainty["flow_cov"].covariance
|
|
941
|
+
result.flow.flow_covariance_inv = head_output_uncertainty["flow_cov"].inv_covariance
|
|
942
|
+
result.flow.flow_covariance_log_det = head_output_uncertainty["flow_cov"].log_det
|
|
943
|
+
|
|
944
|
+
if "keypoint_confidence" in head_output_uncertainty:
|
|
945
|
+
result.keypoint_confidence = head_output_uncertainty["keypoint_confidence"].value.squeeze(1)
|
|
946
|
+
|
|
947
|
+
if "non_occluded_mask" in head_output_uncertainty:
|
|
948
|
+
result.covisibility = UFMMaskFieldOutput(
|
|
949
|
+
mask=head_output_uncertainty["non_occluded_mask"].mask,
|
|
950
|
+
logits=head_output_uncertainty["non_occluded_mask"].logits,
|
|
951
|
+
)
|
|
952
|
+
|
|
953
|
+
# we run the classification head in the autocast environment bacause it is not regression
|
|
954
|
+
if self.classification_head_type == "patch_mlp":
|
|
955
|
+
# concatenate the last encoder feature with final info_sharing feature
|
|
956
|
+
|
|
957
|
+
# use the first encoder feature, because it captures more low-level information, which is needed
|
|
958
|
+
# for refinement of the regressed flow.
|
|
959
|
+
classification_feat_1 = torch.cat(
|
|
960
|
+
[feat1_list[0].float().contiguous(), info_sharing_outputs["1"][-1]], dim=1
|
|
961
|
+
)
|
|
962
|
+
classification_feat_2 = torch.cat(
|
|
963
|
+
[feat2_list[0].float().contiguous(), info_sharing_outputs["2"][-1]], dim=1
|
|
964
|
+
)
|
|
965
|
+
|
|
966
|
+
classification_input = PredictionHeadInput(
|
|
967
|
+
torch.cat([classification_feat_1, classification_feat_2], dim=0)
|
|
968
|
+
)
|
|
969
|
+
|
|
970
|
+
classification_features = self.classification_head(classification_input).decoded_channels
|
|
971
|
+
|
|
972
|
+
if self.use_unet_feature:
|
|
973
|
+
if self.feature_combine_method == "conv":
|
|
974
|
+
combined_features = torch.cat(
|
|
975
|
+
[classification_features, torch.cat([unet_feat1, unet_feat2], dim=0)], dim=1
|
|
976
|
+
)
|
|
977
|
+
|
|
978
|
+
combined_features = self.conv1(combined_features)
|
|
979
|
+
combined_features = nn.functional.relu(combined_features)
|
|
980
|
+
combined_features = self.conv2(combined_features)
|
|
981
|
+
elif self.feature_combine_method == "modulate":
|
|
982
|
+
combined_features = classification_features * torch.tanh(
|
|
983
|
+
torch.cat([unet_feat1, unet_feat2], dim=0)
|
|
984
|
+
)
|
|
985
|
+
combined_features = self.conv2(combined_features)
|
|
986
|
+
|
|
987
|
+
classification_features = combined_features
|
|
988
|
+
|
|
989
|
+
classification_features0, classification_features1 = classification_features.chunk(2, dim=0)
|
|
990
|
+
|
|
991
|
+
# refine the flow prediction with features from the classification head
|
|
992
|
+
for i in range(1):
|
|
993
|
+
residual, log_softmax_attention = self.classification_refinement(
|
|
994
|
+
flow_prediction, classification_features
|
|
995
|
+
)
|
|
996
|
+
flow_prediction = flow_prediction + residual
|
|
997
|
+
|
|
998
|
+
# Fill in the result
|
|
999
|
+
# WARNING: based on how the residual is computed, flow_prediction will have gradient cancelled by mathematics,
|
|
1000
|
+
# so there will be no supervision to the flow prediction at all. We need to use specialized loss function to
|
|
1001
|
+
# supervise the regression_flow_output.
|
|
1002
|
+
result.flow = UFMFlowFieldOutput(
|
|
1003
|
+
flow_output=flow_prediction,
|
|
1004
|
+
)
|
|
1005
|
+
|
|
1006
|
+
result.classification_refinement = UFMClassificationRefinementOutput(
|
|
1007
|
+
regression_flow_output=flow_prediction,
|
|
1008
|
+
residual=residual,
|
|
1009
|
+
log_softmax=log_softmax_attention,
|
|
1010
|
+
feature_map_0=classification_features0,
|
|
1011
|
+
feature_map_1=classification_features1,
|
|
1012
|
+
)
|
|
1013
|
+
|
|
1014
|
+
return result
|
|
1015
|
+
|
|
1016
|
+
# @torch.compile()
|
|
1017
|
+
def classification_refinement(self, flow_prediction, classification_features) -> Tuple[torch.Tensor, torch.Tensor]:
|
|
1018
|
+
"""
|
|
1019
|
+
Use correlation between self feature and features around a local patch of the initial flow prediction
|
|
1020
|
+
to refine the flow prediction using attention-based mechanisms.
|
|
1021
|
+
|
|
1022
|
+
Args:
|
|
1023
|
+
flow_prediction (torch.Tensor): Initial flow prediction tensor with shape (B, 2, H, W)
|
|
1024
|
+
classification_features (torch.Tensor): Classification features tensor with shape (2*B, C, H, W)
|
|
1025
|
+
where the first B samples correspond to view1 and the second B samples to view2
|
|
1026
|
+
|
|
1027
|
+
Returns:
|
|
1028
|
+
Tuple[torch.Tensor, torch.Tensor]: A tuple containing:
|
|
1029
|
+
- residual (torch.Tensor): Flow residual to be added to initial prediction, shape (B, 2, H, W)
|
|
1030
|
+
- log_softmax_attention (torch.Tensor): Log softmax attention weights, shape (B, H, W, P, P)
|
|
1031
|
+
where P is the refinement_range
|
|
1032
|
+
"""
|
|
1033
|
+
|
|
1034
|
+
classification_features1, classification_features2 = classification_features.chunk(2, dim=0)
|
|
1035
|
+
|
|
1036
|
+
neighborhood_features, neighborhood_flow_residual = self.obtain_neighborhood_features(
|
|
1037
|
+
flow_estimation=flow_prediction, other_features=classification_features2, local_patch=self.refinement_range
|
|
1038
|
+
)
|
|
1039
|
+
|
|
1040
|
+
residual, log_softmax_attention = self.compute_refinement_attention(
|
|
1041
|
+
classification_features1, neighborhood_features, neighborhood_flow_residual
|
|
1042
|
+
)
|
|
1043
|
+
|
|
1044
|
+
return residual, log_softmax_attention
|
|
1045
|
+
|
|
1046
|
+
def compute_refinement_attention(self, classification_features1, neighborhood_features, neighborhood_flow_residual):
|
|
1047
|
+
"""
|
|
1048
|
+
Compute the attention weights for flow refinement using feature correlation.
|
|
1049
|
+
|
|
1050
|
+
This method performs attention computation between query features (classification_features1)
|
|
1051
|
+
and key features (neighborhood_features) to generate attention weights for flow residual values.
|
|
1052
|
+
|
|
1053
|
+
Args:
|
|
1054
|
+
classification_features1 (torch.Tensor): Query features from view1 with shape (B, C, H, W)
|
|
1055
|
+
neighborhood_features (torch.Tensor): Key features from neighborhood sampling with shape (B, H, W, P, P, C)
|
|
1056
|
+
where P is the refinement_range
|
|
1057
|
+
neighborhood_flow_residual (torch.Tensor): Flow residual values for each neighborhood position
|
|
1058
|
+
with shape (B, H, W, P, P, 2)
|
|
1059
|
+
|
|
1060
|
+
Returns:
|
|
1061
|
+
Tuple[torch.Tensor, torch.Tensor]: A tuple containing:
|
|
1062
|
+
- residual (torch.Tensor): Weighted flow residual with shape (B, 2, H, W)
|
|
1063
|
+
- log_softmax_attention (torch.Tensor): Log softmax attention weights with shape (B, H, W, P, P)
|
|
1064
|
+
"""
|
|
1065
|
+
|
|
1066
|
+
B, C, H, W = classification_features1.shape
|
|
1067
|
+
P = self.refinement_range
|
|
1068
|
+
|
|
1069
|
+
# reshape Q to B, H, W, 1, 1, C
|
|
1070
|
+
classification_features1 = classification_features1.permute(0, 2, 3, 1).reshape(B * H * W, 1, C)
|
|
1071
|
+
|
|
1072
|
+
# reshape K to B, H, W, 1, P^2, C
|
|
1073
|
+
assert neighborhood_features.shape[0] == B
|
|
1074
|
+
assert neighborhood_features.shape[1] == H
|
|
1075
|
+
assert neighborhood_features.shape[2] == W
|
|
1076
|
+
assert neighborhood_features.shape[3] == P
|
|
1077
|
+
assert neighborhood_features.shape[4] == P
|
|
1078
|
+
assert neighborhood_features.shape[5] == C
|
|
1079
|
+
|
|
1080
|
+
neighborhood_features = neighborhood_features.reshape(B * H * W, P * P, C)
|
|
1081
|
+
|
|
1082
|
+
# reshape V to B, H, W, 1, P^2, 2
|
|
1083
|
+
neighborhood_flow_residual = neighborhood_flow_residual.reshape(-1, P * P, 2)
|
|
1084
|
+
|
|
1085
|
+
# compute the attention
|
|
1086
|
+
attention_score = (
|
|
1087
|
+
torch.matmul(classification_features1, neighborhood_features.permute(0, 2, 1)) / self.temperature
|
|
1088
|
+
)
|
|
1089
|
+
attention_score = attention_score + self.classification_bias
|
|
1090
|
+
|
|
1091
|
+
attention = torch.nn.functional.softmax(attention_score, dim=-1)
|
|
1092
|
+
log_softmax_attention = torch.nn.functional.log_softmax(attention_score, dim=-1)
|
|
1093
|
+
|
|
1094
|
+
# compute the weighted sum
|
|
1095
|
+
residual = torch.matmul(attention, neighborhood_flow_residual)
|
|
1096
|
+
|
|
1097
|
+
# reshape the residual to B, H, W, 2, then B, 2, H, W
|
|
1098
|
+
residual = residual.reshape(B, H, W, 2).permute(0, 3, 1, 2)
|
|
1099
|
+
|
|
1100
|
+
return residual, log_softmax_attention.reshape(B, H, W, P, P)
|
|
1101
|
+
|
|
1102
|
+
def _downstream_head(self, head_num, decout, img_shape):
|
|
1103
|
+
"Run the respective prediction heads"
|
|
1104
|
+
# if self.info_sharing_and_head_structure == "dual+single":
|
|
1105
|
+
|
|
1106
|
+
head = getattr(self, f"head{head_num}") if head_num != "uncertainty" else self.uncertainty_head
|
|
1107
|
+
|
|
1108
|
+
head_num = head_num if head_num != "uncertainty" else 1 # uncertainty head is always from branch 1
|
|
1109
|
+
|
|
1110
|
+
if self.head_type == "linear":
|
|
1111
|
+
head_input = PredictionHeadInput(last_feature=decout[f"{head_num}"])
|
|
1112
|
+
elif self.head_type in ["dpt", "moge_conv"]:
|
|
1113
|
+
head_input = PredictionHeadLayeredInput(list_features=decout[f"{head_num}"], target_output_shape=img_shape)
|
|
1114
|
+
|
|
1115
|
+
return head(head_input)
|
|
1116
|
+
|
|
1117
|
+
def obtain_neighborhood_features(
|
|
1118
|
+
self, flow_estimation: torch.Tensor, other_features: torch.Tensor, local_patch: int = 5
|
|
1119
|
+
) -> Tuple[torch.Tensor, torch.Tensor]:
|
|
1120
|
+
"""
|
|
1121
|
+
Query the other features according to flow estimation to extract neighborhood features.
|
|
1122
|
+
|
|
1123
|
+
This method samples features from the second view at locations determined by the flow estimation
|
|
1124
|
+
plus a local neighborhood around each predicted flow vector. It uses bilinear interpolation
|
|
1125
|
+
to sample features at non-integer coordinates.
|
|
1126
|
+
|
|
1127
|
+
Args:
|
|
1128
|
+
flow_estimation (torch.Tensor): Flow prediction tensor with shape (B, 2, H, W)
|
|
1129
|
+
other_features (torch.Tensor): Features from the second view with shape (B, C, H, W)
|
|
1130
|
+
local_patch (int): Size of the local patch for neighborhood sampling (must be odd), default 5
|
|
1131
|
+
|
|
1132
|
+
Returns:
|
|
1133
|
+
Tuple[torch.Tensor, torch.Tensor]: A tuple containing:
|
|
1134
|
+
- expected_output (torch.Tensor): Sampled neighborhood features with shape (B, H, W, P, P, C)
|
|
1135
|
+
where P is the local_patch size
|
|
1136
|
+
- neighborhood_grid_xy_residual (torch.Tensor): Local coordinate offsets with shape (B, H, W, P, P, 2)
|
|
1137
|
+
|
|
1138
|
+
Raises:
|
|
1139
|
+
AssertionError: If local_patch is not an odd number
|
|
1140
|
+
"""
|
|
1141
|
+
|
|
1142
|
+
assert local_patch % 2 == 1, "local_patch should be odd number"
|
|
1143
|
+
|
|
1144
|
+
P = local_patch
|
|
1145
|
+
R = (P - 1) // 2
|
|
1146
|
+
B, C, H, W = other_features.shape
|
|
1147
|
+
|
|
1148
|
+
device = other_features.device
|
|
1149
|
+
|
|
1150
|
+
# expected_output = torch.zeros(B, H, W, P, P, C, device=other_features.device, dtype=torch.float32)
|
|
1151
|
+
|
|
1152
|
+
neighborhood_grid_ij: torch.Tensor
|
|
1153
|
+
|
|
1154
|
+
i_local, j_local = torch.meshgrid(
|
|
1155
|
+
torch.arange(-R, R + 1, device=device), torch.arange(-R, R + 1, device=device), indexing="ij"
|
|
1156
|
+
)
|
|
1157
|
+
ij_local = torch.stack((i_local, j_local), dim=0) # 2, P, P tensor
|
|
1158
|
+
|
|
1159
|
+
# compute the indices of the fetch
|
|
1160
|
+
base_grid_xy = get_meshgrid_torch(W=W, H=H, device=device).permute(2, 0, 1).reshape(1, 2, H, W)
|
|
1161
|
+
|
|
1162
|
+
target_coordinate_xy_float = flow_estimation + base_grid_xy
|
|
1163
|
+
target_coordinate_xy = target_coordinate_xy_float.view(B, 2, H, W, 1, 1)
|
|
1164
|
+
target_coordinate_ij = target_coordinate_xy[:, [1, 0], ...]
|
|
1165
|
+
|
|
1166
|
+
# compute the neighborhood grid
|
|
1167
|
+
neighborhood_grid_ij = target_coordinate_ij + ij_local.view(1, 2, 1, 1, P, P)
|
|
1168
|
+
|
|
1169
|
+
grid_for_sample = neighborhood_grid_ij[:, [1, 0], ...].permute(0, 2, 3, 4, 5, 1).reshape(B, H, W * P * P, 2)
|
|
1170
|
+
grid_for_sample = (grid_for_sample + 0.5) / torch.tensor([W, H], device=device).view(1, 1, 1, 2)
|
|
1171
|
+
grid_for_sample = grid_for_sample * 2 - 1
|
|
1172
|
+
|
|
1173
|
+
expected_output = torch.nn.functional.grid_sample(
|
|
1174
|
+
other_features, grid=grid_for_sample, mode="bicubic", padding_mode="zeros", align_corners=False
|
|
1175
|
+
).view(B, C, H, W, P, P)
|
|
1176
|
+
|
|
1177
|
+
# transform BCHWPP to BHWPPC
|
|
1178
|
+
expected_output = expected_output.permute(0, 2, 3, 4, 5, 1)
|
|
1179
|
+
|
|
1180
|
+
neighborhood_grid_xy_residual = ij_local[[1, 0], ...].view(1, 2, 1, 1, P, P).to(device).float()
|
|
1181
|
+
neighborhood_grid_xy_residual = neighborhood_grid_xy_residual.permute(0, 2, 3, 4, 5, 1).float()
|
|
1182
|
+
|
|
1183
|
+
return expected_output, neighborhood_grid_xy_residual
|
|
1184
|
+
|
|
1185
|
+
def _initialize_classification_head(self, classification_head_kwargs: Dict[str, Any]):
|
|
1186
|
+
"""
|
|
1187
|
+
Initialize classification head for flow refinement.
|
|
1188
|
+
|
|
1189
|
+
Args:
|
|
1190
|
+
classification_head_kwargs (Dict[str, Any]): Classification head configuration parameters
|
|
1191
|
+
|
|
1192
|
+
Returns:
|
|
1193
|
+
nn.Module: Initialized classification head module
|
|
1194
|
+
|
|
1195
|
+
Raises:
|
|
1196
|
+
ValueError: If unsupported classification head type is specified
|
|
1197
|
+
"""
|
|
1198
|
+
|
|
1199
|
+
if self.classification_head_type == "patch_mlp":
|
|
1200
|
+
return MLPFeature(**classification_head_kwargs)
|
|
1201
|
+
else:
|
|
1202
|
+
raise ValueError(f"Classification head type {self.classification_head_type} not supported.")
|
|
1203
|
+
|
|
1204
|
+
def get_parameter_groups(self) -> Dict[str, torch.nn.ParameterList]:
|
|
1205
|
+
"""
|
|
1206
|
+
Get parameter groups for optimizer. This method guides the optimizer
|
|
1207
|
+
to apply correct learning rate to different parts of the model.
|
|
1208
|
+
|
|
1209
|
+
Returns:
|
|
1210
|
+
Dict[str, torch.nn.ParameterList]: Parameter groups for optimizer with keys:
|
|
1211
|
+
- "encoder": Parameters from the feature encoder
|
|
1212
|
+
- "info_sharing": Parameters from the information sharing module
|
|
1213
|
+
- "output_head": Parameters from the main prediction head
|
|
1214
|
+
- "classification_head": Parameters from the classification refinement head
|
|
1215
|
+
- "unet_feature": Parameters from U-Net features (if use_unet_feature=True)
|
|
1216
|
+
- "uncertainty_head": Parameters from uncertainty head (if present)
|
|
1217
|
+
"""
|
|
1218
|
+
|
|
1219
|
+
if self.use_unet_feature:
|
|
1220
|
+
params_dict = {
|
|
1221
|
+
"encoder": torch.nn.ParameterList(self.encoder.parameters()),
|
|
1222
|
+
"info_sharing": torch.nn.ParameterList(self.info_sharing.parameters()),
|
|
1223
|
+
"output_head": torch.nn.ParameterList(self.head1.parameters()),
|
|
1224
|
+
"classification_head": torch.nn.ParameterList(self.classification_head.parameters()),
|
|
1225
|
+
"unet_feature": torch.nn.ParameterList(
|
|
1226
|
+
list(self.unet_feature.parameters())
|
|
1227
|
+
+ list(self.conv1.parameters())
|
|
1228
|
+
+ list(self.conv2.parameters())
|
|
1229
|
+
+ [self.classification_bias]
|
|
1230
|
+
),
|
|
1231
|
+
}
|
|
1232
|
+
else:
|
|
1233
|
+
params_dict = {
|
|
1234
|
+
"encoder": torch.nn.ParameterList(self.encoder.parameters()),
|
|
1235
|
+
"info_sharing": torch.nn.ParameterList(self.info_sharing.parameters()),
|
|
1236
|
+
"output_head": torch.nn.ParameterList(self.head1.parameters()),
|
|
1237
|
+
"classification_head": torch.nn.ParameterList(self.classification_head.parameters()),
|
|
1238
|
+
}
|
|
1239
|
+
|
|
1240
|
+
if hasattr(self, "uncertainty_head"):
|
|
1241
|
+
params_dict["uncertainty_head"] = torch.nn.ParameterList(self.uncertainty_head.parameters())
|
|
1242
|
+
|
|
1243
|
+
return params_dict
|
|
1244
|
+
|
|
1245
|
+
|
|
1246
|
+
if __name__ == "__main__":
|
|
1247
|
+
import cv2
|
|
1248
|
+
import flow_vis
|
|
1249
|
+
import matplotlib.pyplot as plt
|
|
1250
|
+
import numpy as np
|
|
1251
|
+
import torch
|
|
1252
|
+
|
|
1253
|
+
from uniflowmatch.utils.geometry import get_meshgrid_torch
|
|
1254
|
+
from uniflowmatch.utils.viz import warp_image_with_flow
|
|
1255
|
+
|
|
1256
|
+
refinement_model = False
|
|
1257
|
+
resolution = 560
|
|
1258
|
+
|
|
1259
|
+
if not refinement_model:
|
|
1260
|
+
model = UniFlowMatchConfidence.from_pretrained(
|
|
1261
|
+
"infinity1096/UFM-Base" if resolution == 560 else "infinity1096/UFM-Base-980"
|
|
1262
|
+
)
|
|
1263
|
+
else:
|
|
1264
|
+
model = UniFlowMatchClassificationRefinement.from_pretrained(
|
|
1265
|
+
"infinity1096/UFM-Refine" if resolution == 560 else "infinity1096/UFM-Refine-980"
|
|
1266
|
+
)
|
|
1267
|
+
model.eval()
|
|
1268
|
+
|
|
1269
|
+
# === Load and Prepare Images ===
|
|
1270
|
+
source_path = "examples/image_pairs/fire_academy_0.png"
|
|
1271
|
+
target_path = "examples/image_pairs/fire_academy_1.png"
|
|
1272
|
+
|
|
1273
|
+
source_image = cv2.imread(source_path)
|
|
1274
|
+
target_image = cv2.imread(target_path)
|
|
1275
|
+
|
|
1276
|
+
source_image = cv2.cvtColor(source_image, cv2.COLOR_BGR2RGB)
|
|
1277
|
+
target_image = cv2.cvtColor(target_image, cv2.COLOR_BGR2RGB)
|
|
1278
|
+
|
|
1279
|
+
# === Predict Correspondences ===
|
|
1280
|
+
result = model.predict_correspondences_batched(
|
|
1281
|
+
source_image=torch.from_numpy(source_image),
|
|
1282
|
+
target_image=torch.from_numpy(target_image),
|
|
1283
|
+
)
|
|
1284
|
+
|
|
1285
|
+
flow_output = result.flow.flow_output[0].cpu().numpy()
|
|
1286
|
+
covisibility = result.covisibility.mask[0].cpu().numpy()
|
|
1287
|
+
|
|
1288
|
+
# === Visualize Results ===
|
|
1289
|
+
fig, axs = plt.subplots(2, 3, figsize=(15, 5))
|
|
1290
|
+
|
|
1291
|
+
axs[0, 0].imshow(source_image)
|
|
1292
|
+
axs[0, 0].set_title("Source Image")
|
|
1293
|
+
|
|
1294
|
+
axs[0, 1].imshow(target_image)
|
|
1295
|
+
axs[0, 1].set_title("Target Image")
|
|
1296
|
+
|
|
1297
|
+
# Warp the image using flow
|
|
1298
|
+
warped_image = warp_image_with_flow(source_image, None, target_image, flow_output.transpose(1, 2, 0))
|
|
1299
|
+
warped_image = covisibility[..., None] * warped_image + (1 - covisibility[..., None]) * 255 * np.ones_like(
|
|
1300
|
+
warped_image
|
|
1301
|
+
)
|
|
1302
|
+
warped_image /= 255.0
|
|
1303
|
+
|
|
1304
|
+
axs[0, 2].imshow(warped_image)
|
|
1305
|
+
axs[0, 2].set_title("Warped Image")
|
|
1306
|
+
|
|
1307
|
+
# Flow visualization
|
|
1308
|
+
flow_vis_image = flow_vis.flow_to_color(flow_output.transpose(1, 2, 0))
|
|
1309
|
+
axs[1, 0].imshow(flow_vis_image)
|
|
1310
|
+
axs[1, 0].set_title("Flow Output (Valid at covisible region)")
|
|
1311
|
+
|
|
1312
|
+
# Covisibility mask
|
|
1313
|
+
axs[1, 1].imshow(covisibility > 0.5, cmap="gray", vmin=0, vmax=1)
|
|
1314
|
+
axs[1, 1].set_title("Covisibility Mask (Thresholded by 0.5)")
|
|
1315
|
+
|
|
1316
|
+
heatmap = axs[1, 2].imshow(covisibility, cmap="gray", vmin=0, vmax=1)
|
|
1317
|
+
axs[1, 2].set_title("Covisibility Mask")
|
|
1318
|
+
plt.colorbar(heatmap, ax=axs[1, 2])
|
|
1319
|
+
|
|
1320
|
+
plt.tight_layout()
|
|
1321
|
+
plt.savefig("ufm_output.png")
|
|
1322
|
+
plt.show()
|
|
1323
|
+
print("Saved ufm_output.png")
|