mct-nightly 2.1.0.20240806.441__tar.gz → 2.1.0.20240807.445__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/PKG-INFO +1 -1
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/mct_nightly.egg-info/PKG-INFO +1 -1
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/mct_nightly.egg-info/requires.txt +1 -1
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/__init__.py +1 -1
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/node_quantization_config.py +1 -1
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/qparams_activations_computation.py +3 -4
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/__init__.py +2 -1
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/op_quantization_config.py +18 -4
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1/tp_model.py +5 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_lut/tp_model.py +5 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_pot/tp_model.py +5 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2/tp_model.py +5 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2_lut/tp_model.py +5 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3/tp_model.py +5 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3_lut/tp_model.py +5 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v4/tp_model.py +6 -4
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v4/tpc_pytorch.py +3 -3
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/qnnpack_tpc/v1/tp_model.py +5 -4
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/tflite_tpc/v1/tp_model.py +5 -4
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/LICENSE.md +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/README.md +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/mct_nightly.egg-info/SOURCES.txt +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/mct_nightly.egg-info/dependency_links.txt +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/mct_nightly.egg-info/top_level.txt +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/analyzer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/back2framework/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/back2framework/base_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/base_substitutions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/collectors/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/collectors/base_collector.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/collectors/histogram_collector.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/collectors/mean_collector.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/collectors/min_max_per_channel_collector.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/collectors/statistics_collector.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/framework_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/framework_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/fusion/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/fusion/layer_fusing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/base_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/base_node.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/edge.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/functional_node.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/graph_matchers.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/graph_searches.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/memory_graph/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/memory_graph/bipartite_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/memory_graph/compute_graph_max_cut.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/memory_graph/cut.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/memory_graph/max_cut_astar.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/memory_graph/memory_element.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/memory_graph/memory_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/graph/virtual_activation_weights_node.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/hessian/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/hessian/hessian_info_service.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/hessian/hessian_info_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/hessian/hessian_scores_calculator.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/hessian/hessian_scores_request.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/matchers/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/matchers/base_graph_filter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/matchers/base_matcher.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/matchers/edge_matcher.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/matchers/function.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/matchers/node_matcher.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/matchers/walk_matcher.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/memory_computation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/bit_width_setter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/configurable_quant_id.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/configurable_quantizer_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/distance_weighting.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/mixed_precision_quantization_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/mixed_precision_search_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/mixed_precision_search_manager.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/resource_utilization_tools/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/resource_utilization_tools/resource_utilization.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/resource_utilization_tools/resource_utilization_data.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/resource_utilization_tools/ru_aggregation_methods.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/resource_utilization_tools/ru_functions_mapping.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/resource_utilization_tools/ru_methods.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/search_methods/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/search_methods/linear_programming.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/sensitivity_evaluation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/set_layer_to_bitwidth.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/mixed_precision/solution_refinement_procedure.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/model_builder_mode.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/model_collector.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/model_validation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/network_editors/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/network_editors/actions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/network_editors/edit_network.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/network_editors/node_filters.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/node_prior_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/channels_grouping.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/greedy_mask_calculator.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/importance_metrics/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/importance_metrics/base_importance_metric.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/importance_metrics/importance_metric_factory.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/importance_metrics/lfh_importance_metric.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/mask/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/mask/per_channel_mask.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/mask/per_simd_group_mask.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/memory_calculator.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/prune_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/pruner.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/pruning_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/pruning_framework_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/pruning_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/pruning/pruning_section.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/candidate_node_quantization_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/core_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/debug_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/filter_nodes_candidates.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_fn_selection.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_fn_selection.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/error_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/lut_kmeans_params.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/outlier_filter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/power_of_two_selection.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/qparams_computation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/qparams_search.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/qparams_weights_computation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/symmetric_selection.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantization_params_generation/uniform_selection.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantize_graph_weights.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantize_node.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantizers/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantizers/lut_kmeans_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantizers/quantizers_helpers.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/quantizers/uniform_quantizers.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/quantization/set_node_quantization_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/similarity_analyzer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/statistics_correction/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/statistics_correction/apply_bias_correction_to_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/statistics_correction/apply_second_moment_correction_to_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/statistics_correction/compute_bias_correction_of_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/statistics_correction/statistics_correction.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/apply_substitutions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/batchnorm_folding.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/batchnorm_reconstruction.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/batchnorm_refusing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/linear_collapsing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/linear_collapsing_substitution.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/remove_identity.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/residual_collapsing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/scale_equalization.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/shift_negative_activation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/softmax_shift.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/virtual_activation_weights_composition.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/substitutions/weights_activation_split.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/user_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/visualization/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/visualization/final_config_visualizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/visualization/nn_visualizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/common/visualization/tensorboard_writer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/graph_prep_runner.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/back2framework/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/back2framework/factory_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/back2framework/float_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/back2framework/instance_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/back2framework/keras_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/back2framework/mixed_precision_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/back2framework/quantized_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/custom_layer_validation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/default_framework_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/activation_decomposition.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/batchnorm_folding.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/batchnorm_reconstruction.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/batchnorm_refusing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/concat_threshold_update.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/dwconv_to_conv.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/input_scaling.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/linear_collapsing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/matmul_substitution.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/multi_head_attention_decomposition.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/relu_bound_to_power_of_2.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/remove_identity.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/residual_collapsing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/scale_equalization.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/separableconv_decomposition.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/shift_negative_activation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/softmax_shift.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/virtual_activation_weights_composition.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/graph_substitutions/substitutions/weights_activation_split.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/hessian/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/hessian/activation_hessian_scores_calculator_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/hessian/hessian_scores_calculator_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/hessian/weights_hessian_scores_calculator_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/keras_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/keras_model_validation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/keras_node_prior_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/mixed_precision/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/mixed_precision/configurable_activation_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/mixed_precision/configurable_weights_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/pruning/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/pruning/pruning_keras_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/quantizer/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/quantizer/base_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/quantizer/fake_quant_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/quantizer/lut_fake_quant.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/common.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/connectivity_handler.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/nested_model/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/nested_model/edges_merger.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/nested_model/nested_model_handler.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/nested_model/nodes_merger.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/nested_model/outputs_merger.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/node_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/reader/reader.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/resource_utilization_data_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/statistics_correction/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/statistics_correction/apply_second_moment_correction.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/tf_tensor_numpy.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/keras/visualization/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/factory_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/float_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/instance_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/mixed_precision_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/pytorch_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/quantization_wrapper/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/quantization_wrapper/quantized_layer_wrapper.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/quantization_wrapper/wrapper_quantize_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/back2framework/quantized_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/default_framework_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/batchnorm_folding.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/batchnorm_reconstruction.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/batchnorm_refusing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/concat_threshold_update.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/const_holder_conv.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/functional_batch_norm.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/functional_layer_norm.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/linear_collapsing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/multi_head_attention_decomposition.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/relu_bound_to_power_of_2.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/remove_identity.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/reshape_with_static_shapes.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/residual_collapsing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/scale_equalization.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/shift_negative_activation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/softmax_shift.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/transform_function_call_method.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/virtual_activation_weights_composition.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/graph_substitutions/substitutions/weights_activation_split.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/hessian/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/hessian/activation_hessian_scores_calculator_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/hessian/hessian_scores_calculator_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/hessian/weights_hessian_scores_calculator_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/mixed_precision/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/mixed_precision/configurable_activation_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/mixed_precision/configurable_weights_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/pruning/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/pruning/pruning_pytorch_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/pytorch_device_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/pytorch_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/pytorch_node_prior_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/quantizer/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/quantizer/fake_quant_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/quantizer/lut_fake_quant.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/reader/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/reader/graph_builders.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/reader/node_holders.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/reader/reader.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/resource_utilization_data_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/statistics_correction/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/statistics_correction/apply_second_moment_correction.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/pytorch/utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/quantization_prep_runner.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/core/runner.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/data_generation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/data_generation_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/enums.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/image_pipeline.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/model_info_exctractors.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/common/optimization_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/image_operations.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/image_pipeline.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/keras_data_generation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/model_info_exctractors.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_functions/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_functions/batchnorm_alignment_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_functions/bn_layer_weighting_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_functions/image_initilization.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_functions/lr_scheduler.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_functions/output_loss_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_functions/scheduler_step_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/keras/optimization_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/image_operations.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/image_pipeline.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/model_info_exctractors.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_functions/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_functions/batchnorm_alignment_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_functions/bn_layer_weighting_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_functions/image_initilization.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_functions/lr_scheduler.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_functions/output_loss_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_functions/scheduler_step_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/optimization_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/data_generation/pytorch/pytorch_data_generation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/defaultdict.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/fw_agonstic/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/fw_agonstic/exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/fw_agonstic/quantization_format.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/base_keras_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/export_serialization_format.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/fakely_quant_keras_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/fakely_quant_tflite_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/int8_tflite_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/keras_export_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/keras/mctq_keras_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/pytorch/base_pytorch_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/pytorch/export_serialization_format.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/pytorch/fakely_quant_onnx_pytorch_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/pytorch/fakely_quant_torchscript_pytorch_exporter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_exporter/pytorch/pytorch_export_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/fw_agnostic/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/fw_agnostic/get_inferable_quantizers.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/keras/builder/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/keras/builder/fully_quantized_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/keras/builder/node_to_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/keras/validate_layer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/pytorch/builder/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/pytorch/builder/fully_quantized_model_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/pytorch/builder/node_to_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/exporter/model_wrapper/pytorch/validate_layer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/common/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/common/gptq_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/common/gptq_constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/common/gptq_framework_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/common/gptq_graph.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/common/gptq_training.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/gptq_keras_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/gptq_loss.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/gptq_training.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/graph_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantization_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/base_keras_gptq_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/quant_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/quantization_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/regularization_factory.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/soft_rounding/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/soft_rounding/soft_quantizer_reg.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/soft_rounding/symmetric_soft_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/soft_rounding/uniform_soft_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/ste_rounding/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/keras/quantizer/ste_rounding/symmetric_ste.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/gptq_loss.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/gptq_pytorch_implementation.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/gptq_training.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/graph_info.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantization_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/base_pytorch_gptq_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/quant_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/quantization_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/regularization_factory.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/soft_rounding/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/soft_rounding/soft_quantizer_reg.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/soft_rounding/symmetric_soft_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/soft_rounding/uniform_soft_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/ste_rounding/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/pytorch/quantizer/ste_rounding/symmetric_ste.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/gptq/runner.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/logger.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/metadata.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/pruning/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/pruning/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/pruning/keras/pruning_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/pruning/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/pruning/pytorch/pruning_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/ptq/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/ptq/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/ptq/keras/quantization_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/ptq/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/ptq/pytorch/quantization_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/ptq/runner.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/common/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/common/qat_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantization_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/base_keras_qat_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/lsq/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/lsq/symmetric_lsq.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/lsq/uniform_lsq.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/quant_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/quantization_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/ste_rounding/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/ste_rounding/symmetric_ste.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/keras/quantizer/ste_rounding/uniform_ste.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantization_facade.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/base_pytorch_qat_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/lsq/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/lsq/symmetric_lsq.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/lsq/uniform_lsq.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/quantization_builder.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/quantizer_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/ste_rounding/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/ste_rounding/symmetric_ste.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/qat/pytorch/quantizer/ste_rounding/uniform_ste.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/immutable.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/current_tp_model.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/fusing.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/operators.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/target_platform_model.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/target_platform_model_component.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/targetplatform2framework/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/targetplatform2framework/attribute_filter.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/targetplatform2framework/current_tpc.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/targetplatform2framework/layer_filter_params.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/targetplatform2framework/operations_to_layers.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/targetplatform2framework/target_platform_capabilities.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/target_platform/targetplatform2framework/target_platform_capabilities_component.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/get_target_platform_capabilities.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/latest/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/target_platform_capabilities.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_lut/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_lut/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_lut/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_pot/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_pot/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v1_pot/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2_lut/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2_lut/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v2_lut/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3_lut/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3_lut/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v3_lut/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v4/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/imx500_tpc/v4/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/qnnpack_tpc/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/qnnpack_tpc/latest/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/qnnpack_tpc/target_platform_capabilities.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/qnnpack_tpc/v1/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/qnnpack_tpc/v1/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/qnnpack_tpc/v1/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/tflite_tpc/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/tflite_tpc/latest/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/tflite_tpc/target_platform_capabilities.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/tflite_tpc/v1/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/tflite_tpc/v1/tpc_keras.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/target_platform_capabilities/tpc_models/tflite_tpc/v1/tpc_pytorch.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/common/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/common/base_trainable_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/common/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/common/get_quantizer_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/common/get_quantizers.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/common/quant_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/common/trainable_quantizer_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/keras/base_keras_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/keras/config_serialization.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/keras/load_model.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/keras/quantize_wrapper.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/keras/quantizer_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/trainable_infrastructure/pytorch/base_pytorch_quantizer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/constants.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/core_report_generator.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/dataset_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/framework_report_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/model_analyzer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/model_folding_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/similarity_calculator.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/similarity_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/tensorboard_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/common/xquant_config.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/keras/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/keras/dataset_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/keras/facade_xquant_report.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/keras/keras_report_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/keras/model_analyzer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/keras/similarity_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/keras/tensorboard_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/pytorch/__init__.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/pytorch/dataset_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/pytorch/facade_xquant_report.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/pytorch/model_analyzer.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/pytorch/pytorch_report_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/pytorch/similarity_functions.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/model_compression_toolkit/xquant/pytorch/tensorboard_utils.py +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/setup.cfg +0 -0
- {mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/setup.py +0 -0
@@ -27,4 +27,4 @@ from model_compression_toolkit import data_generation
|
|
27
27
|
from model_compression_toolkit import pruning
|
28
28
|
from model_compression_toolkit.trainable_infrastructure.keras.load_model import keras_load_quantized_model
|
29
29
|
|
30
|
-
__version__ = "2.1.0.
|
30
|
+
__version__ = "2.1.0.20240807.000445"
|
@@ -96,7 +96,7 @@ class NodeActivationQuantizationConfig(BaseNodeQuantizationConfig):
|
|
96
96
|
self.activation_n_bits = op_cfg.activation_n_bits
|
97
97
|
self.relu_bound_to_power_of_2 = qc.relu_bound_to_power_of_2
|
98
98
|
self.enable_activation_quantization = op_cfg.enable_activation_quantization
|
99
|
-
self.
|
99
|
+
self.signedness = op_cfg.signedness
|
100
100
|
self.activation_channel_equalization = qc.activation_channel_equalization
|
101
101
|
self.input_scaling = qc.input_scaling
|
102
102
|
self.min_threshold = qc.min_threshold
|
@@ -15,9 +15,8 @@
|
|
15
15
|
import numpy as np
|
16
16
|
from typing import Dict, Union
|
17
17
|
|
18
|
-
from model_compression_toolkit.target_platform_capabilities.target_platform import QuantizationMethod
|
18
|
+
from model_compression_toolkit.target_platform_capabilities.target_platform import QuantizationMethod, Signedness
|
19
19
|
from model_compression_toolkit.core.common.collectors.statistics_collector import BaseStatsCollector
|
20
|
-
from model_compression_toolkit.constants import SIGNED
|
21
20
|
from model_compression_toolkit.core.common.quantization import quantization_params_generation
|
22
21
|
from model_compression_toolkit.core.common.node_prior_info import NodePriorInfo
|
23
22
|
from model_compression_toolkit.core.common.quantization.node_quantization_config import NodeActivationQuantizationConfig
|
@@ -49,8 +48,8 @@ def get_activations_qparams(activation_quant_cfg: NodeActivationQuantizationConf
|
|
49
48
|
bins_counts)
|
50
49
|
min_value, max_value = out_stats_container.get_min_max_values()
|
51
50
|
|
52
|
-
if activation_quant_cfg.
|
53
|
-
signed = activation_quant_cfg.
|
51
|
+
if activation_quant_cfg.signedness in [Signedness.SIGNED, Signedness.UNSIGNED]:
|
52
|
+
signed = activation_quant_cfg.signedness == Signedness.SIGNED
|
54
53
|
elif nodes_prior_info.is_output_bounded():
|
55
54
|
signed = min_value < 0
|
56
55
|
else:
|
@@ -17,7 +17,8 @@ from model_compression_toolkit.target_platform_capabilities.target_platform.fusi
|
|
17
17
|
from model_compression_toolkit.target_platform_capabilities.target_platform.targetplatform2framework.attribute_filter import AttributeFilter
|
18
18
|
from model_compression_toolkit.target_platform_capabilities.target_platform.targetplatform2framework import TargetPlatformCapabilities, OperationsSetToLayers, Smaller, SmallerEq, NotEq, Eq, GreaterEq, Greater, LayerFilterParams, OperationsToLayers, get_current_tpc
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.target_platform.target_platform_model import get_default_quantization_config_options, TargetPlatformModel
|
20
|
-
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import
|
20
|
+
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
21
|
+
OpQuantizationConfig, QuantizationConfigOptions, AttributeQuantizationConfig, Signedness
|
21
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.operators import OperatorsSet, OperatorSetConcat
|
22
23
|
|
23
24
|
from mct_quantizers import QuantizationMethod
|
@@ -15,12 +15,26 @@
|
|
15
15
|
|
16
16
|
import copy
|
17
17
|
from typing import List, Dict, Union, Any, Tuple
|
18
|
+
from enum import Enum
|
18
19
|
|
19
20
|
from mct_quantizers import QuantizationMethod
|
20
21
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
21
22
|
from model_compression_toolkit.logger import Logger
|
22
23
|
|
23
24
|
|
25
|
+
class Signedness(Enum):
|
26
|
+
"""
|
27
|
+
An enum for choosing the signedness of the quantization method:
|
28
|
+
|
29
|
+
AUTO - Signedness decided automatically by quantization.
|
30
|
+
SIGNED - Force signed quantization.
|
31
|
+
UNSIGNED - Force unsigned quantization.
|
32
|
+
"""
|
33
|
+
AUTO = 0
|
34
|
+
SIGNED = 1
|
35
|
+
UNSIGNED = 2
|
36
|
+
|
37
|
+
|
24
38
|
def clone_and_edit_object_params(obj: Any, **kwargs: Dict) -> Any:
|
25
39
|
"""
|
26
40
|
Clones the given object and edit some of its parameters.
|
@@ -120,7 +134,7 @@ class OpQuantizationConfig:
|
|
120
134
|
fixed_scale: float,
|
121
135
|
fixed_zero_point: int,
|
122
136
|
simd_size: int,
|
123
|
-
|
137
|
+
signedness: Signedness
|
124
138
|
):
|
125
139
|
"""
|
126
140
|
|
@@ -134,8 +148,8 @@ class OpQuantizationConfig:
|
|
134
148
|
quantization_preserving (bool): Whether quantization parameters should be the same for an operator's input and output.
|
135
149
|
fixed_scale (float): Scale to use for an operator quantization parameters.
|
136
150
|
fixed_zero_point (int): Zero-point to use for an operator quantization parameters.
|
137
|
-
is_signed (bool): Force activation quantization signedness (None means don't force).
|
138
151
|
simd_size (int): Per op integer representing the Single Instruction, Multiple Data (SIMD) width of an operator. It indicates the number of data elements that can be fetched and processed simultaneously in a single instruction.
|
152
|
+
signedness (bool): Set activation quantization signedness.
|
139
153
|
|
140
154
|
"""
|
141
155
|
|
@@ -154,7 +168,7 @@ class OpQuantizationConfig:
|
|
154
168
|
self.quantization_preserving = quantization_preserving
|
155
169
|
self.fixed_scale = fixed_scale
|
156
170
|
self.fixed_zero_point = fixed_zero_point
|
157
|
-
self.
|
171
|
+
self.signedness = signedness
|
158
172
|
self.simd_size = simd_size
|
159
173
|
|
160
174
|
def get_info(self):
|
@@ -206,7 +220,7 @@ class OpQuantizationConfig:
|
|
206
220
|
self.activation_n_bits == other.activation_n_bits and \
|
207
221
|
self.supported_input_activation_n_bits == other.supported_input_activation_n_bits and \
|
208
222
|
self.enable_activation_quantization == other.enable_activation_quantization and \
|
209
|
-
self.
|
223
|
+
self.signedness == other.signedness and \
|
210
224
|
self.simd_size == other.simd_size
|
211
225
|
|
212
226
|
@property
|
@@ -18,7 +18,7 @@ import model_compression_toolkit as mct
|
|
18
18
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS
|
20
20
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
21
|
-
TargetPlatformModel
|
21
|
+
TargetPlatformModel, Signedness
|
22
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
23
23
|
AttributeQuantizationConfig
|
24
24
|
|
@@ -98,7 +98,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
98
98
|
quantization_preserving=False,
|
99
99
|
fixed_scale=None,
|
100
100
|
fixed_zero_point=None,
|
101
|
-
simd_size=32
|
101
|
+
simd_size=32,
|
102
|
+
signedness=Signedness.AUTO)
|
102
103
|
|
103
104
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
104
105
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -111,7 +112,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
111
112
|
quantization_preserving=False,
|
112
113
|
fixed_scale=None,
|
113
114
|
fixed_zero_point=None,
|
114
|
-
simd_size=32
|
115
|
+
simd_size=32,
|
116
|
+
signedness=Signedness.AUTO)
|
115
117
|
|
116
118
|
# To quantize a model using mixed-precision, create
|
117
119
|
# a list with more than one OpQuantizationConfig.
|
@@ -19,7 +19,7 @@ from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS, \
|
20
20
|
WEIGHTS_QUANTIZATION_METHOD
|
21
21
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
22
|
-
TargetPlatformModel
|
22
|
+
TargetPlatformModel, Signedness
|
23
23
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
24
24
|
AttributeQuantizationConfig
|
25
25
|
|
@@ -94,7 +94,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
94
94
|
quantization_preserving=False,
|
95
95
|
fixed_scale=None,
|
96
96
|
fixed_zero_point=None,
|
97
|
-
simd_size=32
|
97
|
+
simd_size=32,
|
98
|
+
signedness=Signedness.AUTO)
|
98
99
|
|
99
100
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
100
101
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -107,7 +108,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
107
108
|
quantization_preserving=False,
|
108
109
|
fixed_scale=None,
|
109
110
|
fixed_zero_point=None,
|
110
|
-
simd_size=32
|
111
|
+
simd_size=32,
|
112
|
+
signedness=Signedness.AUTO)
|
111
113
|
|
112
114
|
# To quantize a model using mixed-precision, create
|
113
115
|
# a list with more than one OpQuantizationConfig.
|
@@ -18,7 +18,7 @@ import model_compression_toolkit as mct
|
|
18
18
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS
|
20
20
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
21
|
-
TargetPlatformModel
|
21
|
+
TargetPlatformModel, Signedness
|
22
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
23
23
|
AttributeQuantizationConfig
|
24
24
|
|
@@ -94,7 +94,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
94
94
|
quantization_preserving=False,
|
95
95
|
fixed_scale=None,
|
96
96
|
fixed_zero_point=None,
|
97
|
-
simd_size=32
|
97
|
+
simd_size=32,
|
98
|
+
signedness=Signedness.AUTO)
|
98
99
|
|
99
100
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
100
101
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -107,7 +108,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
107
108
|
quantization_preserving=False,
|
108
109
|
fixed_scale=None,
|
109
110
|
fixed_zero_point=None,
|
110
|
-
simd_size=32
|
111
|
+
simd_size=32,
|
112
|
+
signedness=Signedness.AUTO)
|
111
113
|
|
112
114
|
# To quantize a model using mixed-precision, create
|
113
115
|
# a list with more than one OpQuantizationConfig.
|
@@ -18,7 +18,7 @@ import model_compression_toolkit as mct
|
|
18
18
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS
|
20
20
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
21
|
-
TargetPlatformModel
|
21
|
+
TargetPlatformModel, Signedness
|
22
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
23
23
|
AttributeQuantizationConfig
|
24
24
|
|
@@ -100,7 +100,8 @@ def get_op_quantization_configs() -> \
|
|
100
100
|
quantization_preserving=False,
|
101
101
|
fixed_scale=None,
|
102
102
|
fixed_zero_point=None,
|
103
|
-
simd_size=32
|
103
|
+
simd_size=32,
|
104
|
+
signedness=Signedness.AUTO)
|
104
105
|
|
105
106
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
106
107
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -113,7 +114,8 @@ def get_op_quantization_configs() -> \
|
|
113
114
|
quantization_preserving=False,
|
114
115
|
fixed_scale=None,
|
115
116
|
fixed_zero_point=None,
|
116
|
-
simd_size=32
|
117
|
+
simd_size=32,
|
118
|
+
signedness=Signedness.AUTO)
|
117
119
|
|
118
120
|
# To quantize a model using mixed-precision, create
|
119
121
|
# a list with more than one OpQuantizationConfig.
|
@@ -19,7 +19,7 @@ from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS, \
|
20
20
|
WEIGHTS_QUANTIZATION_METHOD
|
21
21
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
22
|
-
TargetPlatformModel
|
22
|
+
TargetPlatformModel, Signedness
|
23
23
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
24
24
|
AttributeQuantizationConfig
|
25
25
|
|
@@ -96,7 +96,8 @@ def get_op_quantization_configs() -> \
|
|
96
96
|
quantization_preserving=False,
|
97
97
|
fixed_scale=None,
|
98
98
|
fixed_zero_point=None,
|
99
|
-
simd_size=32
|
99
|
+
simd_size=32,
|
100
|
+
signedness=Signedness.AUTO)
|
100
101
|
|
101
102
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
102
103
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -109,7 +110,8 @@ def get_op_quantization_configs() -> \
|
|
109
110
|
quantization_preserving=False,
|
110
111
|
fixed_scale=None,
|
111
112
|
fixed_zero_point=None,
|
112
|
-
simd_size=32
|
113
|
+
simd_size=32,
|
114
|
+
signedness=Signedness.AUTO)
|
113
115
|
|
114
116
|
# To quantize a model using mixed-precision, create
|
115
117
|
# a list with more than one OpQuantizationConfig.
|
@@ -18,7 +18,7 @@ import model_compression_toolkit as mct
|
|
18
18
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS
|
20
20
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
21
|
-
TargetPlatformModel
|
21
|
+
TargetPlatformModel, Signedness
|
22
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
23
23
|
AttributeQuantizationConfig
|
24
24
|
|
@@ -100,7 +100,8 @@ def get_op_quantization_configs() -> \
|
|
100
100
|
quantization_preserving=False,
|
101
101
|
fixed_scale=None,
|
102
102
|
fixed_zero_point=None,
|
103
|
-
simd_size=32
|
103
|
+
simd_size=32,
|
104
|
+
signedness=Signedness.AUTO)
|
104
105
|
|
105
106
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
106
107
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -113,7 +114,8 @@ def get_op_quantization_configs() -> \
|
|
113
114
|
quantization_preserving=False,
|
114
115
|
fixed_scale=None,
|
115
116
|
fixed_zero_point=None,
|
116
|
-
simd_size=32
|
117
|
+
simd_size=32,
|
118
|
+
signedness=Signedness.AUTO)
|
117
119
|
|
118
120
|
# To quantize a model using mixed-precision, create
|
119
121
|
# a list with more than one OpQuantizationConfig.
|
@@ -19,7 +19,7 @@ from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS, \
|
20
20
|
WEIGHTS_QUANTIZATION_METHOD
|
21
21
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
22
|
-
TargetPlatformModel
|
22
|
+
TargetPlatformModel, Signedness
|
23
23
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
24
24
|
AttributeQuantizationConfig
|
25
25
|
|
@@ -96,7 +96,8 @@ def get_op_quantization_configs() -> \
|
|
96
96
|
quantization_preserving=False,
|
97
97
|
fixed_scale=None,
|
98
98
|
fixed_zero_point=None,
|
99
|
-
simd_size=32
|
99
|
+
simd_size=32,
|
100
|
+
signedness=Signedness.AUTO)
|
100
101
|
|
101
102
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
102
103
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -109,7 +110,8 @@ def get_op_quantization_configs() -> \
|
|
109
110
|
quantization_preserving=False,
|
110
111
|
fixed_scale=None,
|
111
112
|
fixed_zero_point=None,
|
112
|
-
simd_size=32
|
113
|
+
simd_size=32,
|
114
|
+
signedness=Signedness.AUTO)
|
113
115
|
|
114
116
|
# To quantize a model using mixed-precision, create
|
115
117
|
# a list with more than one OpQuantizationConfig.
|
@@ -18,7 +18,7 @@ import model_compression_toolkit as mct
|
|
18
18
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR, WEIGHTS_N_BITS
|
20
20
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
21
|
-
TargetPlatformModel
|
21
|
+
TargetPlatformModel, Signedness
|
22
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
23
23
|
AttributeQuantizationConfig
|
24
24
|
|
@@ -100,7 +100,8 @@ def get_op_quantization_configs() -> \
|
|
100
100
|
quantization_preserving=False,
|
101
101
|
fixed_scale=None,
|
102
102
|
fixed_zero_point=None,
|
103
|
-
simd_size=32
|
103
|
+
simd_size=32,
|
104
|
+
signedness=Signedness.AUTO)
|
104
105
|
|
105
106
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
106
107
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -113,7 +114,8 @@ def get_op_quantization_configs() -> \
|
|
113
114
|
quantization_preserving=False,
|
114
115
|
fixed_scale=None,
|
115
116
|
fixed_zero_point=None,
|
116
|
-
simd_size=32
|
117
|
+
simd_size=32,
|
118
|
+
signedness=Signedness.AUTO)
|
117
119
|
|
118
120
|
# To quantize a model using mixed-precision, create
|
119
121
|
# a list with more than one OpQuantizationConfig.
|
@@ -170,7 +172,7 @@ def generate_tp_model(default_config: OpQuantizationConfig,
|
|
170
172
|
const_config_input16 = const_config.clone_and_edit(
|
171
173
|
supported_input_activation_n_bits=(8, 16))
|
172
174
|
const_config_input16_output16 = const_config_input16.clone_and_edit(
|
173
|
-
activation_n_bits=16,
|
175
|
+
activation_n_bits=16, signedness=Signedness.SIGNED)
|
174
176
|
const_configuration_options_inout16 = tp.QuantizationConfigOptions([const_config_input16_output16,
|
175
177
|
const_config_input16],
|
176
178
|
base_config=const_config_input16)
|
@@ -17,7 +17,7 @@ import operator
|
|
17
17
|
|
18
18
|
import torch
|
19
19
|
from torch import add, sub, mul, div, flatten, reshape, split, unsqueeze, dropout, sigmoid, tanh, chunk, unbind, topk, \
|
20
|
-
gather, equal, transpose, permute, argmax, squeeze
|
20
|
+
gather, equal, transpose, permute, argmax, squeeze, multiply, subtract
|
21
21
|
from torch.nn import Conv2d, Linear, ConvTranspose2d, MaxPool2d
|
22
22
|
from torch.nn import Dropout, Flatten, Hardtanh, Identity
|
23
23
|
from torch.nn import ReLU, ReLU6, PReLU, SiLU, Sigmoid, Tanh, Hardswish, LeakyReLU
|
@@ -101,8 +101,8 @@ def generate_pytorch_tpc(name: str, tp_model: tp.TargetPlatformModel):
|
|
101
101
|
tp.LayerFilterParams(hardtanh, min_val=0)])
|
102
102
|
|
103
103
|
tp.OperationsSetToLayers("Add", [operator.add, add])
|
104
|
-
tp.OperationsSetToLayers("Sub", [operator.sub, sub])
|
105
|
-
tp.OperationsSetToLayers("Mul", [operator.mul, mul])
|
104
|
+
tp.OperationsSetToLayers("Sub", [operator.sub, sub, subtract])
|
105
|
+
tp.OperationsSetToLayers("Mul", [operator.mul, mul, multiply])
|
106
106
|
tp.OperationsSetToLayers("Div", [operator.truediv, div])
|
107
107
|
tp.OperationsSetToLayers("PReLU", [PReLU, prelu])
|
108
108
|
tp.OperationsSetToLayers("Swish", [SiLU, silu, Hardswish, hardswish])
|
@@ -18,7 +18,7 @@ import model_compression_toolkit as mct
|
|
18
18
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import KERNEL_ATTR, BIAS_ATTR
|
20
20
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
21
|
-
TargetPlatformModel
|
21
|
+
TargetPlatformModel, Signedness
|
22
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
23
23
|
AttributeQuantizationConfig
|
24
24
|
|
@@ -95,7 +95,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
95
95
|
quantization_preserving=False,
|
96
96
|
fixed_scale=None,
|
97
97
|
fixed_zero_point=None,
|
98
|
-
simd_size=32
|
98
|
+
simd_size=32,
|
99
|
+
signedness=Signedness.AUTO)
|
99
100
|
|
100
101
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
101
102
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -108,8 +109,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
108
109
|
quantization_preserving=False,
|
109
110
|
fixed_scale=None,
|
110
111
|
fixed_zero_point=None,
|
111
|
-
simd_size=None
|
112
|
-
|
112
|
+
simd_size=None,
|
113
|
+
signedness=Signedness.AUTO)
|
113
114
|
|
114
115
|
mixed_precision_cfg_list = [] # No mixed precision
|
115
116
|
|
@@ -18,7 +18,7 @@ import model_compression_toolkit as mct
|
|
18
18
|
from model_compression_toolkit.constants import FLOAT_BITWIDTH
|
19
19
|
from model_compression_toolkit.target_platform_capabilities.constants import BIAS_ATTR, KERNEL_ATTR
|
20
20
|
from model_compression_toolkit.target_platform_capabilities.target_platform import OpQuantizationConfig, \
|
21
|
-
TargetPlatformModel
|
21
|
+
TargetPlatformModel, Signedness
|
22
22
|
from model_compression_toolkit.target_platform_capabilities.target_platform.op_quantization_config import \
|
23
23
|
QuantizationMethod, AttributeQuantizationConfig
|
24
24
|
|
@@ -93,7 +93,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
93
93
|
quantization_preserving=False,
|
94
94
|
fixed_scale=None,
|
95
95
|
fixed_zero_point=None,
|
96
|
-
simd_size=32
|
96
|
+
simd_size=32,
|
97
|
+
signedness=Signedness.AUTO)
|
97
98
|
|
98
99
|
# We define an 8-bit config for linear operations quantization, that include a kernel and bias attributes.
|
99
100
|
linear_eight_bits = tp.OpQuantizationConfig(
|
@@ -106,8 +107,8 @@ def get_op_quantization_configs() -> Tuple[OpQuantizationConfig, List[OpQuantiza
|
|
106
107
|
quantization_preserving=False,
|
107
108
|
fixed_scale=None,
|
108
109
|
fixed_zero_point=None,
|
109
|
-
simd_size=None
|
110
|
-
|
110
|
+
simd_size=None,
|
111
|
+
signedness=Signedness.AUTO)
|
111
112
|
|
112
113
|
mixed_precision_cfg_list = [] # No mixed precision
|
113
114
|
|
File without changes
|
File without changes
|
{mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/mct_nightly.egg-info/SOURCES.txt
RENAMED
File without changes
|
File without changes
|
{mct-nightly-2.1.0.20240806.441 → mct-nightly-2.1.0.20240807.445}/mct_nightly.egg-info/top_level.txt
RENAMED
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|