mct-nightly 2.3.0.20250423.537__py3-none-any.whl → 2.3.0.20250424.534__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: mct-nightly
3
- Version: 2.3.0.20250423.537
3
+ Version: 2.3.0.20250424.534
4
4
  Summary: A Model Compression Toolkit for neural networks
5
5
  Classifier: Programming Language :: Python :: 3
6
6
  Classifier: License :: OSI Approved :: Apache Software License
@@ -34,7 +34,7 @@ Dynamic: summary
34
34
  <div align="center" markdown="1">
35
35
  <p>
36
36
  <a href="https://sony.github.io/model_optimization/" target="_blank">
37
- <img src="https://github.com/sony/model_optimization/blob/main/docsrc/images/mctHeader1-cropped.svg" width="1000"></a>
37
+ <img src="https://raw.githubusercontent.com/sony/model_optimization/refs/heads/main/docsrc/images/mctHeader1-cropped.svg" width="1000"></a>
38
38
  </p>
39
39
 
40
40
  ______________________________________________________________________
@@ -100,7 +100,7 @@ For further details, please see [Supported features and algorithms](#high-level-
100
100
  <div align="center">
101
101
  <p align="center">
102
102
 
103
- <img src="https://github.com/sony/model_optimization/blob/main/docsrc/images/mctDiagram_clean.svg" width="800">
103
+ <img src="https://raw.githubusercontent.com/sony/model_optimization/refs/heads/main/docsrc/images/mctDiagram_clean.svg" width="800">
104
104
  </p>
105
105
  </div>
106
106
 
@@ -181,16 +181,16 @@ Currently, MCT is being tested on various Python, Pytorch and TensorFlow version
181
181
  ## <div align="center">Results</div>
182
182
 
183
183
  <p align="center">
184
- <img src="https://github.com/sony/model_optimization/blob/main/docsrc/images/Classification.png" width="200">
185
- <img src="https://github.com/sony/model_optimization/blob/main/docsrc/images/SemSeg.png" width="200">
186
- <img src="https://github.com/sony/model_optimization/blob/main/docsrc/images/PoseEst.png" width="200">
187
- <img src="https://github.com/sony/model_optimization/blob/main/docsrc/images/ObjDet.png" width="200">
184
+ <img src="https://raw.githubusercontent.com/sony/model_optimization/refs/heads/main/docsrc/images/Classification.png" width="200">
185
+ <img src="https://raw.githubusercontent.com/sony/model_optimization/refs/heads/main/docsrc/images/SemSeg.png" width="200">
186
+ <img src="https://raw.githubusercontent.com/sony/model_optimization/refs/heads/main/docsrc/images/PoseEst.png" width="200">
187
+ <img src="https://raw.githubusercontent.com/sony/model_optimization/refs/heads/main/docsrc/images/ObjDet.png" width="200">
188
188
 
189
189
  MCT can quantize an existing 32-bit floating-point model to an 8-bit fixed-point (or less) model without compromising accuracy.
190
190
  Below is a graph of [MobileNetV2](https://pytorch.org/vision/main/models/generated/torchvision.models.mobilenet_v2.html) accuracy on ImageNet vs average bit-width of weights (X-axis), using **single-precision** quantization, **mixed-precision** quantization, and mixed-precision quantization with GPTQ.
191
191
 
192
192
  <p align="center">
193
- <img src="https://github.com/sony/model_optimization/blob/main/docsrc/images/torch_mobilenetv2.png" width="800">
193
+ <img src="https://raw.githubusercontent.com/sony/model_optimization/refs/heads/main/docsrc/images/torch_mobilenetv2.png" width="800">
194
194
 
195
195
  For more results, please see [1]
196
196
 
@@ -1,5 +1,5 @@
1
- mct_nightly-2.3.0.20250423.537.dist-info/licenses/LICENSE.md,sha256=aYSSIb-5AFPeITTvXm1UAoe0uYBiMmSS8flvXaaFUks,10174
2
- model_compression_toolkit/__init__.py,sha256=NXqhvuAEHQlzpdJpTtyy1rDJW2gxmMwEGZvHIBQE_f0,1557
1
+ mct_nightly-2.3.0.20250424.534.dist-info/licenses/LICENSE.md,sha256=aYSSIb-5AFPeITTvXm1UAoe0uYBiMmSS8flvXaaFUks,10174
2
+ model_compression_toolkit/__init__.py,sha256=HmFpViJmJPVcQg5km-gnodqRgdt3lc5eqANLwoWrMqM,1557
3
3
  model_compression_toolkit/constants.py,sha256=iJ6vfTjC2oFIZWt8wvHoxEw5YJi3yl0Hd4q30_8q0Zc,3958
4
4
  model_compression_toolkit/defaultdict.py,sha256=LSc-sbZYXENMCw3U9F4GiXuv67IKpdn0Qm7Fr11jy-4,2277
5
5
  model_compression_toolkit/logger.py,sha256=L3q7tn3Uht0i_7phnlOWMR2Te2zvzrt2HOz9vYEInts,4529
@@ -101,7 +101,7 @@ model_compression_toolkit/core/common/pruning/mask/__init__.py,sha256=huHoBUcKNB
101
101
  model_compression_toolkit/core/common/pruning/mask/per_channel_mask.py,sha256=77DB1vqq_gHwbUjeCHRaq1Q-V4wEtdVdwkGezcZgToA,5021
102
102
  model_compression_toolkit/core/common/pruning/mask/per_simd_group_mask.py,sha256=_LcDAxLeC5I0KdMHS8jib5XxIKO2ZLavXYuSMIPIQBo,5868
103
103
  model_compression_toolkit/core/common/quantization/__init__.py,sha256=sw7LOPN1bM82o3SkMaklyH0jw-TLGK0-fl2Wq73rffI,697
104
- model_compression_toolkit/core/common/quantization/bit_width_config.py,sha256=0HA3CIZW-ZrA55ra-yJXRvAYnoR8i1SjpbnMDKcWYNQ,12819
104
+ model_compression_toolkit/core/common/quantization/bit_width_config.py,sha256=034kgwe0ydyLXsV83KqxKyyHkoUQH06ai0leLyg0p8I,13019
105
105
  model_compression_toolkit/core/common/quantization/candidate_node_quantization_config.py,sha256=lyWPvnoX8BmulhLKR20r5gT2_Yan7P40d8EcgDhErPk,4905
106
106
  model_compression_toolkit/core/common/quantization/core_config.py,sha256=yxCzWqldcHoe8GGxrH0tp99bhrc5jDT7SgZftnMUUBE,2374
107
107
  model_compression_toolkit/core/common/quantization/debug_config.py,sha256=uH45Uq3Tp9FIyMynex_WY2_y-Kv8LuPw2XXZydnpW5A,1649
@@ -528,7 +528,7 @@ model_compression_toolkit/xquant/pytorch/model_analyzer.py,sha256=b93o800yVB3Z-i
528
528
  model_compression_toolkit/xquant/pytorch/pytorch_report_utils.py,sha256=UVN_S9ULHBEldBpShCOt8-soT8YTQ5oE362y96qF_FA,3950
529
529
  model_compression_toolkit/xquant/pytorch/similarity_functions.py,sha256=CERxq5K8rqaiE-DlwhZBTUd9x69dtYJlkHOPLB54vm8,2354
530
530
  model_compression_toolkit/xquant/pytorch/tensorboard_utils.py,sha256=mkoEktLFFHtEKzzFRn_jCnxjhJolK12TZ5AQeDHzUO8,9767
531
- mct_nightly-2.3.0.20250423.537.dist-info/METADATA,sha256=PeCeasbP-z1tGgn1g1tpCnmFWMlSQ7HtIfby4JXgX68,25413
532
- mct_nightly-2.3.0.20250423.537.dist-info/WHEEL,sha256=pxyMxgL8-pra_rKaQ4drOZAegBVuX-G_4nRHjjgWbmo,91
533
- mct_nightly-2.3.0.20250423.537.dist-info/top_level.txt,sha256=gsYA8juk0Z-ZmQRKULkb3JLGdOdz8jW_cMRjisn9ga4,26
534
- mct_nightly-2.3.0.20250423.537.dist-info/RECORD,,
531
+ mct_nightly-2.3.0.20250424.534.dist-info/METADATA,sha256=wMqM0-nGTBa189h4xpdr-iY2-QUxlm1vVnXkB7ogmzU,25560
532
+ mct_nightly-2.3.0.20250424.534.dist-info/WHEEL,sha256=SmOxYU7pzNKBqASvQJ7DjX3XGUF92lrGhMb3R6_iiqI,91
533
+ mct_nightly-2.3.0.20250424.534.dist-info/top_level.txt,sha256=gsYA8juk0Z-ZmQRKULkb3JLGdOdz8jW_cMRjisn9ga4,26
534
+ mct_nightly-2.3.0.20250424.534.dist-info/RECORD,,
@@ -1,5 +1,5 @@
1
1
  Wheel-Version: 1.0
2
- Generator: setuptools (79.0.0)
2
+ Generator: setuptools (79.0.1)
3
3
  Root-Is-Purelib: true
4
4
  Tag: py3-none-any
5
5
 
@@ -27,4 +27,4 @@ from model_compression_toolkit import data_generation
27
27
  from model_compression_toolkit import pruning
28
28
  from model_compression_toolkit.trainable_infrastructure.keras.load_model import keras_load_quantized_model
29
29
 
30
- __version__ = "2.3.0.20250423.000537"
30
+ __version__ = "2.3.0.20250424.000534"
@@ -20,6 +20,8 @@ from model_compression_toolkit.core.common.matchers.node_matcher import BaseNode
20
20
  from model_compression_toolkit.logger import Logger
21
21
 
22
22
  from model_compression_toolkit.core.common.graph.base_node import WeightAttrT
23
+ from model_compression_toolkit.target_platform_capabilities.constants import POS_ATTR
24
+
23
25
 
24
26
  @dataclass
25
27
  class ManualBitWidthSelection:
@@ -221,9 +223,10 @@ class BitWidthConfig:
221
223
  if isinstance(attr_str, str) and isinstance(manual_bit_width_selection.attr, str):
222
224
  if attr_str.find(manual_bit_width_selection.attr) != -1:
223
225
  attr.append(attr_str)
224
- elif isinstance(attr_str, int) and isinstance(manual_bit_width_selection.attr, int):
225
- if attr_str == manual_bit_width_selection.attr:
226
- attr.append(attr_str)
226
+ # this is a positional attribute, so it needs to be handled separately.
227
+ # Search manual_bit_width_selection's attribute that contain the POS_ATTR string.
228
+ elif isinstance(attr_str, int) and POS_ATTR in manual_bit_width_selection.attr:
229
+ attr.append(POS_ATTR)
227
230
  if len(attr) == 0:
228
231
  Logger.critical(f'The requested attribute {manual_bit_width_selection.attr} to change the bit width for {n} does not exist.')
229
232