mct-nightly 2.2.0.20240918.448__py3-none-any.whl → 2.2.0.20240919.455__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: mct-nightly
3
- Version: 2.2.0.20240918.448
3
+ Version: 2.2.0.20240919.455
4
4
  Summary: A Model Compression Toolkit for neural networks
5
5
  Home-page: UNKNOWN
6
6
  License: UNKNOWN
@@ -1,4 +1,4 @@
1
- model_compression_toolkit/__init__.py,sha256=mbgbcZTqzAzq-hfFdFzcbNZgPkm70zf0uPjjSnCRs4E,1573
1
+ model_compression_toolkit/__init__.py,sha256=C1njbo5gUl6G932-sOcU7MZgh1PntTBHD0seYNlVWRw,1573
2
2
  model_compression_toolkit/constants.py,sha256=i4wYheBkIdQmsQA-axIpcT3YiSO1USNc-jaNiNE8w6E,3920
3
3
  model_compression_toolkit/defaultdict.py,sha256=LSc-sbZYXENMCw3U9F4GiXuv67IKpdn0Qm7Fr11jy-4,2277
4
4
  model_compression_toolkit/logger.py,sha256=3DByV41XHRR3kLTJNbpaMmikL8icd9e1N-nkQAY9oDk,4567
@@ -508,7 +508,7 @@ model_compression_toolkit/trainable_infrastructure/pytorch/__init__.py,sha256=hu
508
508
  model_compression_toolkit/trainable_infrastructure/pytorch/annealing_schedulers.py,sha256=IdUBpZUcOXHLPp2OhwbO_Kytee3OTVuy2032N-tm694,1686
509
509
  model_compression_toolkit/trainable_infrastructure/pytorch/base_pytorch_quantizer.py,sha256=lWc5EG3ptrP85n69EHGKFkIadnrKEBMKnB5YXQ5AmXo,2745
510
510
  model_compression_toolkit/trainable_infrastructure/pytorch/quantizer_utils.py,sha256=1yOXKghUYfw2hmzbqTuNagIXBoM-wR2bP-ul66-mnDw,7767
511
- model_compression_toolkit/trainable_infrastructure/pytorch/util.py,sha256=4Qv_rkfxaDf0YeLD5I_7cepUk8OFsMNvUTrw9wFp_kU,1082
511
+ model_compression_toolkit/trainable_infrastructure/pytorch/util.py,sha256=oKuWi7E07a8zv5x9auhBugYE2RUQ7ojDh2XCs5koYJY,1090
512
512
  model_compression_toolkit/trainable_infrastructure/pytorch/activation_quantizers/__init__.py,sha256=73CXhqqNTvDpsvlJXclrGJq-vsCUYCI64ILu1y2mtvw,1056
513
513
  model_compression_toolkit/trainable_infrastructure/pytorch/activation_quantizers/base_activation_quantizer.py,sha256=X6E6mewWQot_aAkz3UxW5X0-Fjl_aMMjs3A-Af5eL6w,972
514
514
  model_compression_toolkit/trainable_infrastructure/pytorch/activation_quantizers/lsq/__init__.py,sha256=RAe8mgIr1V8dRIQtLf_dSG5zTUCKuQzxyybYx1dzEAs,697
@@ -550,8 +550,8 @@ tests_pytest/pytorch/gptq/test_annealing_cfg.py,sha256=hGC7L6mp3N1ygcJ3OctgS_Fz2
550
550
  tests_pytest/pytorch/gptq/test_gradual_act_quantization.py,sha256=tI01aFIUaiCILL5Qn--p1E_rLBUelxLdSY3k52lwcx0,4594
551
551
  tests_pytest/pytorch/trainable_infrastructure/__init__.py,sha256=RAe8mgIr1V8dRIQtLf_dSG5zTUCKuQzxyybYx1dzEAs,697
552
552
  tests_pytest/pytorch/trainable_infrastructure/test_linear_annealing.py,sha256=eNOpSp0GoLxtEdiRypBp8jaujXfdNxBwKh5Rd-P7WLs,1786
553
- mct_nightly-2.2.0.20240918.448.dist-info/LICENSE.md,sha256=aYSSIb-5AFPeITTvXm1UAoe0uYBiMmSS8flvXaaFUks,10174
554
- mct_nightly-2.2.0.20240918.448.dist-info/METADATA,sha256=Atg7fbRWZ1KvrHeOc1jaJ6Gb2VrUdnOAs9gKc_v26VU,20813
555
- mct_nightly-2.2.0.20240918.448.dist-info/WHEEL,sha256=eOLhNAGa2EW3wWl_TU484h7q1UNgy0JXjjoqKoxAAQc,92
556
- mct_nightly-2.2.0.20240918.448.dist-info/top_level.txt,sha256=csdfSXhtRnpWYRzjZ-dRLIhOmM2TEdVXUxG05A5fgb8,39
557
- mct_nightly-2.2.0.20240918.448.dist-info/RECORD,,
553
+ mct_nightly-2.2.0.20240919.455.dist-info/LICENSE.md,sha256=aYSSIb-5AFPeITTvXm1UAoe0uYBiMmSS8flvXaaFUks,10174
554
+ mct_nightly-2.2.0.20240919.455.dist-info/METADATA,sha256=EFLD1qEibUyY89Wz6xO9shdsPwMuV4Y1vP8SnKlWFh0,20813
555
+ mct_nightly-2.2.0.20240919.455.dist-info/WHEEL,sha256=eOLhNAGa2EW3wWl_TU484h7q1UNgy0JXjjoqKoxAAQc,92
556
+ mct_nightly-2.2.0.20240919.455.dist-info/top_level.txt,sha256=csdfSXhtRnpWYRzjZ-dRLIhOmM2TEdVXUxG05A5fgb8,39
557
+ mct_nightly-2.2.0.20240919.455.dist-info/RECORD,,
@@ -27,4 +27,4 @@ from model_compression_toolkit import data_generation
27
27
  from model_compression_toolkit import pruning
28
28
  from model_compression_toolkit.trainable_infrastructure.keras.load_model import keras_load_quantized_model
29
29
 
30
- __version__ = "2.2.0.20240918.000448"
30
+ __version__ = "2.2.0.20240919.000455"
@@ -12,13 +12,13 @@
12
12
  # See the License for the specific language governing permissions and
13
13
  # limitations under the License.
14
14
  # ==============================================================================
15
- from functools import cache
15
+ from functools import lru_cache
16
16
  from typing import Callable
17
17
 
18
18
  from tqdm import tqdm
19
19
 
20
20
 
21
- @cache
21
+ @lru_cache
22
22
  def get_total_grad_steps(representative_data_gen: Callable) -> int:
23
23
  # dry run on the representative dataset to count number of batches
24
24
  num_batches = 0