ai-edge-torch-nightly 0.6.0.dev20250529__py3-none-any.whl → 0.6.0.dev20250531__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- ai_edge_torch/generative/examples/amd_llama_135m/verify_util.py +6 -6
- ai_edge_torch/generative/examples/deepseek/verify_util.py +6 -6
- ai_edge_torch/generative/examples/llama/verify_util.py +6 -6
- ai_edge_torch/generative/examples/openelm/verify_util.py +6 -6
- ai_edge_torch/generative/examples/phi/verify_util.py +7 -6
- ai_edge_torch/generative/examples/qwen/verify_util.py +7 -6
- ai_edge_torch/generative/examples/smollm/verify_util.py +7 -7
- ai_edge_torch/generative/examples/tiny_llama/verify_util.py +7 -6
- ai_edge_torch/generative/layers/model_config.py +0 -2
- ai_edge_torch/generative/utilities/loader.py +1 -2
- ai_edge_torch/model.py +2 -1
- ai_edge_torch/version.py +1 -1
- {ai_edge_torch_nightly-0.6.0.dev20250529.dist-info → ai_edge_torch_nightly-0.6.0.dev20250531.dist-info}/METADATA +1 -1
- {ai_edge_torch_nightly-0.6.0.dev20250529.dist-info → ai_edge_torch_nightly-0.6.0.dev20250531.dist-info}/RECORD +17 -17
- {ai_edge_torch_nightly-0.6.0.dev20250529.dist-info → ai_edge_torch_nightly-0.6.0.dev20250531.dist-info}/LICENSE +0 -0
- {ai_edge_torch_nightly-0.6.0.dev20250529.dist-info → ai_edge_torch_nightly-0.6.0.dev20250531.dist-info}/WHEEL +0 -0
- {ai_edge_torch_nightly-0.6.0.dev20250529.dist-info → ai_edge_torch_nightly-0.6.0.dev20250531.dist-info}/top_level.txt +0 -0
@@ -16,11 +16,13 @@
|
|
16
16
|
import logging
|
17
17
|
import os
|
18
18
|
import pathlib
|
19
|
+
from typing import Callable, Dict
|
19
20
|
|
20
21
|
from ai_edge_torch.generative.examples.amd_llama_135m import amd_llama_135m
|
21
22
|
from ai_edge_torch.generative.utilities import loader
|
22
23
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
24
|
from ai_edge_torch.generative.utilities import verifier
|
25
|
+
import torch
|
24
26
|
import transformers
|
25
27
|
|
26
28
|
|
@@ -31,8 +33,9 @@ def verify_amd_llama_135m(
|
|
31
33
|
checkpoint_dir: str,
|
32
34
|
weight_filename: str = "model.safetensors",
|
33
35
|
max_new_tokens: int = 30,
|
34
|
-
initialize_from_local: bool = True,
|
35
36
|
prompts: list[str] | None = None,
|
37
|
+
initialize_from_local: bool = True,
|
38
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
36
39
|
) -> bool:
|
37
40
|
"""Verifies the reauthored AMD-Llama-135M model with a custom loader."""
|
38
41
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -41,11 +44,8 @@ def verify_amd_llama_135m(
|
|
41
44
|
)
|
42
45
|
|
43
46
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
44
|
-
custom_loader
|
45
|
-
|
46
|
-
if initialize_from_local
|
47
|
-
else loader.get_custom_loader("", "safetensors")
|
48
|
-
)
|
47
|
+
if custom_loader is None and not initialize_from_local:
|
48
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
49
49
|
|
50
50
|
if initialize_from_local:
|
51
51
|
# Locate the cached dir.
|
@@ -16,11 +16,13 @@
|
|
16
16
|
import logging
|
17
17
|
import os
|
18
18
|
import pathlib
|
19
|
+
from typing import Callable, Dict
|
19
20
|
|
20
21
|
from ai_edge_torch.generative.examples.deepseek import deepseek
|
21
22
|
from ai_edge_torch.generative.utilities import loader
|
22
23
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
24
|
from ai_edge_torch.generative.utilities import verifier
|
25
|
+
import torch
|
24
26
|
import transformers
|
25
27
|
|
26
28
|
|
@@ -31,8 +33,9 @@ def verify_deepseek_r1_distill_1_5b(
|
|
31
33
|
checkpoint_dir: str,
|
32
34
|
weight_filename: str = "model.safetensors",
|
33
35
|
max_new_tokens: int = 30,
|
34
|
-
initialize_from_local: bool = True,
|
35
36
|
prompts: list[str] | None = None,
|
37
|
+
initialize_from_local: bool = True,
|
38
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
36
39
|
) -> bool:
|
37
40
|
"""Verifies the reauthored DeepSeek R1 distilled 1.5B model with a custom loader."""
|
38
41
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -41,11 +44,8 @@ def verify_deepseek_r1_distill_1_5b(
|
|
41
44
|
)
|
42
45
|
|
43
46
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
44
|
-
custom_loader
|
45
|
-
|
46
|
-
if initialize_from_local
|
47
|
-
else loader.get_custom_loader("", "safetensors")
|
48
|
-
)
|
47
|
+
if custom_loader is None and not initialize_from_local:
|
48
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
49
49
|
|
50
50
|
if initialize_from_local:
|
51
51
|
# Locate the cached dir.
|
@@ -16,11 +16,13 @@
|
|
16
16
|
import logging
|
17
17
|
import os
|
18
18
|
import pathlib
|
19
|
+
from typing import Callable, Dict
|
19
20
|
|
20
21
|
from ai_edge_torch.generative.examples.llama import llama
|
21
22
|
from ai_edge_torch.generative.utilities import loader
|
22
23
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
24
|
from ai_edge_torch.generative.utilities import verifier
|
25
|
+
import torch
|
24
26
|
import transformers
|
25
27
|
|
26
28
|
_BUILDER = {
|
@@ -36,8 +38,9 @@ def verify_llama_3_2(
|
|
36
38
|
checkpoint_dir: str,
|
37
39
|
weight_filename: str = "model.safetensors",
|
38
40
|
max_new_tokens: int = 30,
|
39
|
-
initialize_from_local: bool = True,
|
40
41
|
prompts: list[str] | None = None,
|
42
|
+
initialize_from_local: bool = True,
|
43
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
41
44
|
) -> bool:
|
42
45
|
"""Verifies the reauthored Llama 3.2 model with a custom loader."""
|
43
46
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -46,11 +49,8 @@ def verify_llama_3_2(
|
|
46
49
|
)
|
47
50
|
|
48
51
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
49
|
-
custom_loader
|
50
|
-
|
51
|
-
if initialize_from_local
|
52
|
-
else loader.get_custom_loader("", "safetensors")
|
53
|
-
)
|
52
|
+
if custom_loader is None and not initialize_from_local:
|
53
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
54
54
|
|
55
55
|
if initialize_from_local:
|
56
56
|
# Locate the cached dir.
|
@@ -16,11 +16,13 @@
|
|
16
16
|
import logging
|
17
17
|
import os
|
18
18
|
import pathlib
|
19
|
+
from typing import Callable, Dict
|
19
20
|
|
20
21
|
from ai_edge_torch.generative.examples.openelm import openelm
|
21
22
|
from ai_edge_torch.generative.utilities import loader
|
22
23
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
24
|
from ai_edge_torch.generative.utilities import verifier
|
25
|
+
import torch
|
24
26
|
import transformers
|
25
27
|
|
26
28
|
|
@@ -31,8 +33,9 @@ def verify_openelm(
|
|
31
33
|
checkpoint_dir: str,
|
32
34
|
weight_filename: str = "model.safetensors",
|
33
35
|
max_new_tokens: int = 30,
|
34
|
-
initialize_from_local: bool = True,
|
35
36
|
prompts: list[str] | None = None,
|
37
|
+
initialize_from_local: bool = True,
|
38
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
36
39
|
) -> bool:
|
37
40
|
"""Verifies the reauthored OpenELM model with a custom loader."""
|
38
41
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -41,11 +44,8 @@ def verify_openelm(
|
|
41
44
|
)
|
42
45
|
|
43
46
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
44
|
-
custom_loader
|
45
|
-
|
46
|
-
if initialize_from_local
|
47
|
-
else loader.get_custom_loader("", "safetensors")
|
48
|
-
)
|
47
|
+
if custom_loader is None and not initialize_from_local:
|
48
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
49
49
|
|
50
50
|
if initialize_from_local:
|
51
51
|
# Locate the cached dir.
|
@@ -13,14 +13,17 @@
|
|
13
13
|
# limitations under the License.
|
14
14
|
# ==============================================================================
|
15
15
|
"""Utils for verifying the Phi model."""
|
16
|
+
|
16
17
|
import logging
|
17
18
|
import os
|
18
19
|
import pathlib
|
20
|
+
from typing import Callable, Dict
|
19
21
|
|
20
22
|
from ai_edge_torch.generative.examples.phi import phi2, phi3, phi4
|
21
23
|
from ai_edge_torch.generative.utilities import loader
|
22
24
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
25
|
from ai_edge_torch.generative.utilities import verifier
|
26
|
+
import torch
|
24
27
|
import transformers
|
25
28
|
|
26
29
|
|
@@ -38,9 +41,10 @@ def verify_phi(
|
|
38
41
|
checkpoint_dir: str,
|
39
42
|
weight_filename: str = "model.safetensors",
|
40
43
|
max_new_tokens: int = 30,
|
41
|
-
initialize_from_local: bool = True,
|
42
44
|
prompts: list[str] | None = None,
|
43
45
|
atol: float = 1e-04,
|
46
|
+
initialize_from_local: bool = True,
|
47
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
44
48
|
) -> bool:
|
45
49
|
"""Verifies the reauthored Phi model with a custom loader."""
|
46
50
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -49,11 +53,8 @@ def verify_phi(
|
|
49
53
|
)
|
50
54
|
|
51
55
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
52
|
-
custom_loader
|
53
|
-
|
54
|
-
if initialize_from_local
|
55
|
-
else loader.get_custom_loader("", "safetensors")
|
56
|
-
)
|
56
|
+
if custom_loader is None and not initialize_from_local:
|
57
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
57
58
|
|
58
59
|
if initialize_from_local:
|
59
60
|
# Locate the cached dir.
|
@@ -13,14 +13,17 @@
|
|
13
13
|
# limitations under the License.
|
14
14
|
# ==============================================================================
|
15
15
|
"""Utils for verifying the Qwen model."""
|
16
|
+
|
16
17
|
import logging
|
17
18
|
import os
|
18
19
|
import pathlib
|
20
|
+
from typing import Callable, Dict
|
19
21
|
|
20
22
|
from ai_edge_torch.generative.examples.qwen import qwen, qwen3
|
21
23
|
from ai_edge_torch.generative.utilities import loader
|
22
24
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
25
|
from ai_edge_torch.generative.utilities import verifier
|
26
|
+
import torch
|
24
27
|
import transformers
|
25
28
|
|
26
29
|
|
@@ -50,8 +53,9 @@ def verify_qwen(
|
|
50
53
|
checkpoint_dir: str,
|
51
54
|
weight_filename: str = "model.safetensors",
|
52
55
|
max_new_tokens: int = 30,
|
53
|
-
initialize_from_local: bool = True,
|
54
56
|
prompts: list[str] | None = None,
|
57
|
+
initialize_from_local: bool = True,
|
58
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
55
59
|
) -> bool:
|
56
60
|
"""Verifies the reauthored Llama 3.2 model with a custom loader."""
|
57
61
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -60,11 +64,8 @@ def verify_qwen(
|
|
60
64
|
)
|
61
65
|
|
62
66
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
63
|
-
custom_loader
|
64
|
-
|
65
|
-
if initialize_from_local
|
66
|
-
else loader.get_custom_loader("", "safetensors")
|
67
|
-
)
|
67
|
+
if custom_loader is None and not initialize_from_local:
|
68
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
68
69
|
|
69
70
|
if initialize_from_local:
|
70
71
|
# Locate the cached dir.
|
@@ -16,11 +16,12 @@
|
|
16
16
|
import logging
|
17
17
|
import os
|
18
18
|
import pathlib
|
19
|
-
|
19
|
+
from typing import Callable, Dict
|
20
20
|
from ai_edge_torch.generative.examples.smollm import smollm
|
21
21
|
from ai_edge_torch.generative.utilities import loader
|
22
22
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
23
|
from ai_edge_torch.generative.utilities import verifier
|
24
|
+
import torch
|
24
25
|
import transformers
|
25
26
|
|
26
27
|
_BUILDER = {
|
@@ -36,8 +37,9 @@ def verify_smollm_135m(
|
|
36
37
|
checkpoint_dir: str,
|
37
38
|
weight_filename: str = "model.safetensors",
|
38
39
|
max_new_tokens: int = 30,
|
39
|
-
initialize_from_local: bool = True,
|
40
40
|
prompts: list[str] | None = None,
|
41
|
+
initialize_from_local: bool = True,
|
42
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
41
43
|
) -> bool:
|
42
44
|
"""Verifies the reauthored SmoLLM model with a custom loader."""
|
43
45
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -46,11 +48,9 @@ def verify_smollm_135m(
|
|
46
48
|
)
|
47
49
|
|
48
50
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
49
|
-
|
50
|
-
|
51
|
-
|
52
|
-
else loader.get_custom_loader("", "safetensors")
|
53
|
-
)
|
51
|
+
|
52
|
+
if custom_loader is None and not initialize_from_local:
|
53
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
54
54
|
|
55
55
|
if initialize_from_local:
|
56
56
|
# Locate the cached dir.
|
@@ -13,14 +13,17 @@
|
|
13
13
|
# limitations under the License.
|
14
14
|
# ==============================================================================
|
15
15
|
"""Utils for verifying the TinyLlama model."""
|
16
|
+
|
16
17
|
import logging
|
17
18
|
import os
|
18
19
|
import pathlib
|
20
|
+
from typing import Callable, Dict
|
19
21
|
|
20
22
|
from ai_edge_torch.generative.examples.tiny_llama import tiny_llama
|
21
23
|
from ai_edge_torch.generative.utilities import loader
|
22
24
|
from ai_edge_torch.generative.utilities import transformers_verifier
|
23
25
|
from ai_edge_torch.generative.utilities import verifier
|
26
|
+
import torch
|
24
27
|
import transformers
|
25
28
|
|
26
29
|
|
@@ -31,8 +34,9 @@ def verify_tiny_llama(
|
|
31
34
|
checkpoint_dir: str,
|
32
35
|
weight_filename: str = "model.safetensors",
|
33
36
|
max_new_tokens: int = 30,
|
34
|
-
initialize_from_local: bool = True,
|
35
37
|
prompts: list[str] | None = None,
|
38
|
+
initialize_from_local: bool = True,
|
39
|
+
custom_loader: Callable[[str], Dict[str, torch.Tensor]] | None = None,
|
36
40
|
) -> bool:
|
37
41
|
"""Verifies the reauthored TinyLlama model with a custom loader."""
|
38
42
|
logging.info("Loading the original model from: %s", checkpoint_dir)
|
@@ -41,11 +45,8 @@ def verify_tiny_llama(
|
|
41
45
|
)
|
42
46
|
|
43
47
|
logging.info("Building the reauthored model from: %s", checkpoint_dir)
|
44
|
-
custom_loader
|
45
|
-
|
46
|
-
if initialize_from_local
|
47
|
-
else loader.get_custom_loader("", "safetensors")
|
48
|
-
)
|
48
|
+
if custom_loader is None and not initialize_from_local:
|
49
|
+
custom_loader = loader.get_custom_loader("", "safetensors")
|
49
50
|
|
50
51
|
if initialize_from_local:
|
51
52
|
# Locate the cached dir.
|
@@ -134,8 +134,6 @@ class AttentionConfig:
|
|
134
134
|
value_norm_config: NormalizationConfig = dataclasses.field(
|
135
135
|
default_factory=NormalizationConfig
|
136
136
|
)
|
137
|
-
# Whether the KV cache is shared with the previous attention block.
|
138
|
-
kv_shared: bool = False
|
139
137
|
relative_attention_num_buckets: int = 0
|
140
138
|
relative_attention_max_distance: int = 0
|
141
139
|
# Softcap on the output logits.
|
@@ -23,14 +23,13 @@ from safetensors import safe_open
|
|
23
23
|
from safetensors.torch import load_file
|
24
24
|
import torch
|
25
25
|
|
26
|
-
|
27
26
|
def get_custom_loader(
|
28
27
|
checkpoint_path: str,
|
29
28
|
checkpoint_format: Optional[str] = None,
|
30
29
|
) -> Callable[[str], Dict[str, torch.Tensor]]:
|
31
30
|
"""Returns a custom loader for the given checkpoint path.
|
32
31
|
|
33
|
-
Those
|
32
|
+
Those custome loaders can either support state dictionary or safetensors, and
|
34
33
|
the actual data might be fetched from a remote source.
|
35
34
|
|
36
35
|
Args:
|
ai_edge_torch/model.py
CHANGED
@@ -155,7 +155,8 @@ class TfLiteModel(Model):
|
|
155
155
|
Args:
|
156
156
|
path: The path to file to which the model is serialized.
|
157
157
|
"""
|
158
|
-
os.
|
158
|
+
if os.path.dirname(path):
|
159
|
+
os.makedirs(os.path.dirname(path), exist_ok=True)
|
159
160
|
with open(path, 'wb') as file_handle:
|
160
161
|
file_handle.write(self._tflite_model)
|
161
162
|
|
ai_edge_torch/version.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.1
|
2
2
|
Name: ai-edge-torch-nightly
|
3
|
-
Version: 0.6.0.
|
3
|
+
Version: 0.6.0.dev20250531
|
4
4
|
Summary: Supporting PyTorch models with the Google AI Edge TFLite runtime.
|
5
5
|
Home-page: https://github.com/google-ai-edge/ai-edge-torch
|
6
6
|
Keywords: On-Device ML,AI,Google,TFLite,PyTorch,LLMs,GenAI
|
@@ -1,8 +1,8 @@
|
|
1
1
|
ai_edge_torch/__init__.py,sha256=lemyLCNoGYRnJsmDuGZu7qOqLbLqG6CGDFtu3ue1syU,1290
|
2
2
|
ai_edge_torch/_config.py,sha256=AiqhbcheF7j_ozIGDLC89k1we95aVgFDa-tR6h7UI0s,2529
|
3
3
|
ai_edge_torch/conftest.py,sha256=r0GTrhMRhlmOGrrkvumHN8hkmyug6WvF60vWq8wRIBI,758
|
4
|
-
ai_edge_torch/model.py,sha256=
|
5
|
-
ai_edge_torch/version.py,sha256=
|
4
|
+
ai_edge_torch/model.py,sha256=A7loFu8jE9CsXsfMmHYZ-KDFJiaD8Kkqwm_9d3IVzk0,5638
|
5
|
+
ai_edge_torch/version.py,sha256=Z8PdNXw5sVjFtOY2IFehsJi1BwTO9Mm-mSE2MZWv8Hg,806
|
6
6
|
ai_edge_torch/_convert/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIXY_dRQeYudjsrKGf6LZz65g,671
|
7
7
|
ai_edge_torch/_convert/conversion.py,sha256=iQk3R-pLq4c1nfLqPB4xTRj78gghxPGzJCJtILLdg5o,6123
|
8
8
|
ai_edge_torch/_convert/conversion_utils.py,sha256=Sr8qXVcTwc-ZnZmK7yxVrIOOp1S_vNrwzC0zUvLTI2o,2160
|
@@ -56,12 +56,12 @@ ai_edge_torch/generative/examples/amd_llama_135m/__init__.py,sha256=hHLluseD2R0H
|
|
56
56
|
ai_edge_torch/generative/examples/amd_llama_135m/amd_llama_135m.py,sha256=PTKcl-CHQnzExQSfrwG9YC0KPc8zomG7WlPabXtZLx4,2910
|
57
57
|
ai_edge_torch/generative/examples/amd_llama_135m/convert_to_tflite.py,sha256=s2f5TJos6rSgogqeFk0qsOpI30qsR04umk9hAAZ5918,1782
|
58
58
|
ai_edge_torch/generative/examples/amd_llama_135m/verify.py,sha256=uyBg5-trxQEjEHDZMX4qojkcsZgERUiPqIgR9n0_AY4,1311
|
59
|
-
ai_edge_torch/generative/examples/amd_llama_135m/verify_util.py,sha256=
|
59
|
+
ai_edge_torch/generative/examples/amd_llama_135m/verify_util.py,sha256=OJTjULdz-8LVd5DhmX8isxUI1PzhE-dQLj7--5DHfJ8,2961
|
60
60
|
ai_edge_torch/generative/examples/deepseek/__init__.py,sha256=JaAnrFoXTl3RJX97XspklkTyqOHVyAgRJsZtzNDd10c,671
|
61
61
|
ai_edge_torch/generative/examples/deepseek/convert_to_tflite.py,sha256=xTPfT3Mt_4bMfGkrqDKatLecZOuaE0WhxXs3uAsO_uU,1749
|
62
62
|
ai_edge_torch/generative/examples/deepseek/deepseek.py,sha256=9gUnK1IOifQyYpm03f64Mzg-afwbYY9kVWz6-ynq8zY,3014
|
63
63
|
ai_edge_torch/generative/examples/deepseek/verify.py,sha256=HkvgEyGb-V_f6mWfyeN7Ai5uADAVQNzWvkygaKJiLAc,1344
|
64
|
-
ai_edge_torch/generative/examples/deepseek/verify_util.py,sha256=
|
64
|
+
ai_edge_torch/generative/examples/deepseek/verify_util.py,sha256=jui_16J0C0VhICGBJDiO7Br5l5QCrhm-AohXYuUyyqQ,2971
|
65
65
|
ai_edge_torch/generative/examples/gemma/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIXY_dRQeYudjsrKGf6LZz65g,671
|
66
66
|
ai_edge_torch/generative/examples/gemma/convert_gemma1_to_tflite.py,sha256=t2qZTjyM2imPenb14fzbQ-CHj5Cejw4M5xfEZpgX6Uc,1748
|
67
67
|
ai_edge_torch/generative/examples/gemma/convert_gemma2_to_tflite.py,sha256=Yj-b4S9BNxArnGjruRIymCiWrlf7ZvwiG6keTVGldk4,1816
|
@@ -86,7 +86,7 @@ ai_edge_torch/generative/examples/llama/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIX
|
|
86
86
|
ai_edge_torch/generative/examples/llama/convert_to_tflite.py,sha256=4qnMyvJHqhqf9k01wEsO23BKo6tSy2KD7sHdTGimKGg,1957
|
87
87
|
ai_edge_torch/generative/examples/llama/llama.py,sha256=eWPFnuSxhjuk5XZmvtndu_Z1-e9NlZg7-uFfiOqJXfw,6952
|
88
88
|
ai_edge_torch/generative/examples/llama/verify.py,sha256=XoF_-kxdryjt0Bt_YeHnIbLfjwFxSVioTSEG75moDr8,1581
|
89
|
-
ai_edge_torch/generative/examples/llama/verify_util.py,sha256=
|
89
|
+
ai_edge_torch/generative/examples/llama/verify_util.py,sha256=JIAkG7TEaR_5scibyM-zgS9G2KrIfUBgX4vHDJwmaBo,3019
|
90
90
|
ai_edge_torch/generative/examples/moonshine/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIXY_dRQeYudjsrKGf6LZz65g,671
|
91
91
|
ai_edge_torch/generative/examples/moonshine/convert_moonshine_to_tflite.py,sha256=_GkaSkregS3NWN38UGXxj4pED5gtQGaaPZx5_CZ0TVM,1657
|
92
92
|
ai_edge_torch/generative/examples/moonshine/moonshine.py,sha256=nZ2b8u4TmsB5sgdClgAuH8E78bcTv9RCnF9666HqP2M,3394
|
@@ -94,7 +94,7 @@ ai_edge_torch/generative/examples/openelm/__init__.py,sha256=hHLluseD2R0Hh4W6XZR
|
|
94
94
|
ai_edge_torch/generative/examples/openelm/convert_to_tflite.py,sha256=S7OP8PJcOQbm8AHvi_Tc3qnQuVOtjMFNlwaZQ_oirUM,1747
|
95
95
|
ai_edge_torch/generative/examples/openelm/openelm.py,sha256=R_E0hXsg6l8ANEgBBy0R8egz3p4ONJvBmPWs6sXx63M,4692
|
96
96
|
ai_edge_torch/generative/examples/openelm/verify.py,sha256=kRoNEUEsrz51PFSeTPcrYsPBQRLtUmYM3t_-Jl0oFqM,1300
|
97
|
-
ai_edge_torch/generative/examples/openelm/verify_util.py,sha256=
|
97
|
+
ai_edge_torch/generative/examples/openelm/verify_util.py,sha256=arX-ZGWVFfsyZ7n_V8ivaw8fqpmK3CmcxfsS0b1pOBY,2913
|
98
98
|
ai_edge_torch/generative/examples/paligemma/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIXY_dRQeYudjsrKGf6LZz65g,671
|
99
99
|
ai_edge_torch/generative/examples/paligemma/convert_to_tflite.py,sha256=Fl4k-lcpiUaJS0A1E7HVVUW7iTcZAU4FbA4KcSkO5SQ,2212
|
100
100
|
ai_edge_torch/generative/examples/paligemma/decoder.py,sha256=GLlfbJr3ZIzmH643IwXyrG54qKEYMPRsvhU6gXXi7yg,5490
|
@@ -115,7 +115,7 @@ ai_edge_torch/generative/examples/phi/phi4.py,sha256=ZHA0Rq7ifgxiHC_8PJf-y7WCA7i
|
|
115
115
|
ai_edge_torch/generative/examples/phi/verify.py,sha256=fIWgqypLQ3uOQ1u5uuklYiQSJPhKCTYIBACjrp7DbMA,1346
|
116
116
|
ai_edge_torch/generative/examples/phi/verify_phi3.py,sha256=TwIu2xUPQyMUTFdz29E2y75wfq4c1fGJnT3QfA3eS1s,1347
|
117
117
|
ai_edge_torch/generative/examples/phi/verify_phi4.py,sha256=2MlgQrfRkhE7Dya8MIixGwpqEZYdPjQkUGB47Mt1hSI,1343
|
118
|
-
ai_edge_torch/generative/examples/phi/verify_util.py,sha256=
|
118
|
+
ai_edge_torch/generative/examples/phi/verify_util.py,sha256=yIfeGP-VJ9CpeOxzhUe0tDCwMSriQVMGOWXEBxJRYcs,3069
|
119
119
|
ai_edge_torch/generative/examples/qwen/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIXY_dRQeYudjsrKGf6LZz65g,671
|
120
120
|
ai_edge_torch/generative/examples/qwen/convert_to_tflite.py,sha256=TnzyARHQgmWeOdYsV9WpRj5vhKGBH0kAbp3tMj8ZCYw,1998
|
121
121
|
ai_edge_torch/generative/examples/qwen/convert_v3_to_tflite.py,sha256=GVV8CVj3rdgt_ZTOlpLSa6AD1pMMpMnZEuowzN2AIGM,2004
|
@@ -123,7 +123,7 @@ ai_edge_torch/generative/examples/qwen/qwen.py,sha256=EcIHVeBcJLc290TiPkPfE7jdG_
|
|
123
123
|
ai_edge_torch/generative/examples/qwen/qwen3.py,sha256=g6aVHjnlPo4YhLjSdXxONaDcKT3fZOh8cewlvf3cfoQ,5554
|
124
124
|
ai_edge_torch/generative/examples/qwen/verify_qwen2.py,sha256=ry-c2QesH-0KnrSQygfjUFs6d4kOFvJz2ts_8mP156I,1659
|
125
125
|
ai_edge_torch/generative/examples/qwen/verify_qwen3.py,sha256=hmE0gdyzgcDpEDcWiwOzKQcxt4XeAe9DPRspy_I-lc8,1628
|
126
|
-
ai_edge_torch/generative/examples/qwen/verify_util.py,sha256=
|
126
|
+
ai_edge_torch/generative/examples/qwen/verify_util.py,sha256=9XhDQyIHFoJ6BYota1lIzwsBrN7t_Mtugpgbjgv17Gg,3277
|
127
127
|
ai_edge_torch/generative/examples/qwen_vl/__init__.py,sha256=JaAnrFoXTl3RJX97XspklkTyqOHVyAgRJsZtzNDd10c,671
|
128
128
|
ai_edge_torch/generative/examples/qwen_vl/convert_to_tflite.py,sha256=BM-ed7KrmPwzI3MvDs2R7P-kJgE1SK_cNVqIfXhtJjs,2411
|
129
129
|
ai_edge_torch/generative/examples/qwen_vl/decoder.py,sha256=plOi-3LltxReW_HVxhxwee_rYCQq-gsOwbGZtRsM8N8,4443
|
@@ -137,7 +137,7 @@ ai_edge_torch/generative/examples/smollm/convert_to_tflite.py,sha256=QVRX_ovqBQi
|
|
137
137
|
ai_edge_torch/generative/examples/smollm/convert_v2_to_tflite.py,sha256=rOVYSaS68_otJcGewQSconBCPD4GhDEIIyquD4dSUWc,1979
|
138
138
|
ai_edge_torch/generative/examples/smollm/smollm.py,sha256=nQRiq6phJbtl3GAEEsJ_bPP_zrpQmiPumNEWCRrECn0,4028
|
139
139
|
ai_edge_torch/generative/examples/smollm/verify.py,sha256=tXiAnwOnqgwyoa8dI4tCBiGUXkOMfdE9MUkkY_Bc4Ig,1603
|
140
|
-
ai_edge_torch/generative/examples/smollm/verify_util.py,sha256=
|
140
|
+
ai_edge_torch/generative/examples/smollm/verify_util.py,sha256=_Z5jtNzQnULeWrXWC1LIpEawKwL_wA1_FtRc6A4-s4k,3019
|
141
141
|
ai_edge_torch/generative/examples/stable_diffusion/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIXY_dRQeYudjsrKGf6LZz65g,671
|
142
142
|
ai_edge_torch/generative/examples/stable_diffusion/attention.py,sha256=kDWG6MlIGa89zC5KSRcJlw2c4ITuw8KcchtfmF55f4g,3545
|
143
143
|
ai_edge_torch/generative/examples/stable_diffusion/clip.py,sha256=lSCRZsoLjH_kqasRMwCy5IogkhyJdwcHKsPEfyxsXCQ,6112
|
@@ -165,7 +165,7 @@ ai_edge_torch/generative/examples/tiny_llama/__init__.py,sha256=hHLluseD2R0Hh4W6
|
|
165
165
|
ai_edge_torch/generative/examples/tiny_llama/convert_to_tflite.py,sha256=urWkWjOaGzV2gwMXoGEs1mfHNEXfEKgwuXmQ0lrWcbM,1761
|
166
166
|
ai_edge_torch/generative/examples/tiny_llama/tiny_llama.py,sha256=cVNP_a_3UBML0j9ITtcITeVXqCdcC7U1JoYwir09Dk8,2936
|
167
167
|
ai_edge_torch/generative/examples/tiny_llama/verify.py,sha256=qzUsW8tJlAD9mqRxDSAcz5xSUKNlBz_DykA3PwUHMwc,1336
|
168
|
-
ai_edge_torch/generative/examples/tiny_llama/verify_util.py,sha256=
|
168
|
+
ai_edge_torch/generative/examples/tiny_llama/verify_util.py,sha256=AxJeLbpDfKN-ztHYTr2qjoTpWQ-n9yeqYKhorFEBzLs,2937
|
169
169
|
ai_edge_torch/generative/fx_passes/__init__.py,sha256=PFSMsA1vfBfrV9ssBCkYJNl8Hx_bLdWjN01iyjPM5jE,1094
|
170
170
|
ai_edge_torch/generative/fx_passes/remove_sdpa_zero_mask_pass.py,sha256=myGjal5A8yIBoqgArd2k40rZmCgD1Ya369KR7182bhI,2129
|
171
171
|
ai_edge_torch/generative/layers/__init__.py,sha256=hHLluseD2R0Hh4W6XZRIXY_dRQeYudjsrKGf6LZz65g,671
|
@@ -180,7 +180,7 @@ ai_edge_torch/generative/layers/feed_forward.py,sha256=_GmtHxwL068l9gh_F_WFcFk7L
|
|
180
180
|
ai_edge_torch/generative/layers/feed_forward_test.py,sha256=Y5l1eC9NgfYixHcfIfE1W4FGh7oC-9UGGyHdKS9tQKc,1880
|
181
181
|
ai_edge_torch/generative/layers/kv_cache.py,sha256=b-7shzDaKexmvQF7P3SiAmIz4ZofjYWv3m5u71GojsA,10460
|
182
182
|
ai_edge_torch/generative/layers/lora.py,sha256=hsvWLLOnW7HQ0AysOZu30x_cetMquDd1tjfyLz8HCSU,17892
|
183
|
-
ai_edge_torch/generative/layers/model_config.py,sha256=
|
183
|
+
ai_edge_torch/generative/layers/model_config.py,sha256=gGr6SZzgFVRXemfXwX__TZ0OXgAFnU70M6U0eql06TE,10712
|
184
184
|
ai_edge_torch/generative/layers/normalization.py,sha256=ijwCpi22NLX-Sygwy5sK9l9WjGvbPIhZvVwoBAonWAo,7014
|
185
185
|
ai_edge_torch/generative/layers/normalization_test.py,sha256=zwurZly-TgFxdgVVdpzu9vCpcLbd5RYt_gKg9Lfg1jI,2248
|
186
186
|
ai_edge_torch/generative/layers/rotary_position_embedding.py,sha256=975zR202MdIrILJ7blceAcxrNqX1ZCN0ECKG1gz-bV8,2655
|
@@ -210,7 +210,7 @@ ai_edge_torch/generative/test/utils.py,sha256=tF6aCfAGJnc9dmzCnZCEOuKNVimfWOqscv
|
|
210
210
|
ai_edge_torch/generative/utilities/__init__.py,sha256=-_jxnnFnCgnTU4oTm4MnRsvL5lqhomBNdFBbqfmfHPo,720
|
211
211
|
ai_edge_torch/generative/utilities/converter.py,sha256=DuoPb8Uhbxa32uUvr6grV5lssmUJdx298QwYz8cG_1Y,15512
|
212
212
|
ai_edge_torch/generative/utilities/export_config.py,sha256=qjkEbjcvi2AgQikZS5qfgR95Z5z9pm07KX-RN5ibfNE,2280
|
213
|
-
ai_edge_torch/generative/utilities/loader.py,sha256=
|
213
|
+
ai_edge_torch/generative/utilities/loader.py,sha256=drgKBmNibuc3PCdc0kU0pVcp2Nt1_mjLYh67RyXOn7U,15952
|
214
214
|
ai_edge_torch/generative/utilities/model_builder.py,sha256=tBfOcsI_NcneggHqkCSydYN3ZgmkzPc6nW0AJrA81wI,6461
|
215
215
|
ai_edge_torch/generative/utilities/moonshine_loader.py,sha256=_RpFabSqtGH5PHiP3_1f6QfO14qMADUxr_HGRlVDFB0,4891
|
216
216
|
ai_edge_torch/generative/utilities/stable_diffusion_loader.py,sha256=dqPD9qRXEWtU3ombslOC-BE2l_dMwHoCNu7NsIJhsso,36158
|
@@ -268,8 +268,8 @@ ai_edge_torch/testing/__init__.py,sha256=_yGgvnBZWb7T3IN3mc4x1sS4vM96HZwM8pwIcPG
|
|
268
268
|
ai_edge_torch/testing/export.py,sha256=k5mGDGzwc23Z4zaIVDs8CNh-oOt64gsf9MS9NjhbPy4,3293
|
269
269
|
ai_edge_torch/testing/model_coverage/__init__.py,sha256=5P8J6Zk5YYtDvTBucFvB9NGSRI7Gw_24WnrbhXgycEE,765
|
270
270
|
ai_edge_torch/testing/model_coverage/model_coverage.py,sha256=UPB448aMDUyC0HNYVqio2rcJPnDN0tBQMP08J6vPYew,4718
|
271
|
-
ai_edge_torch_nightly-0.6.0.
|
272
|
-
ai_edge_torch_nightly-0.6.0.
|
273
|
-
ai_edge_torch_nightly-0.6.0.
|
274
|
-
ai_edge_torch_nightly-0.6.0.
|
275
|
-
ai_edge_torch_nightly-0.6.0.
|
271
|
+
ai_edge_torch_nightly-0.6.0.dev20250531.dist-info/LICENSE,sha256=z8d0m5b2O9McPEK1xHG_dWgUBT6EfBDz6wA0F7xSPTA,11358
|
272
|
+
ai_edge_torch_nightly-0.6.0.dev20250531.dist-info/METADATA,sha256=hxE-vwPFf28jr_h3KH5MnCqXgvpr6GWwh8IetZ2oK8Q,2074
|
273
|
+
ai_edge_torch_nightly-0.6.0.dev20250531.dist-info/WHEEL,sha256=tZoeGjtWxWRfdplE7E3d45VPlLNQnvbKiYnx7gwAy8A,92
|
274
|
+
ai_edge_torch_nightly-0.6.0.dev20250531.dist-info/top_level.txt,sha256=5KXRaF2hwkApYxf7Y8y_tVb9aulGTlbOoNdbx1aKRkE,14
|
275
|
+
ai_edge_torch_nightly-0.6.0.dev20250531.dist-info/RECORD,,
|
File without changes
|
File without changes
|