liger-kernel-nightly 0.5.3.dev20250219232423__py3-none-any.whl → 0.5.3.dev20250220195514__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- liger_kernel/ops/cross_entropy.py +3 -3
- liger_kernel/ops/fused_linear_cross_entropy.py +3 -3
- liger_kernel/ops/fused_linear_jsd.py +3 -3
- liger_kernel/ops/jsd.py +3 -3
- liger_kernel/ops/layer_norm.py +3 -3
- liger_kernel/ops/utils.py +1 -2
- liger_kernel/transformers/cross_entropy.py +3 -3
- liger_kernel/transformers/fused_linear_cross_entropy.py +3 -3
- liger_kernel/transformers/group_norm.py +6 -6
- liger_kernel/transformers/monkey_patch.py +27 -27
- {liger_kernel_nightly-0.5.3.dev20250219232423.dist-info → liger_kernel_nightly-0.5.3.dev20250220195514.dist-info}/METADATA +6 -1
- {liger_kernel_nightly-0.5.3.dev20250219232423.dist-info → liger_kernel_nightly-0.5.3.dev20250220195514.dist-info}/RECORD +16 -16
- {liger_kernel_nightly-0.5.3.dev20250219232423.dist-info → liger_kernel_nightly-0.5.3.dev20250220195514.dist-info}/LICENSE +0 -0
- {liger_kernel_nightly-0.5.3.dev20250219232423.dist-info → liger_kernel_nightly-0.5.3.dev20250220195514.dist-info}/NOTICE +0 -0
- {liger_kernel_nightly-0.5.3.dev20250219232423.dist-info → liger_kernel_nightly-0.5.3.dev20250220195514.dist-info}/WHEEL +0 -0
- {liger_kernel_nightly-0.5.3.dev20250219232423.dist-info → liger_kernel_nightly-0.5.3.dev20250220195514.dist-info}/top_level.txt +0 -0
@@ -289,9 +289,9 @@ def cross_entropy_forward(
|
|
289
289
|
weight_sum = 0.0
|
290
290
|
if weight is not None:
|
291
291
|
assert weight.shape[0] == V, f"If given, weight has to be a Tensor of size V. Got: {weight.shape}"
|
292
|
-
assert torch.is_floating_point(
|
293
|
-
weight
|
294
|
-
)
|
292
|
+
assert torch.is_floating_point(weight), (
|
293
|
+
f"If given, weight has to be a Tensor of floating point dtype. Got: {weight.dtype}"
|
294
|
+
)
|
295
295
|
sum_non_ignore_weight = torch.gather(weight, dim=0, index=target.masked_select(target_mask)).sum().item()
|
296
296
|
weight_sum = weight.sum().item()
|
297
297
|
# ensure weight is contiguous
|
@@ -58,9 +58,9 @@ def fused_linear_cross_entropy_forward(
|
|
58
58
|
ce_weight_sum = 0.0
|
59
59
|
if ce_weight is not None:
|
60
60
|
assert ce_weight.shape[0] == V, f"If given, weight has to be a Tensor of size V. Got: {ce_weight.shape}"
|
61
|
-
assert torch.is_floating_point(
|
62
|
-
ce_weight
|
63
|
-
)
|
61
|
+
assert torch.is_floating_point(ce_weight), (
|
62
|
+
f"If given, weight has to be a Tensor of floating point dtype. Got: {ce_weight.dtype}"
|
63
|
+
)
|
64
64
|
total_sum_non_ignore_ce_weight = (
|
65
65
|
torch.gather(ce_weight, dim=0, index=target.masked_select(target_mask)).sum().item()
|
66
66
|
)
|
@@ -195,9 +195,9 @@ class LigerFusedLinearJSDFunction(torch.autograd.Function):
|
|
195
195
|
"""
|
196
196
|
has_label = False
|
197
197
|
if shift_labels is not None:
|
198
|
-
assert shift_labels.shape == (
|
199
|
-
|
200
|
-
)
|
198
|
+
assert shift_labels.shape == (teacher_input.shape[0],), (
|
199
|
+
f"the shape of shift_labels must be (BT,). Got: {shift_labels.shape}"
|
200
|
+
)
|
201
201
|
shift_labels = shift_labels.contiguous()
|
202
202
|
has_label = True
|
203
203
|
|
liger_kernel/ops/jsd.py
CHANGED
@@ -157,9 +157,9 @@ class LigerJSDFunction(torch.autograd.Function):
|
|
157
157
|
"""
|
158
158
|
has_label = False
|
159
159
|
if shift_labels is not None:
|
160
|
-
assert shift_labels.shape == (
|
161
|
-
|
162
|
-
)
|
160
|
+
assert shift_labels.shape == (_input.shape[0],), (
|
161
|
+
f"the shape of shift_labels must be (BT,). Got: {shift_labels.shape}"
|
162
|
+
)
|
163
163
|
shift_labels = shift_labels.contiguous()
|
164
164
|
has_label = True
|
165
165
|
|
liger_kernel/ops/layer_norm.py
CHANGED
@@ -147,9 +147,9 @@ def layer_norm_forward(X, W, B, eps):
|
|
147
147
|
Y = torch.empty((n_rows, n_cols), dtype=X.dtype, device=X.device)
|
148
148
|
Mean = torch.empty(n_rows, dtype=X.dtype, device=X.device)
|
149
149
|
RSTD = torch.empty(n_rows, dtype=X.dtype, device=X.device)
|
150
|
-
assert (
|
151
|
-
X.shape[1]
|
152
|
-
)
|
150
|
+
assert X.shape[1] == W.shape[0], (
|
151
|
+
f"Incompatible hidden size dimension between input tensor with shape[1] = {X.shape[1]} and weight tensor with shape[0] = {W.shape[0]}"
|
152
|
+
)
|
153
153
|
|
154
154
|
_layer_norm_forward_kernel[(n_rows,)](
|
155
155
|
Y,
|
liger_kernel/ops/utils.py
CHANGED
@@ -49,8 +49,7 @@ def calculate_settings(n):
|
|
49
49
|
BLOCK_SIZE = triton.next_power_of_2(n)
|
50
50
|
if BLOCK_SIZE > MAX_FUSED_SIZE:
|
51
51
|
raise RuntimeError(
|
52
|
-
f"Cannot launch Triton kernel since n = {n} exceeds "
|
53
|
-
f"the recommended Triton blocksize = {MAX_FUSED_SIZE}."
|
52
|
+
f"Cannot launch Triton kernel since n = {n} exceeds the recommended Triton blocksize = {MAX_FUSED_SIZE}."
|
54
53
|
)
|
55
54
|
|
56
55
|
num_warps = 4
|
@@ -17,9 +17,9 @@ class LigerCrossEntropyLoss(torch.nn.Module):
|
|
17
17
|
return_z_loss: bool = False,
|
18
18
|
):
|
19
19
|
super().__init__()
|
20
|
-
assert (label_smoothing >= 0) and (
|
21
|
-
label_smoothing
|
22
|
-
)
|
20
|
+
assert (label_smoothing >= 0) and (label_smoothing <= 1), (
|
21
|
+
f"label_smoothing must be between 0.0 and 1.0. Got: {label_smoothing}"
|
22
|
+
)
|
23
23
|
assert reduction in {
|
24
24
|
"mean",
|
25
25
|
"sum",
|
@@ -17,9 +17,9 @@ class LigerFusedLinearCrossEntropyLoss(torch.nn.Module):
|
|
17
17
|
return_z_loss: bool = False,
|
18
18
|
):
|
19
19
|
super().__init__()
|
20
|
-
assert (label_smoothing >= 0) and (
|
21
|
-
label_smoothing
|
22
|
-
)
|
20
|
+
assert (label_smoothing >= 0) and (label_smoothing <= 1), (
|
21
|
+
f"label_smoothing must be between 0.0 and 1.0. Got: {label_smoothing}"
|
22
|
+
)
|
23
23
|
assert reduction in {
|
24
24
|
"mean",
|
25
25
|
"sum",
|
@@ -21,9 +21,9 @@ class LigerGroupNorm(nn.Module):
|
|
21
21
|
"zeros",
|
22
22
|
], f"init_fn must be either 'ones' or 'zeros', got {init_fn}"
|
23
23
|
|
24
|
-
assert (
|
25
|
-
num_channels
|
26
|
-
)
|
24
|
+
assert num_channels % num_groups == 0, (
|
25
|
+
f"Number of channels {num_channels} must be divisible by num_groups {num_groups}"
|
26
|
+
)
|
27
27
|
self.num_channels = num_channels
|
28
28
|
self.num_groups = num_groups
|
29
29
|
self.eps = eps
|
@@ -34,9 +34,9 @@ class LigerGroupNorm(nn.Module):
|
|
34
34
|
def forward(self, hidden_states):
|
35
35
|
# hidden_states: (batch_size, num_channels, *)
|
36
36
|
assert hidden_states.dim() >= 3, f"Input must have atleast 3 dimensions, got {hidden_states.dim()}"
|
37
|
-
assert (
|
38
|
-
hidden_states.size(1)
|
39
|
-
)
|
37
|
+
assert hidden_states.size(1) == self.num_channels, (
|
38
|
+
f"Input tensor must have {self.num_channels} channels, got {hidden_states.size(1)}"
|
39
|
+
)
|
40
40
|
return LigerGroupNormFunction.apply(
|
41
41
|
hidden_states,
|
42
42
|
self.weight,
|
@@ -85,9 +85,9 @@ def apply_liger_kernel_to_llama(
|
|
85
85
|
loaded. Default is None.
|
86
86
|
"""
|
87
87
|
|
88
|
-
assert not (
|
89
|
-
cross_entropy and fused_linear_cross_entropy
|
90
|
-
)
|
88
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
89
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
90
|
+
)
|
91
91
|
|
92
92
|
from transformers.models.llama import modeling_llama
|
93
93
|
from transformers.models.llama.modeling_llama import LlamaModel
|
@@ -159,9 +159,9 @@ def apply_liger_kernel_to_mllama(
|
|
159
159
|
loaded. Default is None.
|
160
160
|
"""
|
161
161
|
|
162
|
-
assert not (
|
163
|
-
cross_entropy and fused_linear_cross_entropy
|
164
|
-
)
|
162
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
163
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
164
|
+
)
|
165
165
|
|
166
166
|
from transformers.models.mllama import modeling_mllama
|
167
167
|
from transformers.models.mllama.modeling_mllama import MllamaForCausalLM
|
@@ -261,9 +261,9 @@ def apply_liger_kernel_to_mistral(
|
|
261
261
|
model (PreTrainedModel): The model instance to apply Liger kernels to, if the model has already been
|
262
262
|
loaded. Default is None.
|
263
263
|
"""
|
264
|
-
assert not (
|
265
|
-
cross_entropy and fused_linear_cross_entropy
|
266
|
-
)
|
264
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
265
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
266
|
+
)
|
267
267
|
|
268
268
|
from transformers.models.mistral import modeling_mistral
|
269
269
|
from transformers.models.mistral.modeling_mistral import MistralModel
|
@@ -321,9 +321,9 @@ def apply_liger_kernel_to_mixtral(
|
|
321
321
|
loaded. Default is None.
|
322
322
|
"""
|
323
323
|
|
324
|
-
assert not (
|
325
|
-
cross_entropy and fused_linear_cross_entropy
|
326
|
-
)
|
324
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
325
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
326
|
+
)
|
327
327
|
|
328
328
|
from transformers.models.mixtral import modeling_mixtral
|
329
329
|
from transformers.models.mixtral.modeling_mixtral import MixtralModel
|
@@ -393,9 +393,9 @@ def apply_liger_kernel_to_gemma(
|
|
393
393
|
model (PreTrainedModel): The model instance to apply Liger kernels to, if the model has already been
|
394
394
|
loaded. Default is None.
|
395
395
|
"""
|
396
|
-
assert not (
|
397
|
-
cross_entropy and fused_linear_cross_entropy
|
398
|
-
)
|
396
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
397
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
398
|
+
)
|
399
399
|
|
400
400
|
from transformers.models.gemma import modeling_gemma
|
401
401
|
from transformers.models.gemma.modeling_gemma import GemmaModel
|
@@ -467,9 +467,9 @@ def apply_liger_kernel_to_gemma2(
|
|
467
467
|
model (PreTrainedModel): The model instance to apply Liger kernels to, if the model has already been
|
468
468
|
loaded. Default is None.
|
469
469
|
"""
|
470
|
-
assert not (
|
471
|
-
cross_entropy and fused_linear_cross_entropy
|
472
|
-
)
|
470
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
471
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
472
|
+
)
|
473
473
|
|
474
474
|
from transformers.models.gemma2 import modeling_gemma2
|
475
475
|
from transformers.models.gemma2.modeling_gemma2 import Gemma2Model
|
@@ -544,9 +544,9 @@ def apply_liger_kernel_to_qwen2(
|
|
544
544
|
model (PreTrainedModel): The model instance to apply Liger kernels to, if the model has already been
|
545
545
|
loaded. Default is None.
|
546
546
|
"""
|
547
|
-
assert not (
|
548
|
-
cross_entropy and fused_linear_cross_entropy
|
549
|
-
)
|
547
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
548
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
549
|
+
)
|
550
550
|
|
551
551
|
from transformers.models.qwen2 import modeling_qwen2
|
552
552
|
from transformers.models.qwen2.modeling_qwen2 import Qwen2Model
|
@@ -619,9 +619,9 @@ def apply_liger_kernel_to_qwen2_vl(
|
|
619
619
|
model (PreTrainedModel): The model instance to apply Liger kernels to, if the model has already been
|
620
620
|
loaded. Default is None.
|
621
621
|
"""
|
622
|
-
assert not (
|
623
|
-
cross_entropy and fused_linear_cross_entropy
|
624
|
-
)
|
622
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
623
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
624
|
+
)
|
625
625
|
|
626
626
|
from transformers.models.qwen2_vl import modeling_qwen2_vl
|
627
627
|
from transformers.models.qwen2_vl.modeling_qwen2_vl import Qwen2VLModel
|
@@ -689,9 +689,9 @@ def apply_liger_kernel_to_phi3(
|
|
689
689
|
model (PreTrainedModel): The model instance to apply Liger kernels to, if the model has already been
|
690
690
|
loaded. Default is None.
|
691
691
|
"""
|
692
|
-
assert not (
|
693
|
-
cross_entropy and fused_linear_cross_entropy
|
694
|
-
)
|
692
|
+
assert not (cross_entropy and fused_linear_cross_entropy), (
|
693
|
+
"cross_entropy and fused_linear_cross_entropy cannot both be True."
|
694
|
+
)
|
695
695
|
|
696
696
|
from transformers.models.phi3 import modeling_phi3
|
697
697
|
from transformers.models.phi3.modeling_phi3 import Phi3Model
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.1
|
2
2
|
Name: liger_kernel_nightly
|
3
|
-
Version: 0.5.3.
|
3
|
+
Version: 0.5.3.dev20250220195514
|
4
4
|
Summary: Efficient Triton kernels for LLM Training
|
5
5
|
License: BSD 2-CLAUSE LICENSE
|
6
6
|
Copyright 2024 LinkedIn Corporation
|
@@ -191,6 +191,11 @@ y = orpo_loss(lm_head.weight, x, target)
|
|
191
191
|
- `torch >= 2.5.0` Install according to the instruction in Pytorch official webpage.
|
192
192
|
- `triton >= 3.0.0` Install from pypi. (e.g. `pip install triton==3.0.0`)
|
193
193
|
|
194
|
+
```bash
|
195
|
+
# Need to pass the url when installing
|
196
|
+
pip install -e .[dev] --extra-index-url https://download.pytorch.org/whl/nightly/rocm6.2
|
197
|
+
```
|
198
|
+
|
194
199
|
### Optional Dependencies
|
195
200
|
|
196
201
|
- `transformers >= 4.x`: Required if you plan to use the transformers models patching APIs. The specific model you are working will dictate the minimum version of transformers.
|
@@ -16,33 +16,33 @@ liger_kernel/chunked_loss/kto_loss.py,sha256=eVNW6HVCAm32shpfhbRlk92Flnjd7G32v0g
|
|
16
16
|
liger_kernel/chunked_loss/orpo_loss.py,sha256=yjcrrbVeemLYodoSKT-FMSnaPtyKAZ3aOrvPD6tTY6Y,3617
|
17
17
|
liger_kernel/chunked_loss/simpo_loss.py,sha256=3TTc7U79Orjgi-Wu81WZkWk5MgsdqKXIOBHgIvDazPw,3865
|
18
18
|
liger_kernel/ops/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
19
|
-
liger_kernel/ops/cross_entropy.py,sha256=
|
20
|
-
liger_kernel/ops/fused_linear_cross_entropy.py,sha256=
|
21
|
-
liger_kernel/ops/fused_linear_jsd.py,sha256=
|
19
|
+
liger_kernel/ops/cross_entropy.py,sha256=D6vFFloiuxFXoWfjlIjmfO3tVaWOiYmztw9FKAi5vdU,18608
|
20
|
+
liger_kernel/ops/fused_linear_cross_entropy.py,sha256=1Y3Uk_TCSjqKgoG2eot1ptnWXJXXQESqGvOmqAW1gsM,10912
|
21
|
+
liger_kernel/ops/fused_linear_jsd.py,sha256=Seshez2qaM6HiTQ8_HEqSwhaeVruNT1SvIM4ZrAPBEU,9602
|
22
22
|
liger_kernel/ops/geglu.py,sha256=axGvCIvlBzuluoAIrWTsp2iZM4BFKNInkPov8YVvH9E,4126
|
23
23
|
liger_kernel/ops/group_norm.py,sha256=qD4D4lSjSgVtO52EBNLC2iTseALRgPgqXE50U2woggk,10837
|
24
|
-
liger_kernel/ops/jsd.py,sha256=
|
24
|
+
liger_kernel/ops/jsd.py,sha256=0jNeRxpcNI5ckxCdoCNyO5GEedLIuzx3lz6KAiksc4o,6109
|
25
25
|
liger_kernel/ops/kl_div.py,sha256=MnfuYqqQESON1X2Swy064x1urKtMFdgeSWd60VttBXI,8420
|
26
|
-
liger_kernel/ops/layer_norm.py,sha256=
|
26
|
+
liger_kernel/ops/layer_norm.py,sha256=o5X_N0XNX0t-1AV3dyv43G0KJSyclUxcpNXzHNh35ks,7640
|
27
27
|
liger_kernel/ops/qwen2vl_mrope.py,sha256=3GExhYpLgB4VUtyZyjRk8XjEur3W4EWF6HQ67ML5vBU,8481
|
28
28
|
liger_kernel/ops/rms_norm.py,sha256=PWLJcdIKU5e-8BuYFHd9Cqlq6wmr6fUXKi9zQD4LetU,11727
|
29
29
|
liger_kernel/ops/rope.py,sha256=ofmBOkUpZZO-Q8Z5B_LOFYYLD-YT-8WnJ4vGOrDYouI,8943
|
30
30
|
liger_kernel/ops/swiglu.py,sha256=KmgMjaJQnbLLgZn2nEpbwHU_xpnYRweCyrLQSVvM1vA,3015
|
31
|
-
liger_kernel/ops/utils.py,sha256=
|
31
|
+
liger_kernel/ops/utils.py,sha256=uoFKQqo-34N2TWQNvXMFywqGiOMMXNEVBxVojzlUAa0,3836
|
32
32
|
liger_kernel/ops/experimental/embedding.py,sha256=tolj3tItkzpSb30zWqDN2_yX4ectflaQ8HMyKyFIQc8,4172
|
33
33
|
liger_kernel/ops/experimental/mm_int8int2.py,sha256=TrS9lpwekrik_w5qE7AhMJD1bcq-OidjtbsW80oZ6IM,13314
|
34
34
|
liger_kernel/transformers/__init__.py,sha256=QPmYkL6hosBPpPqCUGqvIvAtD9XzLgvZqZxUyYMZeVk,2008
|
35
35
|
liger_kernel/transformers/auto_model.py,sha256=0qCTRZt280Bj_LcFdzo9hlaR-BWNazawXOGgoCZjgEg,1545
|
36
|
-
liger_kernel/transformers/cross_entropy.py,sha256=
|
36
|
+
liger_kernel/transformers/cross_entropy.py,sha256=z3KTWQnFxr_IZaVjtYt0ZNEWQdDdYThN35xWkHlDGH0,1683
|
37
37
|
liger_kernel/transformers/functional.py,sha256=lDOjch622dJIc78K3ePFK_H1DX00GC5kKjodjcbEgbM,4624
|
38
|
-
liger_kernel/transformers/fused_linear_cross_entropy.py,sha256=
|
38
|
+
liger_kernel/transformers/fused_linear_cross_entropy.py,sha256=09Rt7FZzLH42VOcIbQ4dlQd0o3Rlb4vk6fqiOQ7WTD8,1778
|
39
39
|
liger_kernel/transformers/fused_linear_jsd.py,sha256=bZ4otCvWBuOnA5XdQL-FzZVItJlDt-ht9e_pG7PG93E,3999
|
40
40
|
liger_kernel/transformers/geglu.py,sha256=mrgqzIUVd6lN7fkDKLkw5YaESDxDtFgbot430WwPVOQ,1107
|
41
|
-
liger_kernel/transformers/group_norm.py,sha256=
|
41
|
+
liger_kernel/transformers/group_norm.py,sha256=6qMAWOprr4SzP0YhNVNGQIBpM5aUHplUD2VuGJrMBz0,2173
|
42
42
|
liger_kernel/transformers/jsd.py,sha256=DGqRnxIZxsvxo0_tbbxX3b-sDbDjC_yKufyRIHCcScY,2979
|
43
43
|
liger_kernel/transformers/kl_div.py,sha256=WLffFbh1EExD2Eb1F7lN11fo9JJC-0751WJjZAF1Fj8,409
|
44
44
|
liger_kernel/transformers/layer_norm.py,sha256=c9pk3PEasOKYR0rhe5e5nNrnYKVCEW4VC8S6LpCq9EQ,906
|
45
|
-
liger_kernel/transformers/monkey_patch.py,sha256=
|
45
|
+
liger_kernel/transformers/monkey_patch.py,sha256=DXU00zsQvSjAqCx7l36gKm1O81FuHgILkZMhyx4ZSys,37812
|
46
46
|
liger_kernel/transformers/qwen2vl_mrope.py,sha256=5EwSqrMdsL9MYspeBMXBsNJKvH0MOmRrtJXAJlnnlOI,1047
|
47
47
|
liger_kernel/transformers/rms_norm.py,sha256=GqCEJuGt0YdqqlMcToE0Wp4A8YFquDa4UUSyH2uFW2A,1191
|
48
48
|
liger_kernel/transformers/rope.py,sha256=ZTrTORSAyfcFIKjk6XEeYmk4ROH7xXED9L4g2NFntlE,999
|
@@ -63,9 +63,9 @@ liger_kernel/transformers/trainer/__init__.py,sha256=p7yQfklV8-467qSz_ZMimkbDF7H
|
|
63
63
|
liger_kernel/transformers/trainer/orpo_trainer.py,sha256=pdekW7l6Qg_aqa5SYKYlSWUF8m3lkOFvFLcIMEHrz9s,8338
|
64
64
|
liger_kernel/triton/__init__.py,sha256=qCiCamzCRv6lpV8IqpAc9YMdNKC7GKurClWceQPnlis,92
|
65
65
|
liger_kernel/triton/monkey_patch.py,sha256=Rd0hUHAzDkFfHvnX7-PBaNK5EKnZhtfM_h-fgQH9HPY,1568
|
66
|
-
liger_kernel_nightly-0.5.3.
|
67
|
-
liger_kernel_nightly-0.5.3.
|
68
|
-
liger_kernel_nightly-0.5.3.
|
69
|
-
liger_kernel_nightly-0.5.3.
|
70
|
-
liger_kernel_nightly-0.5.3.
|
71
|
-
liger_kernel_nightly-0.5.3.
|
66
|
+
liger_kernel_nightly-0.5.3.dev20250220195514.dist-info/LICENSE,sha256=OhzLDHJ0to4a8sodVLELZiCFylZ1NAAYLs-HrjPy0ag,1312
|
67
|
+
liger_kernel_nightly-0.5.3.dev20250220195514.dist-info/METADATA,sha256=WkDgh3E1y7TYWCDttILZqikHz2S5b2kxLKoJ7JiWMd8,21766
|
68
|
+
liger_kernel_nightly-0.5.3.dev20250220195514.dist-info/NOTICE,sha256=njwnoPZLh9AN8SJQzxvCGLHi-8X__AvWRze6joNXIY8,2066
|
69
|
+
liger_kernel_nightly-0.5.3.dev20250220195514.dist-info/WHEEL,sha256=P9jw-gEje8ByB7_hXoICnHtVCrEwMQh-630tKvQWehc,91
|
70
|
+
liger_kernel_nightly-0.5.3.dev20250220195514.dist-info/top_level.txt,sha256=2eghu4hA3LnkM7ElW92tQ8zegWKgSbeo-k-aGe1YnvY,13
|
71
|
+
liger_kernel_nightly-0.5.3.dev20250220195514.dist-info/RECORD,,
|
File without changes
|
File without changes
|
File without changes
|