liger-kernel-nightly 0.5.2.dev20250108073340__py3-none-any.whl → 0.5.2.dev20250108102127__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- liger_kernel/ops/cross_entropy.py +8 -24
- liger_kernel/ops/fused_linear_cross_entropy.py +4 -4
- liger_kernel/transformers/cross_entropy.py +0 -3
- {liger_kernel_nightly-0.5.2.dev20250108073340.dist-info → liger_kernel_nightly-0.5.2.dev20250108102127.dist-info}/METADATA +1 -1
- {liger_kernel_nightly-0.5.2.dev20250108073340.dist-info → liger_kernel_nightly-0.5.2.dev20250108102127.dist-info}/RECORD +9 -9
- {liger_kernel_nightly-0.5.2.dev20250108073340.dist-info → liger_kernel_nightly-0.5.2.dev20250108102127.dist-info}/LICENSE +0 -0
- {liger_kernel_nightly-0.5.2.dev20250108073340.dist-info → liger_kernel_nightly-0.5.2.dev20250108102127.dist-info}/NOTICE +0 -0
- {liger_kernel_nightly-0.5.2.dev20250108073340.dist-info → liger_kernel_nightly-0.5.2.dev20250108102127.dist-info}/WHEEL +0 -0
- {liger_kernel_nightly-0.5.2.dev20250108073340.dist-info → liger_kernel_nightly-0.5.2.dev20250108102127.dist-info}/top_level.txt +0 -0
@@ -20,9 +20,6 @@ if compare_version("triton", operator.ge, "3.0.0"):
|
|
20
20
|
else:
|
21
21
|
from triton.language.math import tanh
|
22
22
|
|
23
|
-
_TRUE: tl.constexpr = tl.constexpr(1)
|
24
|
-
_FALSE: tl.constexpr = tl.constexpr(0)
|
25
|
-
|
26
23
|
|
27
24
|
@triton.jit
|
28
25
|
def liger_cross_entropy_kernel(
|
@@ -95,7 +92,7 @@ def liger_cross_entropy_kernel(
|
|
95
92
|
return
|
96
93
|
|
97
94
|
loss_ptr += program_id * loss_stride
|
98
|
-
if RETURN_Z_LOSS
|
95
|
+
if RETURN_Z_LOSS:
|
99
96
|
z_loss_ptr += program_id * loss_stride
|
100
97
|
|
101
98
|
if HAS_WEIGHT:
|
@@ -254,7 +251,7 @@ def liger_cross_entropy_kernel(
|
|
254
251
|
loss += z_loss
|
255
252
|
|
256
253
|
tl.store(loss_ptr, loss)
|
257
|
-
if RETURN_Z_LOSS
|
254
|
+
if RETURN_Z_LOSS:
|
258
255
|
tl.store(z_loss_ptr, z_loss)
|
259
256
|
|
260
257
|
|
@@ -264,12 +261,6 @@ def liger_cross_entropy_kernel(
|
|
264
261
|
MAX_FUSED_SIZE = 65536 // 2 # the best size we found by manually tuning
|
265
262
|
|
266
263
|
|
267
|
-
_bool_to_return_z_loss = {
|
268
|
-
True: _TRUE.value,
|
269
|
-
False: _FALSE.value,
|
270
|
-
}
|
271
|
-
|
272
|
-
|
273
264
|
def cross_entropy_forward(
|
274
265
|
_input,
|
275
266
|
target,
|
@@ -281,11 +272,7 @@ def cross_entropy_forward(
|
|
281
272
|
softcap,
|
282
273
|
return_z_loss,
|
283
274
|
):
|
284
|
-
|
285
|
-
assert return_z_loss in _bool_to_return_z_loss, f"return_z_loss must be True or False. Got: {return_z_loss}"
|
286
|
-
return_z_loss = _bool_to_return_z_loss[return_z_loss]
|
287
|
-
else:
|
288
|
-
assert return_z_loss in _bool_to_return_z_loss, f"return_z_loss must be True or False. Got: {return_z_loss}"
|
275
|
+
assert isinstance(return_z_loss, bool), f"return_z_loss must be True or False. Got: {return_z_loss}"
|
289
276
|
|
290
277
|
BT, V = _input.shape
|
291
278
|
n_rows = BT
|
@@ -294,10 +281,7 @@ def cross_entropy_forward(
|
|
294
281
|
|
295
282
|
# unreduced loss
|
296
283
|
loss_1d = torch.zeros(n_rows, dtype=_input.dtype, device=_input.device)
|
297
|
-
if return_z_loss
|
298
|
-
z_loss_1d = torch.zeros(n_rows, dtype=_input.dtype, device=_input.device)
|
299
|
-
else:
|
300
|
-
z_loss_1d = None # set None when return_z_loss == False
|
284
|
+
z_loss_1d = torch.zeros(n_rows, dtype=_input.dtype, device=_input.device) if return_z_loss else None
|
301
285
|
|
302
286
|
target_mask = target != ignore_index
|
303
287
|
n_non_ignore = target_mask.sum().item()
|
@@ -326,7 +310,7 @@ def cross_entropy_forward(
|
|
326
310
|
X_stride=_input.stride(-2),
|
327
311
|
Y_ptr=target,
|
328
312
|
Y_stride=target.stride(-1), # always 1
|
329
|
-
weight_ptr=weight
|
313
|
+
weight_ptr=weight, # dummy if None
|
330
314
|
loss_ptr=loss_1d,
|
331
315
|
z_loss_ptr=z_loss_1d,
|
332
316
|
loss_stride=loss_1d.stride(-1), # always 1
|
@@ -338,7 +322,7 @@ def cross_entropy_forward(
|
|
338
322
|
lse_square_scale=lse_square_scale,
|
339
323
|
label_smoothing=label_smoothing,
|
340
324
|
reduction=reduction,
|
341
|
-
softcap=softcap
|
325
|
+
softcap=softcap,
|
342
326
|
RETURN_Z_LOSS=return_z_loss,
|
343
327
|
BLOCK_SIZE=BLOCK_SIZE,
|
344
328
|
HAS_WEIGHT=True if weight is not None else False,
|
@@ -350,10 +334,10 @@ def cross_entropy_forward(
|
|
350
334
|
|
351
335
|
if reduction == "none":
|
352
336
|
loss = loss_1d
|
353
|
-
z_loss = z_loss_1d if return_z_loss
|
337
|
+
z_loss = z_loss_1d if return_z_loss else None
|
354
338
|
else:
|
355
339
|
loss = torch.sum(loss_1d)
|
356
|
-
z_loss = torch.sum(z_loss_1d) if return_z_loss
|
340
|
+
z_loss = torch.sum(z_loss_1d) if return_z_loss else None
|
357
341
|
|
358
342
|
return loss, z_loss, _input
|
359
343
|
|
@@ -92,9 +92,9 @@ def fused_linear_cross_entropy_forward(
|
|
92
92
|
X_stride=logits_chunk.stride(-2),
|
93
93
|
Y_ptr=target_chunk,
|
94
94
|
Y_stride=target_chunk.stride(-1), # always 1
|
95
|
-
weight_ptr=ce_weight
|
95
|
+
weight_ptr=ce_weight,
|
96
96
|
loss_ptr=loss_1d_slice,
|
97
|
-
z_loss_ptr=
|
97
|
+
z_loss_ptr=None,
|
98
98
|
loss_stride=loss_1d_slice.stride(-1), # always 1
|
99
99
|
n_cols=V,
|
100
100
|
n_non_ignore=total_n_non_ignore,
|
@@ -104,8 +104,8 @@ def fused_linear_cross_entropy_forward(
|
|
104
104
|
lse_square_scale=lse_square_scale,
|
105
105
|
label_smoothing=label_smoothing,
|
106
106
|
reduction=reduction,
|
107
|
-
softcap=softcap
|
108
|
-
RETURN_Z_LOSS=
|
107
|
+
softcap=softcap,
|
108
|
+
RETURN_Z_LOSS=False,
|
109
109
|
HAS_WEIGHT=True if ce_weight is not None else False,
|
110
110
|
HAS_SOFTCAPPING=True if softcap is not None else False,
|
111
111
|
BLOCK_SIZE=BLOCK_SIZE,
|
@@ -20,9 +20,6 @@ class LigerCrossEntropyLoss(torch.nn.Module):
|
|
20
20
|
assert (label_smoothing >= 0) and (
|
21
21
|
label_smoothing <= 1
|
22
22
|
), f"label_smoothing must be between 0.0 and 1.0. Got: {label_smoothing}"
|
23
|
-
assert (label_smoothing >= 0) and (
|
24
|
-
label_smoothing <= 1
|
25
|
-
), f"label_smoothing must be between 0.0 and 1.0. Got: {label_smoothing}"
|
26
23
|
assert reduction in {
|
27
24
|
"mean",
|
28
25
|
"sum",
|
@@ -11,8 +11,8 @@ liger_kernel/chunked_loss/fused_linear_preference.py,sha256=eQCZmQ3xOL3jpZ7RhOfx
|
|
11
11
|
liger_kernel/chunked_loss/orpo_loss.py,sha256=jbZxx-EjPK71A6CSyNzTOAIEQgAUjfvwSViw6R_pPXQ,3510
|
12
12
|
liger_kernel/chunked_loss/simpo_loss.py,sha256=3TTc7U79Orjgi-Wu81WZkWk5MgsdqKXIOBHgIvDazPw,3865
|
13
13
|
liger_kernel/ops/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
14
|
-
liger_kernel/ops/cross_entropy.py,sha256=
|
15
|
-
liger_kernel/ops/fused_linear_cross_entropy.py,sha256=
|
14
|
+
liger_kernel/ops/cross_entropy.py,sha256=SRzAF9Ek84pBVFy3wqQZs7AhRoorKRIgQ-Td_rtl1Kk,18606
|
15
|
+
liger_kernel/ops/fused_linear_cross_entropy.py,sha256=hezFRwbcPc-HNGZUFqUn5AYUqUpboPpFh4MNqEW4WgU,10108
|
16
16
|
liger_kernel/ops/fused_linear_jsd.py,sha256=eKqaADj7LgWfoYqyH03tjrmhNTfJOF1Dhx_bWzBTnTU,9600
|
17
17
|
liger_kernel/ops/geglu.py,sha256=axGvCIvlBzuluoAIrWTsp2iZM4BFKNInkPov8YVvH9E,4126
|
18
18
|
liger_kernel/ops/group_norm.py,sha256=qD4D4lSjSgVtO52EBNLC2iTseALRgPgqXE50U2woggk,10837
|
@@ -28,7 +28,7 @@ liger_kernel/ops/experimental/embedding.py,sha256=tolj3tItkzpSb30zWqDN2_yX4ectfl
|
|
28
28
|
liger_kernel/ops/experimental/mm_int8int2.py,sha256=TrS9lpwekrik_w5qE7AhMJD1bcq-OidjtbsW80oZ6IM,13314
|
29
29
|
liger_kernel/transformers/__init__.py,sha256=QPmYkL6hosBPpPqCUGqvIvAtD9XzLgvZqZxUyYMZeVk,2008
|
30
30
|
liger_kernel/transformers/auto_model.py,sha256=0qCTRZt280Bj_LcFdzo9hlaR-BWNazawXOGgoCZjgEg,1545
|
31
|
-
liger_kernel/transformers/cross_entropy.py,sha256=
|
31
|
+
liger_kernel/transformers/cross_entropy.py,sha256=LtiHlj_tK2YFpilwvbG_NEVzbf82zKRpWCZMjaFUd4M,1681
|
32
32
|
liger_kernel/transformers/functional.py,sha256=B1wkHWLx-YNhxvXBEXB4Ch1yEwF3mjwTPCeXA5aCV_c,4490
|
33
33
|
liger_kernel/transformers/fused_linear_cross_entropy.py,sha256=LAN8-pjUI2Erz_MnfMer-0ZmxJ0JlKxGzdZGJY-N65g,1569
|
34
34
|
liger_kernel/transformers/fused_linear_jsd.py,sha256=bZ4otCvWBuOnA5XdQL-FzZVItJlDt-ht9e_pG7PG93E,3999
|
@@ -58,9 +58,9 @@ liger_kernel/transformers/trainer/__init__.py,sha256=p7yQfklV8-467qSz_ZMimkbDF7H
|
|
58
58
|
liger_kernel/transformers/trainer/orpo_trainer.py,sha256=MId1S_MfA3pPVQA1rkiKxp-jZDNz8VmvZzXC-Kugol4,7662
|
59
59
|
liger_kernel/triton/__init__.py,sha256=qCiCamzCRv6lpV8IqpAc9YMdNKC7GKurClWceQPnlis,92
|
60
60
|
liger_kernel/triton/monkey_patch.py,sha256=Rd0hUHAzDkFfHvnX7-PBaNK5EKnZhtfM_h-fgQH9HPY,1568
|
61
|
-
liger_kernel_nightly-0.5.2.
|
62
|
-
liger_kernel_nightly-0.5.2.
|
63
|
-
liger_kernel_nightly-0.5.2.
|
64
|
-
liger_kernel_nightly-0.5.2.
|
65
|
-
liger_kernel_nightly-0.5.2.
|
66
|
-
liger_kernel_nightly-0.5.2.
|
61
|
+
liger_kernel_nightly-0.5.2.dev20250108102127.dist-info/LICENSE,sha256=OhzLDHJ0to4a8sodVLELZiCFylZ1NAAYLs-HrjPy0ag,1312
|
62
|
+
liger_kernel_nightly-0.5.2.dev20250108102127.dist-info/METADATA,sha256=XHrJlebOzBW0f6tV-rb0iahG9LNI-f86Ar7s-upwoxo,21055
|
63
|
+
liger_kernel_nightly-0.5.2.dev20250108102127.dist-info/NOTICE,sha256=njwnoPZLh9AN8SJQzxvCGLHi-8X__AvWRze6joNXIY8,2066
|
64
|
+
liger_kernel_nightly-0.5.2.dev20250108102127.dist-info/WHEEL,sha256=P9jw-gEje8ByB7_hXoICnHtVCrEwMQh-630tKvQWehc,91
|
65
|
+
liger_kernel_nightly-0.5.2.dev20250108102127.dist-info/top_level.txt,sha256=2eghu4hA3LnkM7ElW92tQ8zegWKgSbeo-k-aGe1YnvY,13
|
66
|
+
liger_kernel_nightly-0.5.2.dev20250108102127.dist-info/RECORD,,
|
File without changes
|
File without changes
|
File without changes
|