@genai-fi/nanogpt 0.6.3 → 0.7.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (140) hide show
  1. package/dist/Generator.js +11 -11
  2. package/dist/NanoGPTModel.d.ts +2 -2
  3. package/dist/NanoGPTModel.js +104 -136
  4. package/dist/{RealDiv-BYViZwhN.js → RealDiv-C4hOvYOZ.js} +26 -25
  5. package/dist/{Reshape-t7Kcikjk.js → Reshape-BLijOA8h.js} +5 -5
  6. package/dist/TeachableLLM.js +5 -5
  7. package/dist/{TiedEmbedding-9WeDwvjO.js → TiedEmbedding-BLltddza.js} +4 -4
  8. package/dist/{axis_util-Bu4h7XWV.js → axis_util-DaAl5MER.js} +3 -3
  9. package/dist/backend.d.ts +1 -0
  10. package/dist/backend.js +7 -0
  11. package/dist/backend_util-DWiwsi2N.js +749 -0
  12. package/dist/{broadcast_to-DARN-DBD.js → broadcast_to-C4v-j9yA.js} +2 -2
  13. package/dist/{concat-5aPGqw3Z.js → concat-CsHeR4zV.js} +8 -8
  14. package/dist/{dataset-pgqp-YfL.js → dataset-JDyjG3QR.js} +3 -3
  15. package/dist/{dropout-Bciw46HT.js → dropout-hpDwECTe.js} +7 -7
  16. package/dist/{gather-DjyCjmOD.js → gather-D0_gPiBz.js} +4 -4
  17. package/dist/gelu-uyHP1x1f.js +26 -0
  18. package/dist/gpgpu_math-DJm3ZTAf.js +2371 -0
  19. package/dist/index-BPPzKVdR.js +12099 -0
  20. package/dist/{index-BAzbokzv.js → index-C0dhsYom.js} +405 -389
  21. package/dist/{kernel_funcs_utils-CUxJCg0g.js → kernel_funcs_utils-CwRTFqrc.js} +31 -30
  22. package/dist/layers/BaseLayer.js +2 -2
  23. package/dist/layers/CausalSelfAttention.js +6 -6
  24. package/dist/layers/MLP.js +5 -5
  25. package/dist/layers/RMSNorm.js +3 -3
  26. package/dist/layers/RoPECache.js +4 -4
  27. package/dist/layers/TiedEmbedding.js +5 -5
  28. package/dist/layers/TransformerBlock.js +1 -1
  29. package/dist/loader/loadTransformers.js +1 -1
  30. package/dist/loader/oldZipLoad.js +5 -5
  31. package/dist/{log_sum_exp-YEo2h3gb.js → log_sum_exp-D086OgZJ.js} +15 -15
  32. package/dist/main.d.ts +2 -0
  33. package/dist/main.js +9 -5
  34. package/dist/{mat_mul-7121rsJk.js → mat_mul-1nwdPkQ_.js} +4 -4
  35. package/dist/{max-DtlIuVeW.js → max-BQc2Aj-I.js} +4 -4
  36. package/dist/{mulmat_packed_gpu-D4nKF7Je.js → mulmat_packed_gpu-Gzf3I9UV.js} +1 -1
  37. package/dist/non_max_suppression_impl-CsEgBuMA.js +134 -0
  38. package/dist/{ones-BBlSRqn1.js → ones-D63HpSF_.js} +2 -2
  39. package/dist/ops/appendCache.js +3 -3
  40. package/dist/ops/attentionMask.js +1 -1
  41. package/dist/ops/cpu/appendCache.js +8 -8
  42. package/dist/ops/cpu/attentionMask.js +9 -9
  43. package/dist/ops/cpu/fusedSoftmax.js +17 -11
  44. package/dist/ops/cpu/gatherSub.js +7 -7
  45. package/dist/ops/cpu/gelu.js +13 -13
  46. package/dist/ops/cpu/matMulGelu.js +36 -24
  47. package/dist/ops/cpu/matMulMul.js +14 -8
  48. package/dist/ops/cpu/mulDropout.js +9 -3
  49. package/dist/ops/cpu/normRMS.js +5 -5
  50. package/dist/ops/cpu/qkv.js +3 -3
  51. package/dist/ops/cpu/rope.js +5 -5
  52. package/dist/ops/cpu/scatterSub.js +11 -11
  53. package/dist/ops/fusedSoftmax.js +1 -1
  54. package/dist/ops/gatherSub.js +1 -1
  55. package/dist/ops/gelu.js +2 -2
  56. package/dist/ops/grads/attentionMask.js +1 -1
  57. package/dist/ops/grads/fusedSoftmax.js +2 -2
  58. package/dist/ops/grads/gelu.js +3 -24
  59. package/dist/ops/grads/matMulGelu.js +5 -5
  60. package/dist/ops/grads/normRMS.js +6 -6
  61. package/dist/ops/grads/qkv.js +1 -1
  62. package/dist/ops/grads/rope.js +3 -3
  63. package/dist/ops/matMulGelu.js +1 -1
  64. package/dist/ops/matMulMul.js +1 -1
  65. package/dist/ops/mulDrop.js +1 -1
  66. package/dist/ops/normRMS.js +1 -1
  67. package/dist/ops/qkv.js +1 -1
  68. package/dist/ops/rope.js +4 -4
  69. package/dist/ops/scatterSub.js +1 -1
  70. package/dist/ops/webgl/appendCache.js +1 -1
  71. package/dist/ops/webgl/attentionMask.js +1 -1
  72. package/dist/ops/webgl/fusedSoftmax.js +4 -4
  73. package/dist/ops/webgl/gatherSub.js +1 -1
  74. package/dist/ops/webgl/gelu.js +2 -2
  75. package/dist/ops/webgl/log.js +5 -5
  76. package/dist/ops/webgl/matMulGelu.js +17 -17
  77. package/dist/ops/webgl/matMulMul.js +1 -1
  78. package/dist/ops/webgl/mulDropout.js +4 -4
  79. package/dist/ops/webgl/normRMS.js +2 -2
  80. package/dist/ops/webgl/qkv.js +1 -1
  81. package/dist/ops/webgl/rope.js +1 -1
  82. package/dist/ops/webgl/scatterSub.js +1 -1
  83. package/dist/ops/webgpu/appendCache.d.ts +1 -0
  84. package/dist/ops/webgpu/appendCache.js +56 -0
  85. package/dist/ops/webgpu/attentionMask.d.ts +1 -0
  86. package/dist/ops/webgpu/attentionMask.js +64 -0
  87. package/dist/ops/webgpu/gatherSub.d.ts +1 -0
  88. package/dist/ops/webgpu/gatherSub.js +37 -0
  89. package/dist/ops/webgpu/gelu.d.ts +14 -0
  90. package/dist/ops/webgpu/gelu.js +86 -0
  91. package/dist/ops/webgpu/index.d.ts +0 -0
  92. package/dist/ops/webgpu/index.js +8 -0
  93. package/dist/ops/webgpu/normRMS.d.ts +1 -0
  94. package/dist/ops/webgpu/normRMS.js +115 -0
  95. package/dist/ops/webgpu/qkv.d.ts +1 -0
  96. package/dist/ops/webgpu/qkv.js +56 -0
  97. package/dist/ops/webgpu/rope.d.ts +1 -0
  98. package/dist/ops/webgpu/rope.js +68 -0
  99. package/dist/ops/webgpu/scatterSub.d.ts +1 -0
  100. package/dist/ops/webgpu/scatterSub.js +37 -0
  101. package/dist/{ops-C0sQEcPw.js → ops-CIQLNshk.js} +452 -503
  102. package/dist/{random_width-DWzaOgrn.js → random_width-DkYP8W8N.js} +143 -144
  103. package/dist/{range-DYsrnfiy.js → range-CYzpQY53.js} +1 -1
  104. package/dist/{reciprocal-CJQeasVa.js → reciprocal-_A9yv27J.js} +1 -1
  105. package/dist/{register_all_kernels-BfFCQAqs.js → register_all_kernels-guvSxp7M.js} +202 -200
  106. package/dist/{reshape-krWGKraP.js → reshape-BMUzc1UY.js} +3 -3
  107. package/dist/{scatter_nd_util-93ln7Hut.js → scatter_nd_util-IRBqKz_b.js} +3 -3
  108. package/dist/{selu_util-sntGesxr.js → selu_util-Dt_iuXaq.js} +6 -6
  109. package/dist/shared-BNa2q6jD.js +69 -0
  110. package/dist/{shared-Ca6iDobD.js → shared-CDu9S76h.js} +541 -606
  111. package/dist/{sin-D_h-qCSx.js → sin-Cocju-BY.js} +6 -6
  112. package/dist/{softmax-fsdtf6JC.js → softmax-GPNK3o-U.js} +3 -3
  113. package/dist/{split-eiktj-6L.js → split-CHzJjxDv.js} +4 -4
  114. package/dist/{stack-dfEEz2OY.js → stack-Dpgg_1W1.js} +2 -2
  115. package/dist/{sum-BE_Irnim.js → sum-B8wEpKsg.js} +5 -5
  116. package/dist/{tensor-Xyi595sG.js → tensor-RvZVNmg0.js} +1 -1
  117. package/dist/{tensor2d-CPEkynbH.js → tensor2d-B_kyod7_.js} +1 -1
  118. package/dist/training/AdamExt.js +1 -1
  119. package/dist/training/DatasetBuilder.js +2 -2
  120. package/dist/training/Evaluator.js +1 -1
  121. package/dist/training/FullTrainer.js +20 -20
  122. package/dist/training/Trainer.d.ts +5 -6
  123. package/dist/training/Trainer.js +59 -60
  124. package/dist/training/sparseCrossEntropy.js +4 -4
  125. package/dist/utilities/dummy.js +19 -19
  126. package/dist/utilities/generate.js +15 -16
  127. package/dist/utilities/multinomialCPU.d.ts +2 -0
  128. package/dist/utilities/multinomialCPU.js +13 -0
  129. package/dist/utilities/performance.d.ts +2 -0
  130. package/dist/utilities/performance.js +16 -0
  131. package/dist/utilities/profile.d.ts +1 -0
  132. package/dist/utilities/profile.js +9 -6
  133. package/dist/utilities/safetensors.js +2 -2
  134. package/dist/utilities/weights.js +2 -2
  135. package/dist/{variable-wSS22xj5.js → variable-DXEUOwew.js} +1 -1
  136. package/dist/webgpu_util-g13LvDIv.js +625 -0
  137. package/dist/{zeros-YJDE7oRb.js → zeros-DCPCdFGq.js} +8 -8
  138. package/package.json +2 -1
  139. package/dist/gpgpu_math-CNslybmD.js +0 -3115
  140. package/dist/norm-CzltS9Fz.js +0 -86
@@ -1,13 +1,13 @@
1
- import { r as d } from "../../index-BAzbokzv.js";
2
- import { c as h } from "../../concat-5aPGqw3Z.js";
1
+ import { f as d } from "../../index-C0dhsYom.js";
2
+ import { c as h } from "../../concat-CsHeR4zV.js";
3
3
  function u(p) {
4
- const { cache: n, item: s } = p.inputs, { maxSize: r, pastLen: c } = p.attrs, t = n.shape[0], o = n.shape[1], a = n.shape[3], e = s.shape[2];
5
- if (c + e <= r) {
6
- const f = n.slice([0, 0, 0, 0], [t, o, c, a]), m = n.slice([0, 0, c + e, 0], [t, o, r - c - e, a]), i = e < e ? s.slice([0, 0, 0, 0], [t, o, e, a]) : s, k = h([f, i, m], 2);
7
- return f.dispose(), m.dispose(), i !== s && i.dispose(), k;
4
+ const { cache: n, item: s } = p.inputs, { maxSize: i, pastLen: c } = p.attrs, t = n.shape[0], o = n.shape[1], a = n.shape[3], e = s.shape[2];
5
+ if (c + e <= i) {
6
+ const l = n.slice([0, 0, 0, 0], [t, o, c, a]), m = n.slice([0, 0, c + e, 0], [t, o, i - c - e, a]), r = e < e ? s.slice([0, 0, 0, 0], [t, o, e, a]) : s, k = h([l, r, m], 2);
7
+ return l.dispose(), m.dispose(), r !== s && r.dispose(), k;
8
8
  }
9
- const l = n.slice([0, 0, e, 0], [t, o, r - e, a]), C = h([l, s], 2);
10
- return l.dispose(), C;
9
+ const f = n.slice([0, 0, e, 0], [t, o, i - e, a]), C = h([f, s], 2);
10
+ return f.dispose(), C;
11
11
  }
12
12
  const w = {
13
13
  kernelName: "AppendCache",
@@ -1,21 +1,21 @@
1
- import { r as a, f as p, s as u } from "../../index-BAzbokzv.js";
2
- import { l as N, w as b } from "../../ops-C0sQEcPw.js";
3
- import { o as A } from "../../ones-BBlSRqn1.js";
4
- import { z as I } from "../../zeros-YJDE7oRb.js";
5
- import { m as g } from "../../mat_mul-7121rsJk.js";
1
+ import { f as a, h as p, b as u } from "../../index-C0dhsYom.js";
2
+ import { l as N, w as b } from "../../ops-CIQLNshk.js";
3
+ import { o as A } from "../../ones-D63HpSF_.js";
4
+ import { z as I } from "../../zeros-DCPCdFGq.js";
5
+ import { m as g } from "../../mat_mul-1nwdPkQ_.js";
6
6
  function o(n) {
7
7
  const { q: s, k: e } = n.inputs, { divisor: r } = n.attrs, c = s.shape[2], t = e.shape[2], m = N.bandPart(A([t, t]), -1, 0).cast("bool"), l = I([t, t]), i = p([t, t], Number.NEGATIVE_INFINITY), f = b(m, l, i), k = g(s, e, !1, !0).mul(u(r)), d = f.slice([0, 0], [c, t]).expandDims(0).expandDims(0);
8
8
  return k.add(d);
9
9
  }
10
- const w = {
10
+ const h = {
11
11
  kernelName: "AttentionMask",
12
12
  backendName: "cpu",
13
13
  kernelFunc: o
14
14
  };
15
- a(w);
16
- const M = {
15
+ a(h);
16
+ const w = {
17
17
  kernelName: "AttentionMask",
18
18
  backendName: "tensorflow",
19
19
  kernelFunc: o
20
20
  };
21
- a(M);
21
+ a(w);
@@ -1,23 +1,29 @@
1
- import { r as n } from "../../index-BAzbokzv.js";
2
- import { s as f } from "../../softmax-fsdtf6JC.js";
3
- function r(t) {
4
- const { inputs: s, attrs: i } = t, { logits: o } = s, { dim: a, dropoutRate: e } = i;
1
+ import { f as e } from "../../index-C0dhsYom.js";
2
+ import { s as f } from "../../softmax-GPNK3o-U.js";
3
+ function n(t) {
4
+ const { inputs: s, attrs: a } = t, { logits: o } = s, { dim: i, dropoutRate: r } = a;
5
5
  if (!o)
6
6
  throw new Error("Error in softmax: input logits is null");
7
- return e !== void 0 && e > 0 && console.warn("Dropout in fusedSoftmax not implemented for CPU backend, skipping dropout."), f(o, a);
7
+ return r !== void 0 && r > 0 && console.warn("Dropout in fusedSoftmax not implemented for CPU backend, skipping dropout."), f(o, i);
8
8
  }
9
9
  const m = {
10
10
  kernelName: "FusedSoftmax",
11
11
  backendName: "cpu",
12
- kernelFunc: r
12
+ kernelFunc: n
13
13
  };
14
- n(m);
15
- const l = {
14
+ e(m);
15
+ const u = {
16
16
  kernelName: "FusedSoftmax",
17
17
  backendName: "tensorflow",
18
- kernelFunc: r
18
+ kernelFunc: n
19
+ };
20
+ e(u);
21
+ const l = {
22
+ kernelName: "FusedSoftmax",
23
+ backendName: "webgpu",
24
+ kernelFunc: n
19
25
  };
20
- n(l);
26
+ e(l);
21
27
  export {
22
- r as softmaxCPU
28
+ n as softmaxCPU
23
29
  };
@@ -1,6 +1,6 @@
1
- import { o as u, q as c, E as g, Y as h, r as m, b as p } from "../../index-BAzbokzv.js";
2
- import { r as l } from "../../range-DYsrnfiy.js";
3
- import { s as N } from "../../stack-dfEEz2OY.js";
1
+ import { x as u, y as c, E as g, $ as h, f as m, c as p } from "../../index-C0dhsYom.js";
2
+ import { r as f } from "../../range-CYzpQY53.js";
3
+ import { s as l } from "../../stack-Dpgg_1W1.js";
4
4
  /**
5
5
  * @license
6
6
  * Copyright 2018 Google LLC. All Rights Reserved.
@@ -17,14 +17,14 @@ import { s as N } from "../../stack-dfEEz2OY.js";
17
17
  * limitations under the License.
18
18
  * =============================================================================
19
19
  */
20
- function b(e, t) {
20
+ function N(e, t) {
21
21
  const n = c(t, "indices", "gatherND", "int32"), s = { params: c(e, "x", "gatherND", "string_or_numeric"), indices: n };
22
22
  return g.runKernel(h, s);
23
23
  }
24
- const f = /* @__PURE__ */ u({ gatherND_: b });
24
+ const b = /* @__PURE__ */ u({ gatherND_: N });
25
25
  function d(e) {
26
- const { values: t, labels: n, logits: r } = e.inputs, s = n.shape[0], o = l(0, s, 1, "int32"), a = N([o, n], 1), i = f(r, a);
27
- return p(t, i);
26
+ const { values: t, labels: n, logits: r } = e.inputs, s = n.shape[0], a = f(0, s, 1, "int32"), i = l([a, n], 1), o = b(r, i);
27
+ return p(t, o);
28
28
  }
29
29
  const k = {
30
30
  kernelName: "EfficientGatherSub",
@@ -1,29 +1,29 @@
1
- import { r as t, t as d } from "../../index-BAzbokzv.js";
1
+ import { f as t, t as d } from "../../index-C0dhsYom.js";
2
2
  const o = 0.7978845608028654, c = 0.044715;
3
- function m(u) {
4
- const { inputs: l } = u, { x: e } = l, n = e;
3
+ function m(r) {
4
+ const { inputs: u } = r, { x: n } = u, e = n;
5
5
  return d(() => {
6
- const r = n.mul(n).mul(n), s = n.add(r.mul(c)).mul(o).tanh().add(1).mul(0.5);
7
- return n.mul(s);
6
+ const l = e.pow(3), s = e.add(l.mul(c)).mul(o).tanh().add(1).mul(0.5);
7
+ return e.mul(s);
8
8
  });
9
9
  }
10
- const K = {
10
+ const p = {
11
11
  kernelName: "Gelu",
12
12
  backendName: "cpu",
13
13
  kernelFunc: m
14
14
  };
15
- t(K);
16
- const p = {
15
+ t(p);
16
+ const K = {
17
17
  kernelName: "Gelu",
18
18
  backendName: "tensorflow",
19
19
  kernelFunc: m
20
20
  };
21
- t(p);
22
- function i(u) {
23
- const { dy: l, x: e } = u.inputs;
21
+ t(K);
22
+ function i(r) {
23
+ const { dy: u, x: n } = r.inputs;
24
24
  return d(() => {
25
- const n = e.square(), r = n.mul(e), a = e.add(r.mul(c)).mul(o).tanh(), f = a.square().neg().add(1), k = n.mul(3 * c).add(1), g = a.add(1).mul(0.5), G = e.mul(f).mul(o).mul(k).mul(0.5), N = g.add(G);
26
- return l.mul(N);
25
+ const e = n.square(), l = e.mul(n), a = n.add(l.mul(c)).mul(o).tanh(), f = a.square().neg().add(1), k = e.mul(3 * c).add(1), g = a.add(1).mul(0.5), G = n.mul(f).mul(o).mul(k).mul(0.5), N = g.add(G);
26
+ return u.mul(N);
27
27
  });
28
28
  }
29
29
  const x = {
@@ -1,40 +1,52 @@
1
- import { r as a, t as i } from "../../index-BAzbokzv.js";
2
- const c = 0.7978845608028654, m = 0.044715;
3
- function M(o) {
4
- const { inputs: s } = o, { x: t, kernel: l } = s, e = t, u = l;
5
- return i(() => {
6
- const n = e.matMul(u), d = n.mul(n).mul(n), r = n.add(d.mul(m)).mul(c).tanh().add(1).mul(0.5);
7
- return n.mul(r);
1
+ import { f as e, t as m } from "../../index-C0dhsYom.js";
2
+ import { g as d, d as M } from "../../gelu-uyHP1x1f.js";
3
+ function c(t) {
4
+ const { inputs: u } = t, { x: n, kernel: l } = u, a = n, r = l;
5
+ return m(() => {
6
+ const o = a.matMul(r);
7
+ return d(o);
8
8
  });
9
9
  }
10
- const b = {
10
+ const G = {
11
11
  kernelName: "MatMulGelu",
12
12
  backendName: "cpu",
13
- kernelFunc: M
13
+ kernelFunc: c
14
14
  };
15
- a(b);
16
- const w = {
15
+ e(G);
16
+ const i = {
17
17
  kernelName: "MatMulGelu",
18
18
  backendName: "tensorflow",
19
- kernelFunc: M
19
+ kernelFunc: c
20
20
  };
21
- a(w);
22
- function f(o) {
23
- const { dy: s, x: t, kernel: l } = o.inputs;
24
- return i(() => {
25
- const e = t.matMul(l), u = e.square(), n = u.mul(e), r = e.add(n.mul(m)).mul(c).tanh(), G = r.square().neg().add(1), g = u.mul(3 * m).add(1), p = r.add(1).mul(0.5), N = e.mul(G).mul(c).mul(g).mul(0.5), K = p.add(N), k = s.mul(K), h = k.matMul(l.transpose()), F = t.transpose().matMul(k);
26
- return [h, F];
21
+ e(i);
22
+ const f = {
23
+ kernelName: "MatMulGelu",
24
+ backendName: "webgpu",
25
+ kernelFunc: c
26
+ };
27
+ e(f);
28
+ function s(t) {
29
+ const { dy: u, x: n, kernel: l } = t.inputs;
30
+ return m(() => {
31
+ const a = n.matMul(l), r = M(u, a), o = r.matMul(l.transpose()), k = n.transpose().matMul(r);
32
+ return [o, k];
27
33
  });
28
34
  }
29
- const x = {
35
+ const g = {
30
36
  kernelName: "MatMulGeluGrad",
31
37
  backendName: "cpu",
32
- kernelFunc: f
38
+ kernelFunc: s
33
39
  };
34
- a(x);
35
- const C = {
40
+ e(g);
41
+ const p = {
36
42
  kernelName: "MatMulGeluGrad",
37
43
  backendName: "tensorflow",
38
- kernelFunc: f
44
+ kernelFunc: s
45
+ };
46
+ e(p);
47
+ const N = {
48
+ kernelName: "MatMulGeluGrad",
49
+ backendName: "webgpu",
50
+ kernelFunc: s
39
51
  };
40
- a(C);
52
+ e(N);
@@ -1,17 +1,23 @@
1
- import { r as n, t as M } from "../../index-BAzbokzv.js";
2
- function e(t) {
3
- const { inputs: r, attrs: o } = t, { transposeA: s, transposeB: l } = o, { x: c, kernel: u, y: a } = r, m = c, i = u, k = a;
4
- return M(() => m.matMul(i, s, l).mul(k));
1
+ import { f as e, t as i } from "../../index-C0dhsYom.js";
2
+ function n(t) {
3
+ const { inputs: r, attrs: l } = t, { transposeA: o, transposeB: s } = l, { x: u, kernel: a, y: c } = r, m = u, k = a, M = c;
4
+ return i(() => m.matMul(k, o, s).mul(M));
5
5
  }
6
6
  const f = {
7
7
  kernelName: "MatMulMul",
8
8
  backendName: "cpu",
9
- kernelFunc: e
9
+ kernelFunc: n
10
10
  };
11
- n(f);
11
+ e(f);
12
12
  const p = {
13
13
  kernelName: "MatMulMul",
14
14
  backendName: "tensorflow",
15
- kernelFunc: e
15
+ kernelFunc: n
16
16
  };
17
- n(p);
17
+ e(p);
18
+ const g = {
19
+ kernelName: "MatMulMul",
20
+ backendName: "webgpu",
21
+ kernelFunc: n
22
+ };
23
+ e(g);
@@ -1,7 +1,7 @@
1
- import { r as e, a as u } from "../../index-BAzbokzv.js";
1
+ import { f as e, a as t } from "../../index-C0dhsYom.js";
2
2
  function n(o) {
3
- const { inputs: r } = o, { a: l, b: t } = r;
4
- return console.warn("Using fallback mulDrop implementation without dropout."), u(l, t);
3
+ const { inputs: r } = o, { a: l, b: u } = r;
4
+ return console.warn("Using fallback mulDrop implementation without dropout."), t(l, u);
5
5
  }
6
6
  const a = {
7
7
  kernelName: "MulDropout",
@@ -15,3 +15,9 @@ const c = {
15
15
  kernelFunc: n
16
16
  };
17
17
  e(c);
18
+ const m = {
19
+ kernelName: "MulDropout",
20
+ backendName: "webgpu",
21
+ kernelFunc: n
22
+ };
23
+ e(m);
@@ -1,4 +1,4 @@
1
- import { r as o, t as d } from "../../index-BAzbokzv.js";
1
+ import { f as o, t as d } from "../../index-C0dhsYom.js";
2
2
  function i(t) {
3
3
  const { inputs: e } = t, { x: n, gamma: s } = e, r = n, a = s;
4
4
  return d(() => {
@@ -6,12 +6,12 @@ function i(t) {
6
6
  return r.mul(u).mul(a);
7
7
  });
8
8
  }
9
- const f = {
9
+ const k = {
10
10
  kernelName: "RMSNorm",
11
11
  backendName: "cpu",
12
12
  kernelFunc: i
13
13
  };
14
- o(f);
14
+ o(k);
15
15
  const g = {
16
16
  kernelName: "RMSNorm",
17
17
  backendName: "tensorflow",
@@ -21,8 +21,8 @@ o(g);
21
21
  function N(t) {
22
22
  const { dy: e, x: n, gamma: s } = t.inputs;
23
23
  return d(() => {
24
- const r = n.shape[n.shape.length - 1], a = n.square().mean(-1, !0), m = a.add(1e-8).rsqrt(), u = n.mul(m), l = e.mul(u).sum([0, 1]), c = e.mul(s), k = c.mul(n).sum(-1, !0).div(r);
25
- return [c.mul(m).sub(n.mul(k).mul(m).div(a.add(1e-8))), l];
24
+ const r = n.shape[n.shape.length - 1], a = n.square().mean(-1, !0), m = a.add(1e-8).rsqrt(), u = n.mul(m), l = e.mul(u).sum([0, 1]), c = e.mul(s), f = c.mul(n).sum(-1, !0).div(r);
25
+ return [c.mul(m).sub(n.mul(f).mul(m).div(a.add(1e-8))), l];
26
26
  });
27
27
  }
28
28
  const S = {
@@ -1,6 +1,6 @@
1
- import { r as q } from "../../index-BAzbokzv.js";
2
- import { r as o } from "../../reshape-krWGKraP.js";
3
- import { s as x } from "../../split-eiktj-6L.js";
1
+ import { f as q } from "../../index-C0dhsYom.js";
2
+ import { r as o } from "../../reshape-BMUzc1UY.js";
3
+ import { s as x } from "../../split-CHzJjxDv.js";
4
4
  function v(p) {
5
5
  const { x: c, kernel: K } = p.inputs, { heads: n } = p.attrs, [s, e, t] = c.shape, a = o(c, [s * e, t]), i = a.dot(K);
6
6
  a.dispose();
@@ -1,8 +1,8 @@
1
- import { r as S } from "../../index-BAzbokzv.js";
2
- import { r as F } from "../../range-DYsrnfiy.js";
3
- import { g as I } from "../../gather-DjyCjmOD.js";
4
- import { s as E } from "../../stack-dfEEz2OY.js";
5
- import { c as T } from "../../concat-5aPGqw3Z.js";
1
+ import { f as S } from "../../index-C0dhsYom.js";
2
+ import { r as F } from "../../range-CYzpQY53.js";
3
+ import { g as I } from "../../gather-D0_gPiBz.js";
4
+ import { s as E } from "../../stack-Dpgg_1W1.js";
5
+ import { c as T } from "../../concat-CsHeR4zV.js";
6
6
  function U(t, c, p, o, r) {
7
7
  const n = o.shape[3], s = p;
8
8
  if (s > n) return o;
@@ -1,8 +1,8 @@
1
- import { o as f, x as g, q as r, E as l, X as N, r as b, b as S, a as h } from "../../index-BAzbokzv.js";
2
- import { v as D } from "../../scatter_nd_util-93ln7Hut.js";
3
- import { r as k } from "../../range-DYsrnfiy.js";
4
- import { s as v } from "../../stack-dfEEz2OY.js";
5
- import { o as E } from "../../ones-BBlSRqn1.js";
1
+ import { x as f, C as g, y as r, E as l, _ as N, f as b, c as S, a as h } from "../../index-C0dhsYom.js";
2
+ import { v as D } from "../../scatter_nd_util-IRBqKz_b.js";
3
+ import { r as k } from "../../range-CYzpQY53.js";
4
+ import { s as v } from "../../stack-Dpgg_1W1.js";
5
+ import { o as E } from "../../ones-D63HpSF_.js";
6
6
  /**
7
7
  * @license
8
8
  * Copyright 2018 Google LLC. All Rights Reserved.
@@ -26,14 +26,14 @@ function I(a, e, s) {
26
26
  const c = { indices: n, updates: t }, o = { shape: s };
27
27
  return l.runKernel(N, c, o);
28
28
  }
29
- const K = /* @__PURE__ */ f({ scatterND_: I });
30
- function L(a) {
31
- const { logits: e, labels: s, dy: n } = a.inputs, t = s.shape[0], c = e.shape[1], o = k(0, t, 1, "int32"), i = v([o, s], 1), d = E([t]), u = K(i, d, [t, c]), p = S(e, u), m = n.reshape([t, 1]);
29
+ const y = /* @__PURE__ */ f({ scatterND_: I });
30
+ function C(a) {
31
+ const { logits: e, labels: s, dy: n } = a.inputs, t = s.shape[0], c = e.shape[1], o = k(0, t, 1, "int32"), i = v([o, s], 1), d = E([t]), u = y(i, d, [t, c]), p = S(e, u), m = n.reshape([t, 1]);
32
32
  return h(p, m);
33
33
  }
34
- const T = {
34
+ const K = {
35
35
  kernelName: "EfficientScatterSub",
36
36
  backendName: "cpu",
37
- kernelFunc: L
37
+ kernelFunc: C
38
38
  };
39
- b(T);
39
+ b(K);
@@ -1,4 +1,4 @@
1
- import { e as t } from "../index-BAzbokzv.js";
1
+ import { e as t } from "../index-C0dhsYom.js";
2
2
  import "./cpu/fusedSoftmax.js";
3
3
  import "./webgl/fusedSoftmax.js";
4
4
  import "./grads/fusedSoftmax.js";
@@ -1,4 +1,4 @@
1
- import { e as n } from "../index-BAzbokzv.js";
1
+ import { e as n } from "../index-C0dhsYom.js";
2
2
  import "./cpu/gatherSub.js";
3
3
  import "./webgl/gatherSub.js";
4
4
  function f(r, e, t) {
package/dist/ops/gelu.js CHANGED
@@ -1,7 +1,7 @@
1
- import "../index-BAzbokzv.js";
1
+ import "../index-C0dhsYom.js";
2
2
  import "./cpu/gelu.js";
3
3
  import "./webgl/gelu.js";
4
- import { d as e, g as i } from "./grads/gelu.js";
4
+ import { d as e, g as i } from "../gelu-uyHP1x1f.js";
5
5
  export {
6
6
  e as dGelu,
7
7
  i as gelu
@@ -1,4 +1,4 @@
1
- import { c as m, s as i } from "../../index-BAzbokzv.js";
1
+ import { i as m, b as i } from "../../index-C0dhsYom.js";
2
2
  import { matMulMul as u } from "../matMulMul.js";
3
3
  const p = {
4
4
  kernelName: "AttentionMask",
@@ -1,6 +1,6 @@
1
- import { c as f, a as i, b as l } from "../../index-BAzbokzv.js";
1
+ import { i as f, a as i, c as l } from "../../index-C0dhsYom.js";
2
2
  import { mulDrop as g } from "../mulDrop.js";
3
- import { s as T } from "../../sum-BE_Irnim.js";
3
+ import { s as T } from "../../sum-B8wEpKsg.js";
4
4
  const Y = {
5
5
  kernelName: "FusedSoftmax",
6
6
  outputsToSave: [!0],
@@ -1,26 +1,5 @@
1
- import { c as t, e as n } from "../../index-BAzbokzv.js";
2
- import "../cpu/gelu.js";
3
- import "../webgl/gelu.js";
4
- const o = {
5
- kernelName: "Gelu",
6
- inputsToSave: ["x"],
7
- outputsToSave: [],
8
- gradFunc: (e, r) => {
9
- const [u] = r;
10
- return {
11
- x: () => a(e, u)
12
- };
13
- }
14
- };
15
- t(o);
16
- function g(e) {
17
- return n().runKernel("Gelu", { x: e });
18
- }
19
- function a(e, r) {
20
- return n().runKernel("GeluGrad", { dy: e, x: r });
21
- }
1
+ import "../../index-C0dhsYom.js";
2
+ import { a as e } from "../../gelu-uyHP1x1f.js";
22
3
  export {
23
- a as d,
24
- g,
25
- o as geluGradConfig
4
+ e as geluGradConfig
26
5
  };
@@ -1,17 +1,17 @@
1
- import { c as a, e as o } from "../../index-BAzbokzv.js";
2
- function s(e, n, r) {
1
+ import { i as a, e as o } from "../../index-C0dhsYom.js";
2
+ function i(e, n, r) {
3
3
  return o().runKernel("MatMulGeluGrad", { dy: e, x: n, kernel: r });
4
4
  }
5
- const d = {
5
+ const s = {
6
6
  kernelName: "MatMulGelu",
7
7
  inputsToSave: ["x", "kernel"],
8
8
  outputsToSave: [],
9
9
  gradFunc: (e, n) => {
10
- const [r, t] = n, [u, l] = s(e, r, t);
10
+ const [r, t] = n, [u, l] = i(e, r, t);
11
11
  return {
12
12
  x: () => u,
13
13
  kernel: () => l
14
14
  };
15
15
  }
16
16
  };
17
- a(d);
17
+ a(s);
@@ -1,20 +1,20 @@
1
- import { c as t, e as g } from "../../index-BAzbokzv.js";
2
- function s(r, a, n) {
1
+ import { i as t, e as g } from "../../index-C0dhsYom.js";
2
+ function i(r, a, n) {
3
3
  return g().runKernel("RMSNormGrad", { dy: r, x: a, gamma: n });
4
4
  }
5
- const u = {
5
+ const s = {
6
6
  kernelName: "RMSNorm",
7
7
  inputsToSave: ["x", "gamma"],
8
8
  outputsToSave: [],
9
9
  gradFunc: (r, a) => {
10
- const [n, e] = a, [m, o] = s(r, n, e);
10
+ const [n, e] = a, [m, o] = i(r, n, e);
11
11
  return {
12
12
  x: () => m,
13
13
  gamma: () => o
14
14
  };
15
15
  }
16
16
  };
17
- t(u);
17
+ t(s);
18
18
  export {
19
- u as normRMSGradConfig
19
+ s as normRMSGradConfig
20
20
  };
@@ -1,4 +1,4 @@
1
- import { c as Q } from "../../index-BAzbokzv.js";
1
+ import { i as Q } from "../../index-C0dhsYom.js";
2
2
  const V = {
3
3
  kernelName: "QKV",
4
4
  inputsToSave: ["x", "kernel"],
@@ -1,6 +1,6 @@
1
- import { c as a, e as i } from "../../index-BAzbokzv.js";
1
+ import { i, e as a } from "../../index-C0dhsYom.js";
2
2
  function p(n, e, s, o) {
3
- return i().runKernel("Rope", { x: n, sin: e, cos: s }, { pastLen: o });
3
+ return a().runKernel("Rope", { x: n, sin: e, cos: s }, { pastLen: o });
4
4
  }
5
5
  const c = {
6
6
  kernelName: "Rope",
@@ -11,4 +11,4 @@ const c = {
11
11
  return t.dispose(), { x: () => r };
12
12
  }
13
13
  };
14
- a(c);
14
+ i(c);
@@ -1,4 +1,4 @@
1
- import { e as u } from "../index-BAzbokzv.js";
1
+ import { e as u } from "../index-C0dhsYom.js";
2
2
  import "./cpu/matMulGelu.js";
3
3
  import "./webgl/matMulGelu.js";
4
4
  import "./grads/matMulGelu.js";
@@ -1,4 +1,4 @@
1
- import { e as u } from "../index-BAzbokzv.js";
1
+ import { e as u } from "../index-C0dhsYom.js";
2
2
  import "./cpu/matMulMul.js";
3
3
  import "./webgl/matMulMul.js";
4
4
  function m(e, r, t, l = !1, n = !1) {
@@ -1,4 +1,4 @@
1
- import { e as t } from "../index-BAzbokzv.js";
1
+ import { e as t } from "../index-C0dhsYom.js";
2
2
  import "./cpu/mulDropout.js";
3
3
  import "./webgl/mulDropout.js";
4
4
  function m(r, o, e, n) {
@@ -1,4 +1,4 @@
1
- import { e as n } from "../index-BAzbokzv.js";
1
+ import { e as n } from "../index-C0dhsYom.js";
2
2
  import "./cpu/normRMS.js";
3
3
  import "./webgl/normRMS.js";
4
4
  import "./grads/normRMS.js";
package/dist/ops/qkv.js CHANGED
@@ -1,4 +1,4 @@
1
- import { e as o } from "../index-BAzbokzv.js";
1
+ import { e as o } from "../index-C0dhsYom.js";
2
2
  import "./cpu/qkv.js";
3
3
  import "./webgl/qkv.js";
4
4
  import "./grads/qkv.js";
package/dist/ops/rope.js CHANGED
@@ -1,8 +1,8 @@
1
- import { e as p } from "../index-BAzbokzv.js";
2
- import "../random_width-DWzaOgrn.js";
3
- import "../register_all_kernels-BfFCQAqs.js";
1
+ import { e as p } from "../index-C0dhsYom.js";
2
+ import "../random_width-DkYP8W8N.js";
3
+ import "../register_all_kernels-guvSxp7M.js";
4
4
  import "../index-Tf7vU29b.js";
5
- import "../dataset-pgqp-YfL.js";
5
+ import "../dataset-JDyjG3QR.js";
6
6
  import "./cpu/rope.js";
7
7
  import "./webgl/rope.js";
8
8
  import "./grads/rope.js";
@@ -1,4 +1,4 @@
1
- import { e as i } from "../index-BAzbokzv.js";
1
+ import { e as i } from "../index-C0dhsYom.js";
2
2
  import "./cpu/scatterSub.js";
3
3
  import "./webgl/scatterSub.js";
4
4
  function c(t, r, e) {
@@ -1,4 +1,4 @@
1
- import { r as p } from "../../index-BAzbokzv.js";
1
+ import { f as p } from "../../index-C0dhsYom.js";
2
2
  class m {
3
3
  variableNames = ["cache", "item"];
4
4
  outputShape;
@@ -1,4 +1,4 @@
1
- import { r as m } from "../../index-BAzbokzv.js";
1
+ import { f as m } from "../../index-C0dhsYom.js";
2
2
  class h {
3
3
  variableNames = ["q", "k"];
4
4
  outputShape;
@@ -1,7 +1,7 @@
1
- import { m as b, s as I, r as k } from "../../RealDiv-BYViZwhN.js";
2
- import { r as v } from "../../Reshape-t7Kcikjk.js";
3
- import { r as w, p as P } from "../../index-BAzbokzv.js";
4
- import { e as S } from "../../axis_util-Bu4h7XWV.js";
1
+ import { m as b, s as I, r as k } from "../../RealDiv-C4hOvYOZ.js";
2
+ import { r as v } from "../../Reshape-BLijOA8h.js";
3
+ import { f as w, p as P } from "../../index-C0dhsYom.js";
4
+ import { e as S } from "../../axis_util-DaAl5MER.js";
5
5
  class T {
6
6
  variableNames = ["logits", "maxLogits"];
7
7
  outputShape;
@@ -1,4 +1,4 @@
1
- import { r as l } from "../../index-BAzbokzv.js";
1
+ import { f as l } from "../../index-C0dhsYom.js";
2
2
  class u {
3
3
  variableNames = ["labels", "logits", "values"];
4
4
  outputShape;