@genai-fi/nanogpt 0.10.1 → 0.10.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (206) hide show
  1. package/dist/Generator.js +14 -14
  2. package/dist/{RealDiv-DgA3z9oO.js → RealDiv-zz7FpkKX.js} +17 -17
  3. package/dist/{Reshape-CF6odzV4.js → Reshape-CDVLyVfz.js} +3 -3
  4. package/dist/{Reshape-_kILl6tK.js → Reshape-CHdUjC72.js} +4 -4
  5. package/dist/TeachableLLM.js +8 -8
  6. package/dist/{axis_util-BvHEw88j.js → axis_util-BsIr9ZNu.js} +1 -1
  7. package/dist/backend.js +2 -2
  8. package/dist/{backend_util-D-rUb2ty.js → backend_util-B1XRLuq9.js} +31 -31
  9. package/dist/{backend_webgpu-B0u2ndUn.js → backend_webgpu-CqpfEImu.js} +5 -5
  10. package/dist/{broadcast_to-CwF7XIeu.js → broadcast_to-B0ChcDaz.js} +4 -4
  11. package/dist/checks/appendCache.js +2 -2
  12. package/dist/checks/attentionMask.js +3 -3
  13. package/dist/checks/gelu.js +2 -2
  14. package/dist/checks/matMulGelu.js +5 -5
  15. package/dist/checks/normRMS.js +4 -4
  16. package/dist/checks/normRMSGrad.js +3 -3
  17. package/dist/checks/packUnpack.js +2 -2
  18. package/dist/checks/qkv.js +3 -3
  19. package/dist/checks/rope.js +2 -2
  20. package/dist/{complex-CSlYz-2T.js → complex-BBiRlsVq.js} +3 -3
  21. package/dist/{concat-BHlIJeyT.js → concat-DmBLPVGC.js} +3 -3
  22. package/dist/{concat_util-DcJk7YHS.js → concat_util-iBYIyuQe.js} +1 -1
  23. package/dist/{dataset-0xP8GjwI.js → dataset-D2P7rHAw.js} +5 -5
  24. package/dist/{dropout-C1pM3f11.js → dropout-B1x1kYMa.js} +3 -3
  25. package/dist/{expand_dims-BPG4fwBP.js → expand_dims-ouvfxQ1n.js} +3 -3
  26. package/dist/{exports_initializers-xuidcwI4.js → exports_initializers-CZSUJoVE.js} +1 -1
  27. package/dist/{gather-DykLGqmW.js → gather-CH9sdacz.js} +2 -2
  28. package/dist/{gelu-CNLFZWea.js → gelu-Bmhopi0J.js} +2 -2
  29. package/dist/{gpgpu_math-DDVJCn6-.js → gpgpu_math-DsCcikas.js} +3 -3
  30. package/dist/{index-ZyQhjEPo.js → index-D6Q1lPZO.js} +55 -55
  31. package/dist/{index-CjOj7j-u.js → index-DRyE072i.js} +15 -15
  32. package/dist/{kernel_funcs_utils-Dg_-E44D.js → kernel_funcs_utils-CWfOAPGO.js} +9 -9
  33. package/dist/layers/BaseLayer.js +10 -10
  34. package/dist/layers/CausalSelfAttention.js +6 -6
  35. package/dist/layers/MLP.js +4 -4
  36. package/dist/layers/PositionEmbedding.js +5 -5
  37. package/dist/layers/RMSNorm.js +3 -3
  38. package/dist/layers/RoPECache.js +4 -4
  39. package/dist/layers/TiedEmbedding.js +6 -6
  40. package/dist/layers/TransformerBlock.js +1 -1
  41. package/dist/loader/loadTransformers.js +1 -1
  42. package/dist/loader/oldZipLoad.js +8 -8
  43. package/dist/{log_sum_exp-DWI-76TI.js → log_sum_exp-D3ftBNY5.js} +6 -6
  44. package/dist/main.js +8 -8
  45. package/dist/{matMul16--R5hOwDG.js → matMul16-fEAJ4smh.js} +4 -4
  46. package/dist/{mat_mul-DeAh4uTH.js → mat_mul-C59XWcJd.js} +2 -2
  47. package/dist/{mod-Gt1rMB4n.js → mod-DESSvHIU.js} +2 -2
  48. package/dist/models/NanoGPTV1.js +2 -2
  49. package/dist/models/model.js +8 -8
  50. package/dist/{mulmat_packed_gpu-BMFhLwta.js → mulmat_packed_gpu-Coh6qbJk.js} +1 -1
  51. package/dist/{ones-CAMiP4I2.js → ones-jU9jlQvM.js} +4 -4
  52. package/dist/ops/adamAdjust.js +1 -1
  53. package/dist/ops/adamMoments.js +1 -1
  54. package/dist/ops/add16.js +1 -1
  55. package/dist/ops/appendCache.js +3 -3
  56. package/dist/ops/attentionMask.js +1 -1
  57. package/dist/ops/concat16.js +2 -2
  58. package/dist/ops/cpu/adamAdjust.js +2 -2
  59. package/dist/ops/cpu/adamMoments.js +3 -3
  60. package/dist/ops/cpu/appendCache.js +3 -3
  61. package/dist/ops/cpu/attentionMask.js +6 -6
  62. package/dist/ops/cpu/fusedSoftmax.js +3 -3
  63. package/dist/ops/cpu/gatherSub.js +4 -4
  64. package/dist/ops/cpu/gelu.js +2 -2
  65. package/dist/ops/cpu/matMul16.js +3 -3
  66. package/dist/ops/cpu/matMulGelu.js +4 -4
  67. package/dist/ops/cpu/matMulMul.js +2 -2
  68. package/dist/ops/cpu/mulDropout.js +2 -2
  69. package/dist/ops/cpu/normRMS.js +2 -2
  70. package/dist/ops/cpu/qkv.js +4 -4
  71. package/dist/ops/cpu/rope.js +6 -6
  72. package/dist/ops/cpu/scatterSub.js +7 -7
  73. package/dist/ops/dot16.js +2 -2
  74. package/dist/ops/gatherSub.js +1 -1
  75. package/dist/ops/gelu.js +2 -2
  76. package/dist/ops/grads/add16.js +2 -2
  77. package/dist/ops/grads/attentionMask.js +3 -3
  78. package/dist/ops/grads/gelu.js +3 -3
  79. package/dist/ops/grads/matMul16.js +4 -4
  80. package/dist/ops/grads/matMulGelu.js +2 -2
  81. package/dist/ops/grads/normRMS.js +2 -2
  82. package/dist/ops/grads/pack16.js +4 -4
  83. package/dist/ops/grads/qkv.js +4 -4
  84. package/dist/ops/grads/rope.js +3 -3
  85. package/dist/ops/grads/softmax16.js +2 -2
  86. package/dist/ops/grads/unpack16.js +3 -3
  87. package/dist/ops/matMul16.js +3 -3
  88. package/dist/ops/matMulGelu.js +1 -1
  89. package/dist/ops/matMulMul.js +1 -1
  90. package/dist/ops/mul16.js +1 -1
  91. package/dist/ops/mulDrop.js +1 -1
  92. package/dist/ops/normRMS.js +1 -1
  93. package/dist/ops/pack16.js +2 -2
  94. package/dist/ops/qkv.js +1 -1
  95. package/dist/ops/reshape16.js +3 -3
  96. package/dist/ops/rope.js +5 -5
  97. package/dist/ops/scatterSub.js +1 -1
  98. package/dist/ops/slice16.js +2 -2
  99. package/dist/ops/softmax16.js +1 -1
  100. package/dist/ops/sub16.js +1 -1
  101. package/dist/ops/sum16.js +2 -2
  102. package/dist/ops/transpose16.js +4 -4
  103. package/dist/ops/unpack16.js +2 -2
  104. package/dist/ops/webgl/adamAdjust.js +3 -3
  105. package/dist/ops/webgl/adamMoments.js +2 -2
  106. package/dist/ops/webgl/appendCache.js +2 -2
  107. package/dist/ops/webgl/attentionMask.js +2 -2
  108. package/dist/ops/webgl/fusedSoftmax.js +6 -6
  109. package/dist/ops/webgl/gatherSub.js +2 -2
  110. package/dist/ops/webgl/gelu.js +3 -3
  111. package/dist/ops/webgl/log.js +4 -4
  112. package/dist/ops/webgl/matMul16.js +5 -5
  113. package/dist/ops/webgl/matMulGelu.js +6 -6
  114. package/dist/ops/webgl/matMulMul.js +2 -2
  115. package/dist/ops/webgl/mulDropout.js +2 -2
  116. package/dist/ops/webgl/normRMS.js +3 -3
  117. package/dist/ops/webgl/qkv.js +2 -2
  118. package/dist/ops/webgl/rope.js +2 -2
  119. package/dist/ops/webgl/scatterSub.js +2 -2
  120. package/dist/ops/webgpu/adamAdjust.js +5 -5
  121. package/dist/ops/webgpu/adamMoments.js +5 -5
  122. package/dist/ops/webgpu/add16.js +2 -2
  123. package/dist/ops/webgpu/appendCache.js +5 -5
  124. package/dist/ops/webgpu/attentionMask.js +4 -4
  125. package/dist/ops/webgpu/attentionMask32_program.js +2 -2
  126. package/dist/ops/webgpu/concat16.js +7 -7
  127. package/dist/ops/webgpu/gatherSub.js +5 -5
  128. package/dist/ops/webgpu/gelu.js +4 -4
  129. package/dist/ops/webgpu/matMul16.js +6 -6
  130. package/dist/ops/webgpu/matMul16_program.js +3 -3
  131. package/dist/ops/webgpu/mul16.js +2 -2
  132. package/dist/ops/webgpu/normRMS.js +4 -4
  133. package/dist/ops/webgpu/normRMSGrad.js +6 -6
  134. package/dist/ops/webgpu/pack16.js +2 -2
  135. package/dist/ops/webgpu/pack16_program.js +2 -2
  136. package/dist/ops/webgpu/qkv.js +4 -4
  137. package/dist/ops/webgpu/rope.js +5 -5
  138. package/dist/ops/webgpu/scatterSub.js +5 -5
  139. package/dist/ops/webgpu/slice16.js +6 -6
  140. package/dist/ops/webgpu/softmax16.js +4 -4
  141. package/dist/ops/webgpu/softmax16_program.js +2 -2
  142. package/dist/ops/webgpu/softmax16_subgroup_program.js +2 -2
  143. package/dist/ops/webgpu/softmax16grad.js +2 -2
  144. package/dist/ops/webgpu/sub16.js +2 -2
  145. package/dist/ops/webgpu/sum16.js +5 -5
  146. package/dist/ops/webgpu/transpose16.js +3 -3
  147. package/dist/ops/webgpu/transpose16_program.js +2 -2
  148. package/dist/ops/webgpu/transpose16_shared_program.js +4 -4
  149. package/dist/ops/webgpu/unpack16.js +4 -4
  150. package/dist/ops/webgpu/utils/binary_op.js +4 -4
  151. package/dist/ops/webgpu/utils/reductions.js +5 -5
  152. package/dist/{ops-CNI3TwqM.js → ops-BFDtP6th.js} +24 -24
  153. package/dist/{pack16-CFUqumar.js → pack16-CmVZs6af.js} +3 -3
  154. package/dist/patches/PackedTensor.js +1 -1
  155. package/dist/patches/engine.js +7 -5
  156. package/dist/patches/tape.js +1 -1
  157. package/dist/patches/webgpu_backend.js +5 -5
  158. package/dist/patches/webgpu_base.js +1 -1
  159. package/dist/patches/webgpu_program.js +3 -3
  160. package/dist/{random_width-DY6Kk2Dl.js → random_width-BVV9HveY.js} +31 -31
  161. package/dist/{range-BMS52eQi.js → range-ZZZD60Fx.js} +2 -2
  162. package/dist/{reciprocal-CTmshQ9J.js → reciprocal-CrYlsAGD.js} +2 -2
  163. package/dist/{register_all_kernels-Bwu1PTuU.js → register_all_kernels-nvj2k7OC.js} +41 -41
  164. package/dist/{relu-yZ2-7WxU.js → relu-BYDneVPn.js} +2 -2
  165. package/dist/{reshape-DevtBWtf.js → reshape-CaPQzFvz.js} +2 -2
  166. package/dist/{rope-B5UUMsPi.js → rope-s4W2XO9B.js} +5 -5
  167. package/dist/{scatter_nd_util-5EL-8VAQ.js → scatter_nd_util-C7zXRT_h.js} +1 -1
  168. package/dist/{selu_util-D1w6yyTO.js → selu_util-BGPXmd4B.js} +16 -16
  169. package/dist/{shared-BRksrJb3.js → shared-CHhxz-O5.js} +1 -1
  170. package/dist/{shared-BuAXb4CI.js → shared-D2NP_CpY.js} +8 -8
  171. package/dist/{sin-BGfy2HZo.js → sin-Djs4aQiu.js} +2 -2
  172. package/dist/{slice-D_gkkqZK.js → slice-DvovR5wq.js} +2 -2
  173. package/dist/{slice_util-DtEldBfK.js → slice_util-DyjSAD0u.js} +1 -1
  174. package/dist/{softmax-ZHVebtR1.js → softmax-C9JQEtnO.js} +2 -2
  175. package/dist/{split-DrfihRpZ.js → split-DBck65sX.js} +2 -2
  176. package/dist/{squeeze-DZEpeblb.js → squeeze-C00Ipm_7.js} +3 -3
  177. package/dist/{stack-yOIAalTq.js → stack-ChnHwRpX.js} +3 -3
  178. package/dist/{sum-_fzj5ZTB.js → sum-ywRJj3Zr.js} +2 -2
  179. package/dist/{tensor-f35l8Odg.js → tensor-0r5yOo2R.js} +1 -1
  180. package/dist/{tensor-DdQUJZlz.js → tensor-CzmOBsdf.js} +21 -21
  181. package/dist/{tensor1d-CeZuc-Rv.js → tensor1d-BlUT89BP.js} +2 -2
  182. package/dist/{tensor2d-G4Ys2GxX.js → tensor2d-CSB4KOb0.js} +2 -2
  183. package/dist/{tensor4d-B8roDgtc.js → tensor4d-D7bLqGqz.js} +2 -2
  184. package/dist/{tensor_util-DV-FP5Q3.js → tensor_util-DfwaWayG.js} +12 -12
  185. package/dist/{tfjs_backend-kNyO5L2d.js → tfjs_backend-CNkSTL0c.js} +38 -38
  186. package/dist/{tile-BzyEiF-F.js → tile-CR074jmp.js} +3 -3
  187. package/dist/training/Adam.js +2 -2
  188. package/dist/training/AdamExt.js +1 -1
  189. package/dist/training/DatasetBuilder.js +2 -2
  190. package/dist/training/FullTrainer.js +1 -1
  191. package/dist/training/Trainer.js +2 -2
  192. package/dist/training/sparseCrossEntropy.js +3 -3
  193. package/dist/{transpose-DKELTqhe.js → transpose-DH4gmHvu.js} +4 -4
  194. package/dist/utilities/dummy.js +3 -3
  195. package/dist/utilities/multinomialCPU.js +2 -2
  196. package/dist/utilities/packed.js +338 -304
  197. package/dist/utilities/performance.js +1 -1
  198. package/dist/utilities/profile.js +1 -1
  199. package/dist/utilities/safetensors.js +2 -2
  200. package/dist/utilities/sentences.js +5 -5
  201. package/dist/utilities/weights.js +2 -2
  202. package/dist/{variable-Bhn5bHYv.js → variable-DzfrwYuP.js} +1 -1
  203. package/dist/{webgpu_program-Cigz-7RF.js → webgpu_program-DzaQiqel.js} +2 -2
  204. package/dist/{webgpu_util-BBCnKm2X.js → webgpu_util-0_ubCEHJ.js} +2 -2
  205. package/dist/{zeros-2gldETuK.js → zeros-DBFVbpv5.js} +3 -3
  206. package/package.json +1 -1
@@ -1,6 +1,6 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { a as i } from "../../pack16-CFUqumar.js";
3
- import "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { a as i } from "../../pack16-CmVZs6af.js";
3
+ import "../../tensor_util-DfwaWayG.js";
4
4
  export {
5
5
  i as unpackGradConfig
6
6
  };
@@ -1,9 +1,9 @@
1
- import "../index-ZyQhjEPo.js";
2
- import { b as p, c as u, d as i, e as s, m as M } from "../matMul16--R5hOwDG.js";
1
+ import "../index-D6Q1lPZO.js";
2
+ import { b as p, c as u, d as i, e as s, m as M } from "../matMul16-fEAJ4smh.js";
3
3
  import "./webgl/matMul16.js";
4
4
  import "./cpu/matMul16.js";
5
5
  import "../utilities/packed.js";
6
- import "../pack16-CFUqumar.js";
6
+ import "../pack16-CmVZs6af.js";
7
7
  export {
8
8
  p as matMul16,
9
9
  u as matMul16Gelu,
@@ -1,4 +1,4 @@
1
- import { e as u } from "../index-ZyQhjEPo.js";
1
+ import { e as u } from "../index-D6Q1lPZO.js";
2
2
  import "./cpu/matMulGelu.js";
3
3
  import "./webgl/matMulGelu.js";
4
4
  import "./grads/matMulGelu.js";
@@ -1,4 +1,4 @@
1
- import { e as u } from "../index-ZyQhjEPo.js";
1
+ import { e as u } from "../index-D6Q1lPZO.js";
2
2
  import "./cpu/matMulMul.js";
3
3
  import "./webgl/matMulMul.js";
4
4
  function m(e, r, t, l = !1, n = !1) {
package/dist/ops/mul16.js CHANGED
@@ -1,4 +1,4 @@
1
- import { m as t, e as u } from "../index-ZyQhjEPo.js";
1
+ import { m as t, e as u } from "../index-D6Q1lPZO.js";
2
2
  import { isPackedTensor as n } from "../utilities/packed.js";
3
3
  function i(r, e) {
4
4
  return !n(r) && !n(e) ? t(r, e) : u().runKernel("Mul16", { a: r, b: e });
@@ -1,4 +1,4 @@
1
- import { e as t } from "../index-ZyQhjEPo.js";
1
+ import { e as t } from "../index-D6Q1lPZO.js";
2
2
  import "./cpu/mulDropout.js";
3
3
  import "./webgl/mulDropout.js";
4
4
  function m(r, o, e, n) {
@@ -1,4 +1,4 @@
1
- import { e as n } from "../index-ZyQhjEPo.js";
1
+ import { e as n } from "../index-D6Q1lPZO.js";
2
2
  import "./cpu/normRMS.js";
3
3
  import "./webgl/normRMS.js";
4
4
  import "./grads/normRMS.js";
@@ -1,6 +1,6 @@
1
1
  import "../utilities/packed.js";
2
- import "../index-ZyQhjEPo.js";
3
- import { p as t } from "../pack16-CFUqumar.js";
2
+ import "../index-D6Q1lPZO.js";
3
+ import { p as t } from "../pack16-CmVZs6af.js";
4
4
  export {
5
5
  t as pack16
6
6
  };
package/dist/ops/qkv.js CHANGED
@@ -1,4 +1,4 @@
1
- import { e as m } from "../index-ZyQhjEPo.js";
1
+ import { e as m } from "../index-D6Q1lPZO.js";
2
2
  import "./cpu/qkv.js";
3
3
  import "./webgl/qkv.js";
4
4
  import "./grads/qkv.js";
@@ -1,7 +1,7 @@
1
- import { e as c } from "../index-ZyQhjEPo.js";
1
+ import { e as c } from "../index-D6Q1lPZO.js";
2
2
  import { isPackedTensor as u, packTensor as i } from "../utilities/packed.js";
3
- import { r as p } from "../reshape-DevtBWtf.js";
4
- import { a as l, r as t } from "../tensor_util-DV-FP5Q3.js";
3
+ import { r as p } from "../reshape-CaPQzFvz.js";
4
+ import { a as l, r as t } from "../tensor_util-DfwaWayG.js";
5
5
  const m = {
6
6
  kernelName: "Reshape16",
7
7
  inputsToSave: ["x"],
package/dist/ops/rope.js CHANGED
@@ -1,11 +1,11 @@
1
- import "../index-ZyQhjEPo.js";
2
- import "../random_width-DY6Kk2Dl.js";
3
- import "../register_all_kernels-Bwu1PTuU.js";
1
+ import "../index-D6Q1lPZO.js";
2
+ import "../random_width-BVV9HveY.js";
3
+ import "../register_all_kernels-nvj2k7OC.js";
4
4
  import "../index-Cp39cXWe.js";
5
- import "../dataset-0xP8GjwI.js";
5
+ import "../dataset-D2P7rHAw.js";
6
6
  import "./cpu/rope.js";
7
7
  import "./webgl/rope.js";
8
- import { r as x } from "../rope-B5UUMsPi.js";
8
+ import { r as x } from "../rope-s4W2XO9B.js";
9
9
  import "../utilities/packed.js";
10
10
  export {
11
11
  x as rope
@@ -1,4 +1,4 @@
1
- import { e as i } from "../index-ZyQhjEPo.js";
1
+ import { e as i } from "../index-D6Q1lPZO.js";
2
2
  import "./cpu/scatterSub.js";
3
3
  import "./webgl/scatterSub.js";
4
4
  function c(t, r, e) {
@@ -1,6 +1,6 @@
1
1
  import { isPackedTensor as n } from "../utilities/packed.js";
2
- import { e as c } from "../index-ZyQhjEPo.js";
3
- import { s as i } from "../slice-D_gkkqZK.js";
2
+ import { e as c } from "../index-D6Q1lPZO.js";
3
+ import { s as i } from "../slice-DvovR5wq.js";
4
4
  function a(r, e, o) {
5
5
  return n(r) ? c().runKernel("Slice16", { x: r }, { begin: e, size: o }) : i(r, e, o);
6
6
  }
@@ -1,4 +1,4 @@
1
- import { e } from "../index-ZyQhjEPo.js";
1
+ import { e } from "../index-D6Q1lPZO.js";
2
2
  import "./grads/softmax16.js";
3
3
  import { isPackedTensor as m, packTensor as a } from "../utilities/packed.js";
4
4
  function p(r) {
package/dist/ops/sub16.js CHANGED
@@ -1,4 +1,4 @@
1
- import { c as s, e as t } from "../index-ZyQhjEPo.js";
1
+ import { c as s, e as t } from "../index-D6Q1lPZO.js";
2
2
  import { isPackedTensor as n } from "../utilities/packed.js";
3
3
  function c(r, e) {
4
4
  return !n(r) && !n(e) ? s(r, e) : t().runKernel("Sub16", { a: r, b: e });
package/dist/ops/sum16.js CHANGED
@@ -1,6 +1,6 @@
1
- import { e as t } from "../index-ZyQhjEPo.js";
1
+ import { e as t } from "../index-D6Q1lPZO.js";
2
2
  import { isPackedTensor as s } from "../utilities/packed.js";
3
- import { s as n } from "../sum-_fzj5ZTB.js";
3
+ import { s as n } from "../sum-ywRJj3Zr.js";
4
4
  function p(r, o, e = !1) {
5
5
  if (!s(r))
6
6
  return n(r, o, e);
@@ -1,9 +1,9 @@
1
- import { e as i } from "../index-ZyQhjEPo.js";
1
+ import { e as i } from "../index-D6Q1lPZO.js";
2
2
  import { forceInt as u, forceFloat as l } from "./grads/utils.js";
3
- import { g as m } from "../axis_util-BvHEw88j.js";
3
+ import { g as m } from "../axis_util-BsIr9ZNu.js";
4
4
  import { isPackedTensor as f, packTensor as g } from "../utilities/packed.js";
5
- import { t as a } from "../transpose-DKELTqhe.js";
6
- import { a as d, r as p } from "../tensor_util-DV-FP5Q3.js";
5
+ import { t as a } from "../transpose-DH4gmHvu.js";
6
+ import { a as d, r as p } from "../tensor_util-DfwaWayG.js";
7
7
  const k = {
8
8
  kernelName: "Transpose16",
9
9
  gradFunc: (e, s, o) => {
@@ -1,5 +1,5 @@
1
- import "../index-ZyQhjEPo.js";
2
- import { u as t } from "../pack16-CFUqumar.js";
1
+ import "../index-D6Q1lPZO.js";
2
+ import { u as t } from "../pack16-CmVZs6af.js";
3
3
  import "../utilities/packed.js";
4
4
  export {
5
5
  t as unpack16
@@ -1,6 +1,6 @@
1
- import { r as n } from "../../Reshape-_kILl6tK.js";
2
- import "../../index-ZyQhjEPo.js";
3
- import { r as f } from "../../tensor_util-DV-FP5Q3.js";
1
+ import { r as n } from "../../Reshape-CHdUjC72.js";
2
+ import "../../index-D6Q1lPZO.js";
3
+ import { r as f } from "../../tensor_util-DfwaWayG.js";
4
4
  class v {
5
5
  variableNames = ["moments", "value"];
6
6
  outputShape;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as m } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as m } from "../../tensor_util-DfwaWayG.js";
3
3
  class i {
4
4
  variableNames = ["moments", "gradient"];
5
5
  outputShape;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as p } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as p } from "../../tensor_util-DfwaWayG.js";
3
3
  class m {
4
4
  variableNames = ["cache", "item"];
5
5
  outputShape;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as d } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as d } from "../../tensor_util-DfwaWayG.js";
3
3
  class h {
4
4
  variableNames = ["q", "k"];
5
5
  outputShape;
@@ -1,9 +1,9 @@
1
- import { m as b, s as I, r as k } from "../../RealDiv-DgA3z9oO.js";
2
- import { r as v } from "../../Reshape-_kILl6tK.js";
3
- import "../../index-ZyQhjEPo.js";
4
- import { r as w } from "../../tensor_util-DV-FP5Q3.js";
5
- import { p as P } from "../../tensor-DdQUJZlz.js";
6
- import { e as S } from "../../axis_util-BvHEw88j.js";
1
+ import { m as b, s as I, r as k } from "../../RealDiv-zz7FpkKX.js";
2
+ import { r as v } from "../../Reshape-CHdUjC72.js";
3
+ import "../../index-D6Q1lPZO.js";
4
+ import { r as w } from "../../tensor_util-DfwaWayG.js";
5
+ import { p as P } from "../../tensor-CzmOBsdf.js";
6
+ import { e as S } from "../../axis_util-BsIr9ZNu.js";
7
7
  class T {
8
8
  variableNames = ["logits", "maxLogits"];
9
9
  outputShape;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as i } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as i } from "../../tensor_util-DfwaWayG.js";
3
3
  class l {
4
4
  variableNames = ["labels", "logits", "values"];
5
5
  outputShape;
@@ -1,6 +1,6 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { u as s, C as i } from "../../kernel_funcs_utils-Dg_-E44D.js";
3
- import { r as a } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { u as s, C as i } from "../../kernel_funcs_utils-CWfOAPGO.js";
3
+ import { r as a } from "../../tensor_util-DfwaWayG.js";
4
4
  const t = 0.7978845608028654, r = 0.044715, c = i + `
5
5
  float x3 = x * x * x;
6
6
  float inner = x + ${r} * x3;
@@ -1,7 +1,7 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { u as r, l as e } from "../../kernel_funcs_utils-Dg_-E44D.js";
3
- import { y as s } from "../../shared-BRksrJb3.js";
4
- import { r as N, L as l } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { u as r, l as e } from "../../kernel_funcs_utils-CWfOAPGO.js";
3
+ import { y as s } from "../../shared-CHhxz-O5.js";
4
+ import { r as N, L as l } from "../../tensor_util-DfwaWayG.js";
5
5
  const t = e + `
6
6
  return x < 0.0 ? NAN : log(x);
7
7
  `, a = `
@@ -1,11 +1,11 @@
1
1
  import { isPackedTensor as k } from "../../utilities/packed.js";
2
- import { m as M, b as m } from "../../index-ZyQhjEPo.js";
2
+ import { m as M, b as m } from "../../index-D6Q1lPZO.js";
3
3
  import { matMulMul as g } from "../matMulMul.js";
4
4
  import { matMulGelu as N } from "../matMulGelu.js";
5
- import { r as U } from "../../tensor_util-DV-FP5Q3.js";
6
- import { m as G } from "../../mat_mul-DeAh4uTH.js";
7
- import { r as w } from "../../reshape-DevtBWtf.js";
8
- import { t as h } from "../../transpose-DKELTqhe.js";
5
+ import { r as U } from "../../tensor_util-DfwaWayG.js";
6
+ import { m as G } from "../../mat_mul-C59XWcJd.js";
7
+ import { r as w } from "../../reshape-CaPQzFvz.js";
8
+ import { t as h } from "../../transpose-DH4gmHvu.js";
9
9
  function P(p) {
10
10
  const { A: r, B: o } = p.inputs, { transposeA: l, transposeB: c, scale: u, activation: A, scaleA: f, scaleB: d, forceOutputShape: t, perm: n } = p.attrs, B = !k(r), v = !k(o);
11
11
  if (B && v) {
@@ -1,9 +1,9 @@
1
- import { t as R, e as C, j as N } from "../../index-ZyQhjEPo.js";
2
- import { r as f } from "../../Reshape-_kILl6tK.js";
3
- import { M as H } from "../../mulmat_packed_gpu-BMFhLwta.js";
4
- import { r as E, u as O } from "../../tensor_util-DV-FP5Q3.js";
5
- import { m as $ } from "../../mat_mul-DeAh4uTH.js";
6
- import { s as A, a as U } from "../../tensor-DdQUJZlz.js";
1
+ import { t as R, e as C, j as N } from "../../index-D6Q1lPZO.js";
2
+ import { r as f } from "../../Reshape-CHdUjC72.js";
3
+ import { M as H } from "../../mulmat_packed_gpu-Coh6qbJk.js";
4
+ import { r as E, u as O } from "../../tensor_util-DfwaWayG.js";
5
+ import { m as $ } from "../../mat_mul-C59XWcJd.js";
6
+ import { s as A, a as U } from "../../tensor-CzmOBsdf.js";
7
7
  const M = 0.7978845608028654, g = 0.044715, j = `
8
8
  vec4 x3 = x * x * x;
9
9
  vec4 inner = x + ${g} * x3;
@@ -1,6 +1,6 @@
1
- import "../../index-ZyQhjEPo.js";
1
+ import "../../index-D6Q1lPZO.js";
2
2
  import { batchMatMulGeluImpl as u } from "./matMulGelu.js";
3
- import { r as c } from "../../tensor_util-DV-FP5Q3.js";
3
+ import { r as c } from "../../tensor_util-DfwaWayG.js";
4
4
  const p = `
5
5
  return a * b;
6
6
  `;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as m } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as m } from "../../tensor_util-DfwaWayG.js";
3
3
  class f {
4
4
  variableNames = ["a", "b"];
5
5
  outputShape;
@@ -1,6 +1,6 @@
1
- import { e as G } from "../../index-ZyQhjEPo.js";
2
- import { r as g } from "../../tensor_util-DV-FP5Q3.js";
3
- import { s as x } from "../../sum-_fzj5ZTB.js";
1
+ import { e as G } from "../../index-D6Q1lPZO.js";
2
+ import { r as g } from "../../tensor_util-DfwaWayG.js";
3
+ import { s as x } from "../../sum-ywRJj3Zr.js";
4
4
  class y {
5
5
  variableNames = ["x", "meanSquare", "gamma"];
6
6
  outputShape;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as i } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as i } from "../../tensor_util-DfwaWayG.js";
3
3
  class m {
4
4
  variableNames = ["x", "kernel"];
5
5
  outputShape;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as h } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as h } from "../../tensor_util-DfwaWayG.js";
3
3
  class g {
4
4
  variableNames = ["x", "sin", "cos"];
5
5
  outputShape;
@@ -1,5 +1,5 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { r as i } from "../../tensor_util-DV-FP5Q3.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { r as i } from "../../tensor_util-DfwaWayG.js";
3
3
  class u {
4
4
  variableNames = ["labels", "softmaxProbs", "dy"];
5
5
  outputShape;
@@ -1,8 +1,8 @@
1
- import { e as p } from "../../webgpu_program-Cigz-7RF.js";
2
- import { f as d, c as l } from "../../webgpu_util-BBCnKm2X.js";
3
- import "../../index-ZyQhjEPo.js";
4
- import { j as f } from "../../tensor-DdQUJZlz.js";
5
- import { r as c } from "../../tensor_util-DV-FP5Q3.js";
1
+ import { e as p } from "../../webgpu_program-DzaQiqel.js";
2
+ import { f as d, c as l } from "../../webgpu_util-0_ubCEHJ.js";
3
+ import "../../index-D6Q1lPZO.js";
4
+ import { j as f } from "../../tensor-CzmOBsdf.js";
5
+ import { r as c } from "../../tensor_util-DfwaWayG.js";
6
6
  class h {
7
7
  variableNames = ["moments", "value"];
8
8
  outputShape;
@@ -1,8 +1,8 @@
1
- import { e as p } from "../../webgpu_program-Cigz-7RF.js";
2
- import { f as u, c as d } from "../../webgpu_util-BBCnKm2X.js";
3
- import "../../index-ZyQhjEPo.js";
4
- import { j as f } from "../../tensor-DdQUJZlz.js";
5
- import { r as c } from "../../tensor_util-DV-FP5Q3.js";
1
+ import { e as p } from "../../webgpu_program-DzaQiqel.js";
2
+ import { f as u, c as d } from "../../webgpu_util-0_ubCEHJ.js";
3
+ import "../../index-D6Q1lPZO.js";
4
+ import { j as f } from "../../tensor-CzmOBsdf.js";
5
+ import { r as c } from "../../tensor_util-DfwaWayG.js";
6
6
  class l {
7
7
  variableNames = ["moments", "gradient"];
8
8
  outputShape;
@@ -1,7 +1,7 @@
1
- import "../../index-ZyQhjEPo.js";
1
+ import "../../index-D6Q1lPZO.js";
2
2
  import { BinaryOpProgram as p } from "./utils/binary_op.js";
3
3
  import { B as m } from "../../binary_op_util-pKXltfxI.js";
4
- import { r as c } from "../../tensor_util-DV-FP5Q3.js";
4
+ import { r as c } from "../../tensor_util-DfwaWayG.js";
5
5
  function i(r) {
6
6
  const { a: e, b: n } = r.inputs, a = r.backend, t = new p(m.ADD, e.shape, n.shape), o = a.runWebGPUProgram(t, [e, n], "int32");
7
7
  return o.packed = !0, o;
@@ -1,9 +1,9 @@
1
1
  import { isPackedTensor as S } from "../../utilities/packed.js";
2
- import { e as d } from "../../webgpu_program-Cigz-7RF.js";
3
- import { f as u, c as m } from "../../webgpu_util-BBCnKm2X.js";
4
- import "../../index-ZyQhjEPo.js";
5
- import { j as g } from "../../tensor-DdQUJZlz.js";
6
- import { r as x } from "../../tensor_util-DV-FP5Q3.js";
2
+ import { e as d } from "../../webgpu_program-DzaQiqel.js";
3
+ import { f as u, c as m } from "../../webgpu_util-0_ubCEHJ.js";
4
+ import "../../index-D6Q1lPZO.js";
5
+ import { j as g } from "../../tensor-CzmOBsdf.js";
6
+ import { r as x } from "../../tensor_util-DfwaWayG.js";
7
7
  class b {
8
8
  variableNames = ["cache", "item"];
9
9
  outputShape;
@@ -1,9 +1,9 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { j as d } from "../../tensor-DdQUJZlz.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { j as d } from "../../tensor-CzmOBsdf.js";
3
3
  import { isPackedTensor as p } from "../../utilities/packed.js";
4
- import { b } from "../../matMul16--R5hOwDG.js";
4
+ import { b } from "../../matMul16-fEAJ4smh.js";
5
5
  import l from "./attentionMask32_program.js";
6
- import { r as M } from "../../tensor_util-DV-FP5Q3.js";
6
+ import { r as M } from "../../tensor_util-DfwaWayG.js";
7
7
  function w(n) {
8
8
  const { q: t, k: e } = n.inputs, { divisor: a, pastLen: o } = n.attrs, m = n.backend;
9
9
  if (p(t) && p(e))
@@ -1,5 +1,5 @@
1
- import { e as r } from "../../webgpu_program-Cigz-7RF.js";
2
- import { f as a, c as u } from "../../webgpu_util-BBCnKm2X.js";
1
+ import { e as r } from "../../webgpu_program-DzaQiqel.js";
2
+ import { f as a, c as u } from "../../webgpu_util-0_ubCEHJ.js";
3
3
  class p {
4
4
  variableNames = ["q", "k"];
5
5
  outputShape;
@@ -1,10 +1,10 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { e as x } from "../../webgpu_program-Cigz-7RF.js";
3
- import { f as I, c as D } from "../../webgpu_util-BBCnKm2X.js";
4
- import { r as y } from "../../Reshape-CF6odzV4.js";
5
- import { r as $ } from "../../tensor_util-DV-FP5Q3.js";
6
- import { p as F, s as c } from "../../tensor-DdQUJZlz.js";
7
- import { a as L, c as d } from "../../concat_util-DcJk7YHS.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { e as x } from "../../webgpu_program-DzaQiqel.js";
3
+ import { f as I, c as D } from "../../webgpu_util-0_ubCEHJ.js";
4
+ import { r as y } from "../../Reshape-CDVLyVfz.js";
5
+ import { r as $ } from "../../tensor_util-DfwaWayG.js";
6
+ import { p as F, s as c } from "../../tensor-CzmOBsdf.js";
7
+ import { a as L, c as d } from "../../concat_util-iBYIyuQe.js";
8
8
  class T {
9
9
  outputShape;
10
10
  shaderKey;
@@ -1,8 +1,8 @@
1
- import { e as u } from "../../webgpu_program-Cigz-7RF.js";
2
- import { f as p, c as h } from "../../webgpu_util-BBCnKm2X.js";
3
- import "../../index-ZyQhjEPo.js";
4
- import { j as s } from "../../tensor-DdQUJZlz.js";
5
- import { r as c } from "../../tensor_util-DV-FP5Q3.js";
1
+ import { e as u } from "../../webgpu_program-DzaQiqel.js";
2
+ import { f as p, c as h } from "../../webgpu_util-0_ubCEHJ.js";
3
+ import "../../index-D6Q1lPZO.js";
4
+ import { j as s } from "../../tensor-CzmOBsdf.js";
5
+ import { r as c } from "../../tensor_util-DfwaWayG.js";
6
6
  class l {
7
7
  variableNames = ["labels", "logits", "values"];
8
8
  outputShape;
@@ -1,8 +1,8 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { e as s } from "../../webgpu_program-Cigz-7RF.js";
3
- import { f as o, c as p } from "../../webgpu_util-BBCnKm2X.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { e as s } from "../../webgpu_program-DzaQiqel.js";
3
+ import { f as o, c as p } from "../../webgpu_util-0_ubCEHJ.js";
4
4
  import { isPackedTensor as l } from "../../utilities/packed.js";
5
- import { r as h } from "../../tensor_util-DV-FP5Q3.js";
5
+ import { r as h } from "../../tensor_util-DfwaWayG.js";
6
6
  const r = 0.7978845608028654, u = 0.044715;
7
7
  class x {
8
8
  outputShape;
@@ -1,14 +1,14 @@
1
- import { m as y, b as B, j as Q } from "../../index-ZyQhjEPo.js";
1
+ import { m as y, b as B, j as Q } from "../../index-D6Q1lPZO.js";
2
2
  import { isPackedTensor as R } from "../../utilities/packed.js";
3
3
  import { reshape16 as U } from "../reshape16.js";
4
4
  import { matMulMul as V } from "../matMulMul.js";
5
5
  import { matMulGelu as X } from "../matMulGelu.js";
6
6
  import Y from "./matMul16_program.js";
7
- import { r as Z } from "../../tensor_util-DV-FP5Q3.js";
8
- import { m as _ } from "../../mat_mul-DeAh4uTH.js";
9
- import { r as x } from "../../reshape-DevtBWtf.js";
10
- import { t as C } from "../../transpose-DKELTqhe.js";
11
- import { s as E } from "../../tensor-DdQUJZlz.js";
7
+ import { r as Z } from "../../tensor_util-DfwaWayG.js";
8
+ import { m as _ } from "../../mat_mul-C59XWcJd.js";
9
+ import { r as x } from "../../reshape-CaPQzFvz.js";
10
+ import { t as C } from "../../transpose-DH4gmHvu.js";
11
+ import { s as E } from "../../tensor-CzmOBsdf.js";
12
12
  function $(p) {
13
13
  const { A: e, B: s } = p.inputs, { transposeA: d, transposeB: f, scale: i, activation: k, scaleA: c, scaleB: u, forceOutputShape: o, perm: m, causalMask: g, pastLen: W } = p.attrs, z = p.backend, S = !R(e), M = !R(s);
14
14
  if (S && M) {
@@ -1,6 +1,6 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { e as h } from "../../webgpu_program-Cigz-7RF.js";
3
- import { s as f } from "../../tensor-DdQUJZlz.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { e as h } from "../../webgpu_program-DzaQiqel.js";
3
+ import { s as f } from "../../tensor-CzmOBsdf.js";
4
4
  class A {
5
5
  variableNames = ["A", "B"];
6
6
  outputShape;
@@ -1,7 +1,7 @@
1
- import "../../index-ZyQhjEPo.js";
1
+ import "../../index-D6Q1lPZO.js";
2
2
  import { BinaryOpProgram as m } from "./utils/binary_op.js";
3
3
  import { B as p } from "../../binary_op_util-pKXltfxI.js";
4
- import { r as c } from "../../tensor_util-DV-FP5Q3.js";
4
+ import { r as c } from "../../tensor_util-DfwaWayG.js";
5
5
  function i(r) {
6
6
  const { a: e, b: n } = r.inputs, t = r.backend, a = new m(p.MUL, e.shape, n.shape), o = t.runWebGPUProgram(a, [e, n], "int32");
7
7
  return o.packed = !0, o;
@@ -1,12 +1,12 @@
1
- import "../../index-ZyQhjEPo.js";
1
+ import "../../index-D6Q1lPZO.js";
2
2
  import { createReduceInfo as g, reduce as l } from "./utils/reductions.js";
3
- import { j as w } from "../../tensor-DdQUJZlz.js";
3
+ import { j as w } from "../../tensor-CzmOBsdf.js";
4
4
  import { isPackedTensor as f } from "../../utilities/packed.js";
5
- import { p as k } from "../../pack16-CFUqumar.js";
5
+ import { p as k } from "../../pack16-CmVZs6af.js";
6
6
  import S from "./normRMS16_program.js";
7
7
  import z from "./normRMS32_program.js";
8
8
  import N from "./utils/deviceInfo.js";
9
- import { r as b } from "../../tensor_util-DV-FP5Q3.js";
9
+ import { r as b } from "../../tensor_util-DfwaWayG.js";
10
10
  function P(m) {
11
11
  const { x: e, gamma: n } = m.inputs, c = m.backend, i = N(c), s = f(e), a = f(n), o = s || a, r = !o || s ? e : k(e), p = !o || a ? n : k(n), h = [r, p], t = g(h, -1), u = o ? new S(i, t) : new z(i, t);
12
12
  if (w(p.shape, [r.shape[r.shape.length - 1]], "Error in RMSNorm: "), e.shape.length !== 3)
@@ -1,14 +1,14 @@
1
- import { e as _ } from "../../index-ZyQhjEPo.js";
1
+ import { e as _ } from "../../index-D6Q1lPZO.js";
2
2
  import { createReduceInfo as D } from "./utils/reductions.js";
3
- import { f as X } from "../../webgpu_util-BBCnKm2X.js";
4
- import { e as $ } from "../../webgpu_program-Cigz-7RF.js";
5
- import { j as z } from "../../tensor-DdQUJZlz.js";
6
- import { p as k, u as M } from "../../pack16-CFUqumar.js";
3
+ import { f as X } from "../../webgpu_util-0_ubCEHJ.js";
4
+ import { e as $ } from "../../webgpu_program-DzaQiqel.js";
5
+ import { j as z } from "../../tensor-CzmOBsdf.js";
6
+ import { p as k, u as M } from "../../pack16-CmVZs6af.js";
7
7
  import { isPackedTensor as h } from "../../utilities/packed.js";
8
8
  import { reshape16 as R } from "../reshape16.js";
9
9
  import { sum16 as L } from "../sum16.js";
10
10
  import { slice16 as w } from "../slice16.js";
11
- import { r as P } from "../../tensor_util-DV-FP5Q3.js";
11
+ import { r as P } from "../../tensor_util-DfwaWayG.js";
12
12
  class N {
13
13
  outputShape;
14
14
  shaderKey = "RMSNormGrad";
@@ -1,6 +1,6 @@
1
- import "../../index-ZyQhjEPo.js";
1
+ import "../../index-D6Q1lPZO.js";
2
2
  import c from "./pack16_program.js";
3
- import { r as p } from "../../tensor_util-DV-FP5Q3.js";
3
+ import { r as p } from "../../tensor_util-DfwaWayG.js";
4
4
  function m(n) {
5
5
  const { x: e } = n.inputs, { scaling: t, padding: r } = n.attrs, i = n.backend;
6
6
  if (e.shape[e.shape.length - 1] % 2 !== 0)
@@ -1,5 +1,5 @@
1
- import { f as o, c as a } from "../../webgpu_util-BBCnKm2X.js";
2
- import { e as s } from "../../webgpu_program-Cigz-7RF.js";
1
+ import { f as o, c as a } from "../../webgpu_util-0_ubCEHJ.js";
2
+ import { e as s } from "../../webgpu_program-DzaQiqel.js";
3
3
  class h {
4
4
  outputShape;
5
5
  shaderKey = "Pack16";
@@ -1,9 +1,9 @@
1
- import "../../index-ZyQhjEPo.js";
2
- import { j as h } from "../../tensor-DdQUJZlz.js";
3
- import { b as f } from "../../matMul16--R5hOwDG.js";
1
+ import "../../index-D6Q1lPZO.js";
2
+ import { j as h } from "../../tensor-CzmOBsdf.js";
3
+ import { b as f } from "../../matMul16-fEAJ4smh.js";
4
4
  import { slice16 as a } from "../slice16.js";
5
5
  import { isPackedTensor as l } from "../../utilities/packed.js";
6
- import { r as u } from "../../tensor_util-DV-FP5Q3.js";
6
+ import { r as u } from "../../tensor_util-DfwaWayG.js";
7
7
  function k(i) {
8
8
  const { x: r, kernel: c } = i.inputs, { heads: e } = i.attrs, t = r.shape[0], n = r.shape[1], s = r.shape[2], m = l(r);
9
9
  if (h(c.shape, [m ? s * 2 : s, 3 * s], "Error in QKV: "), s % e !== 0)