@genai-fi/nanogpt 0.9.1 → 0.10.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (343) hide show
  1. package/README.md +352 -14
  2. package/dist/Generator.js +69 -78
  3. package/dist/{RealDiv-D4EzDsC0.js → RealDiv-DgA3z9oO.js} +32 -206
  4. package/dist/Reshape-CF6odzV4.js +16 -0
  5. package/dist/Reshape-_kILl6tK.js +81 -0
  6. package/dist/TeachableLLM.js +28 -22
  7. package/dist/Trainer.d.ts +2 -0
  8. package/dist/Trainer.js +3 -2
  9. package/dist/{axis_util-TbGYJ208.js → axis_util-BvHEw88j.js} +7 -23
  10. package/dist/backend.d.ts +2 -1
  11. package/dist/backend.js +10 -4
  12. package/dist/backend_util-D-rUb2ty.js +474 -0
  13. package/dist/backend_webgpu-B0u2ndUn.js +547 -0
  14. package/dist/binary_op_util-pKXltfxI.js +192 -0
  15. package/dist/broadcast_to-CwF7XIeu.js +30 -0
  16. package/dist/checks/appendCache.js +2 -2
  17. package/dist/checks/attentionMask.js +3 -3
  18. package/dist/checks/check.d.ts +1 -1
  19. package/dist/checks/check.js +8 -8
  20. package/dist/checks/gelu.js +2 -2
  21. package/dist/checks/index.d.ts +2 -0
  22. package/dist/checks/index.js +7 -5
  23. package/dist/checks/matMulGelu.js +6 -6
  24. package/dist/checks/normRMS.js +7 -7
  25. package/dist/checks/normRMSGrad.js +3 -3
  26. package/dist/checks/packUnpack.d.ts +1 -0
  27. package/dist/checks/packUnpack.js +18 -0
  28. package/dist/checks/qkv.js +12 -27
  29. package/dist/checks/rope.js +2 -2
  30. package/dist/checks/weights.js +18 -16
  31. package/dist/complex-CSlYz-2T.js +13 -0
  32. package/dist/complex_util-Yc1A_gV1.js +55 -0
  33. package/dist/concat-BHlIJeyT.js +19 -0
  34. package/dist/concat_util-DcJk7YHS.js +22 -0
  35. package/dist/data/docx.js +1 -1
  36. package/dist/data/parquet.js +2 -2
  37. package/dist/data/pdf.js +1 -1
  38. package/dist/data/textLoader.js +1 -1
  39. package/dist/{dataset-DlZtKmBq.js → dataset-0xP8GjwI.js} +136 -236
  40. package/dist/dropout-C1pM3f11.js +99 -0
  41. package/dist/expand_dims-BPG4fwBP.js +13 -0
  42. package/dist/exports_initializers-xuidcwI4.js +7 -0
  43. package/dist/gather-DykLGqmW.js +10 -0
  44. package/dist/{gelu-Bp_-935b.js → gelu-CNLFZWea.js} +11 -10
  45. package/dist/{gpgpu_math-CDaYiyE_.js → gpgpu_math-DDVJCn6-.js} +90 -265
  46. package/dist/{index-C4L8Cm77.js → index-CieiGp4Y.js} +14 -14
  47. package/dist/index-CjOj7j-u.js +7308 -0
  48. package/dist/{index-Tf7vU29b.js → index-Cp39cXWe.js} +3 -10
  49. package/dist/{index-Dwqa6Zy2.js → index-DvYrXKkX.js} +2 -2
  50. package/dist/index-ZyQhjEPo.js +2157 -0
  51. package/dist/{jszip.min-CjP2V1VV.js → jszip.min-Bz5-11Bk.js} +56 -57
  52. package/dist/kernel_funcs_utils-Dg_-E44D.js +308 -0
  53. package/dist/layers/BaseLayer.d.ts +1 -0
  54. package/dist/layers/BaseLayer.js +7 -6
  55. package/dist/layers/CausalSelfAttention.d.ts +0 -1
  56. package/dist/layers/CausalSelfAttention.js +56 -55
  57. package/dist/layers/MLP.js +15 -16
  58. package/dist/layers/PositionEmbedding.js +5 -14
  59. package/dist/layers/RMSNorm.js +3 -3
  60. package/dist/layers/RoPECache.d.ts +2 -0
  61. package/dist/layers/RoPECache.js +22 -17
  62. package/dist/layers/TiedEmbedding.js +22 -17
  63. package/dist/layers/TransformerBlock.js +21 -20
  64. package/dist/loader/load.js +1 -1
  65. package/dist/loader/loadTransformers.js +1 -1
  66. package/dist/loader/oldZipLoad.js +39 -33
  67. package/dist/loader/save.js +1 -1
  68. package/dist/log_sum_exp-DWI-76TI.js +41 -0
  69. package/dist/main.d.ts +8 -0
  70. package/dist/main.js +63 -52
  71. package/dist/matMul16--R5hOwDG.js +77 -0
  72. package/dist/mat_mul-DeAh4uTH.js +12 -0
  73. package/dist/mod-Gt1rMB4n.js +12 -0
  74. package/dist/models/NanoGPTV1.js +40 -31
  75. package/dist/models/model.d.ts +2 -0
  76. package/dist/models/model.js +37 -29
  77. package/dist/{mulmat_packed_gpu-BT60jmzP.js → mulmat_packed_gpu-BMFhLwta.js} +1 -17
  78. package/dist/{non_max_suppression_impl-CsEgBuMA.js → non_max_suppression_impl-B2W7YjZB.js} +0 -32
  79. package/dist/ones-CAMiP4I2.js +15 -0
  80. package/dist/ops/adamAdjust.js +1 -1
  81. package/dist/ops/adamMoments.d.ts +1 -1
  82. package/dist/ops/adamMoments.js +4 -4
  83. package/dist/ops/add16.d.ts +2 -0
  84. package/dist/ops/add16.js +9 -0
  85. package/dist/ops/appendCache.js +16 -9
  86. package/dist/ops/attentionMask.js +4 -4
  87. package/dist/ops/concat16.d.ts +2 -0
  88. package/dist/ops/concat16.js +9 -0
  89. package/dist/ops/cpu/adamAdjust.js +14 -13
  90. package/dist/ops/cpu/adamMoments.js +10 -9
  91. package/dist/ops/cpu/appendCache.js +9 -8
  92. package/dist/ops/cpu/attentionMask.js +15 -14
  93. package/dist/ops/cpu/fusedSoftmax.js +13 -12
  94. package/dist/ops/cpu/gatherSub.js +9 -24
  95. package/dist/ops/cpu/gelu.js +13 -12
  96. package/dist/ops/cpu/matMul16.d.ts +1 -0
  97. package/dist/ops/cpu/matMul16.js +16 -0
  98. package/dist/ops/cpu/matMulGelu.js +18 -16
  99. package/dist/ops/cpu/matMulMul.js +8 -7
  100. package/dist/ops/cpu/mulDropout.js +4 -3
  101. package/dist/ops/cpu/normRMS.js +11 -10
  102. package/dist/ops/cpu/qkv.js +17 -13
  103. package/dist/ops/cpu/rope.js +23 -22
  104. package/dist/ops/cpu/scatterSub.js +16 -30
  105. package/dist/ops/dot16.d.ts +2 -0
  106. package/dist/ops/dot16.js +42 -0
  107. package/dist/ops/gatherSub.js +1 -1
  108. package/dist/ops/gelu.js +2 -2
  109. package/dist/ops/grads/add16.d.ts +1 -0
  110. package/dist/ops/grads/add16.js +27 -0
  111. package/dist/ops/grads/attentionMask.js +12 -19
  112. package/dist/ops/grads/gelu.js +4 -3
  113. package/dist/ops/grads/matMul16.d.ts +2 -0
  114. package/dist/ops/grads/matMul16.js +9 -0
  115. package/dist/ops/grads/matMulGelu.js +8 -7
  116. package/dist/ops/grads/normRMS.js +8 -7
  117. package/dist/ops/grads/{fusedSoftmax.d.ts → pack16.d.ts} +1 -1
  118. package/dist/ops/grads/pack16.js +7 -0
  119. package/dist/ops/grads/qkv.d.ts +3 -1
  120. package/dist/ops/grads/qkv.js +28 -22
  121. package/dist/ops/grads/rope.d.ts +2 -1
  122. package/dist/ops/grads/rope.js +6 -13
  123. package/dist/ops/grads/softmax16.d.ts +2 -0
  124. package/dist/ops/grads/softmax16.js +26 -0
  125. package/dist/ops/grads/unpack16.d.ts +2 -0
  126. package/dist/ops/grads/unpack16.js +6 -0
  127. package/dist/ops/grads/utils.d.ts +3 -0
  128. package/dist/ops/grads/utils.js +10 -0
  129. package/dist/ops/matMul16.d.ts +15 -0
  130. package/dist/ops/matMul16.js +13 -0
  131. package/dist/ops/matMulGelu.js +1 -1
  132. package/dist/ops/matMulMul.js +1 -1
  133. package/dist/ops/mul16.d.ts +2 -0
  134. package/dist/ops/mul16.js +8 -0
  135. package/dist/ops/mulDrop.js +1 -1
  136. package/dist/ops/normRMS.js +1 -1
  137. package/dist/ops/pack16.d.ts +2 -0
  138. package/dist/ops/pack16.js +6 -0
  139. package/dist/ops/qkv.d.ts +1 -1
  140. package/dist/ops/qkv.js +8 -4
  141. package/dist/ops/reshape16.d.ts +2 -0
  142. package/dist/ops/reshape16.js +43 -0
  143. package/dist/ops/rope.d.ts +1 -1
  144. package/dist/ops/rope.js +8 -10
  145. package/dist/ops/scatterSub.js +1 -1
  146. package/dist/ops/slice16.d.ts +2 -0
  147. package/dist/ops/slice16.js +9 -0
  148. package/dist/ops/softmax16.d.ts +2 -0
  149. package/dist/ops/softmax16.js +12 -0
  150. package/dist/ops/sub16.d.ts +2 -0
  151. package/dist/ops/sub16.js +8 -0
  152. package/dist/ops/sum16.d.ts +2 -0
  153. package/dist/ops/sum16.js +13 -0
  154. package/dist/ops/transpose16.d.ts +3 -0
  155. package/dist/ops/transpose16.js +41 -0
  156. package/dist/ops/unpack16.d.ts +2 -0
  157. package/dist/ops/unpack16.js +6 -0
  158. package/dist/ops/webgl/adamAdjust.js +3 -2
  159. package/dist/ops/webgl/adamMoments.js +2 -1
  160. package/dist/ops/webgl/appendCache.js +2 -1
  161. package/dist/ops/webgl/attentionMask.js +5 -4
  162. package/dist/ops/webgl/fusedSoftmax.js +6 -4
  163. package/dist/ops/webgl/gatherSub.js +7 -6
  164. package/dist/ops/webgl/gelu.js +3 -2
  165. package/dist/ops/webgl/log.js +12 -27
  166. package/dist/ops/webgl/matMul16.d.ts +1 -0
  167. package/dist/ops/webgl/matMul16.js +37 -0
  168. package/dist/ops/webgl/matMulGelu.js +17 -15
  169. package/dist/ops/webgl/matMulMul.js +13 -12
  170. package/dist/ops/webgl/mulDropout.js +9 -8
  171. package/dist/ops/webgl/normRMS.js +8 -7
  172. package/dist/ops/webgl/qkv.js +6 -5
  173. package/dist/ops/webgl/rope.js +11 -10
  174. package/dist/ops/webgl/scatterSub.js +6 -5
  175. package/dist/ops/webgpu/adamAdjust.js +12 -10
  176. package/dist/ops/webgpu/adamMoments.js +27 -22
  177. package/dist/ops/webgpu/add16.d.ts +1 -0
  178. package/dist/ops/webgpu/add16.js +14 -0
  179. package/dist/ops/webgpu/appendCache.js +64 -17
  180. package/dist/ops/webgpu/attentionMask.js +19 -62
  181. package/dist/ops/webgpu/attentionMask32_program.d.ts +19 -0
  182. package/dist/ops/webgpu/attentionMask32_program.js +54 -0
  183. package/dist/ops/webgpu/concat16.d.ts +19 -0
  184. package/dist/ops/webgpu/concat16.js +128 -0
  185. package/dist/ops/webgpu/gatherSub.js +9 -7
  186. package/dist/ops/webgpu/gelu.js +78 -31
  187. package/dist/ops/webgpu/index.js +12 -0
  188. package/dist/ops/webgpu/matMul16.d.ts +1 -0
  189. package/dist/ops/webgpu/matMul16.js +58 -0
  190. package/dist/ops/webgpu/matMul16_program.d.ts +42 -0
  191. package/dist/ops/webgpu/matMul16_program.js +336 -0
  192. package/dist/ops/webgpu/mul16.d.ts +1 -0
  193. package/dist/ops/webgpu/mul16.js +14 -0
  194. package/dist/ops/webgpu/normRMS.js +21 -40
  195. package/dist/ops/webgpu/normRMS16_program.d.ts +9 -0
  196. package/dist/ops/webgpu/normRMS16_program.js +24 -0
  197. package/dist/ops/webgpu/normRMS32_program.d.ts +9 -0
  198. package/dist/ops/webgpu/normRMS32_program.js +24 -0
  199. package/dist/ops/webgpu/normRMSGrad.js +113 -64
  200. package/dist/ops/webgpu/pack16.d.ts +1 -0
  201. package/dist/ops/webgpu/pack16.js +19 -0
  202. package/dist/ops/webgpu/pack16_program.d.ts +19 -0
  203. package/dist/ops/webgpu/pack16_program.js +92 -0
  204. package/dist/ops/webgpu/qkv.js +20 -55
  205. package/dist/ops/webgpu/rope.js +77 -22
  206. package/dist/ops/webgpu/scatterSub.js +9 -7
  207. package/dist/ops/webgpu/slice16.d.ts +7 -0
  208. package/dist/ops/webgpu/slice16.js +71 -0
  209. package/dist/{variable-Bm2OFwGI.js → ops/webgpu/softmax16.d.ts} +2 -8
  210. package/dist/ops/webgpu/softmax16.js +23 -0
  211. package/dist/ops/webgpu/softmax16_program.d.ts +13 -0
  212. package/dist/ops/webgpu/softmax16_program.js +73 -0
  213. package/dist/ops/webgpu/softmax16_subgroup_program.d.ts +17 -0
  214. package/dist/ops/webgpu/softmax16_subgroup_program.js +75 -0
  215. package/dist/ops/webgpu/softmax16grad.d.ts +1 -0
  216. package/dist/ops/webgpu/softmax16grad.js +38 -0
  217. package/dist/ops/webgpu/sub16.d.ts +1 -0
  218. package/dist/ops/webgpu/sub16.js +14 -0
  219. package/dist/ops/webgpu/sum16.d.ts +1 -0
  220. package/dist/ops/webgpu/sum16.js +40 -0
  221. package/dist/ops/webgpu/transpose16.d.ts +1 -0
  222. package/dist/ops/webgpu/transpose16.js +35 -0
  223. package/dist/ops/webgpu/transpose16_program.d.ts +16 -0
  224. package/dist/ops/webgpu/transpose16_program.js +50 -0
  225. package/dist/ops/webgpu/transpose16_shared_program.d.ts +15 -0
  226. package/dist/ops/webgpu/transpose16_shared_program.js +71 -0
  227. package/dist/ops/webgpu/unpack16.d.ts +1 -0
  228. package/dist/ops/webgpu/unpack16.js +49 -0
  229. package/dist/ops/webgpu/utils/binary_op.d.ts +19 -0
  230. package/dist/ops/webgpu/utils/binary_op.js +79 -0
  231. package/dist/ops/webgpu/utils/deviceInfo.d.ts +7 -0
  232. package/dist/ops/webgpu/utils/deviceInfo.js +11 -0
  233. package/dist/ops/webgpu/utils/reductions.d.ts +32 -4
  234. package/dist/ops/webgpu/utils/reductions.js +236 -45
  235. package/dist/ops-CNI3TwqM.js +645 -0
  236. package/dist/pack16-CFUqumar.js +41 -0
  237. package/dist/{papaparse.min-C8l2Kvo1.js → papaparse.min-C0cScC2i.js} +2 -8
  238. package/dist/{parquet-C0Tlmv9c.js → parquet-BE8MU_ge.js} +201 -278
  239. package/dist/patches/PackedTensor.d.ts +12 -0
  240. package/dist/patches/PackedTensor.js +11 -0
  241. package/dist/patches/engine.d.ts +261 -0
  242. package/dist/patches/engine.js +10 -0
  243. package/dist/patches/tape.d.ts +12 -0
  244. package/dist/patches/tape.js +5 -0
  245. package/dist/patches/webgpu_backend.d.ts +18 -0
  246. package/dist/patches/webgpu_backend.js +57 -0
  247. package/dist/{tensor-CZr4dh61.js → patches/webgpu_base.d.ts} +5 -8
  248. package/dist/patches/webgpu_base.js +34 -0
  249. package/dist/patches/webgpu_program.d.ts +36 -0
  250. package/dist/patches/webgpu_program.js +401 -0
  251. package/dist/{pdf-kJD-f258.js → pdf-NIhmP3sq.js} +424 -428
  252. package/dist/random_width-DY6Kk2Dl.js +10051 -0
  253. package/dist/range-BMS52eQi.js +11 -0
  254. package/dist/reciprocal-CTmshQ9J.js +10 -0
  255. package/dist/{register_all_kernels-DIGpEwcf.js → register_all_kernels-Bwu1PTuU.js} +719 -9766
  256. package/dist/relu-yZ2-7WxU.js +10 -0
  257. package/dist/reshape-DevtBWtf.js +10 -0
  258. package/dist/rope-B5UUMsPi.js +32 -0
  259. package/dist/{scatter_nd_util-BQdz--Gn.js → scatter_nd_util-5EL-8VAQ.js} +1 -1
  260. package/dist/selu_util-D1w6yyTO.js +303 -0
  261. package/dist/{shared-DuP7ue-R.js → shared-BRksrJb3.js} +1 -17
  262. package/dist/shared-BuAXb4CI.js +2145 -0
  263. package/dist/sin-BGfy2HZo.js +16 -0
  264. package/dist/slice-D_gkkqZK.js +13 -0
  265. package/dist/slice_util-DtEldBfK.js +261 -0
  266. package/dist/softmax-ZHVebtR1.js +13 -0
  267. package/dist/split-DrfihRpZ.js +10 -0
  268. package/dist/squeeze-DZEpeblb.js +11 -0
  269. package/dist/stack-yOIAalTq.js +13 -0
  270. package/dist/sum-_fzj5ZTB.js +12 -0
  271. package/dist/tensor-DdQUJZlz.js +909 -0
  272. package/dist/tensor-f35l8Odg.js +8 -0
  273. package/dist/tensor1d-CeZuc-Rv.js +12 -0
  274. package/dist/tensor2d-G4Ys2GxX.js +15 -0
  275. package/dist/tensor4d-B8roDgtc.js +15 -0
  276. package/dist/tensor_util-DV-FP5Q3.js +523 -0
  277. package/dist/tfjs_backend-kNyO5L2d.js +653 -0
  278. package/dist/tile-BzyEiF-F.js +13 -0
  279. package/dist/tokeniser/CharTokeniser.js +1 -1
  280. package/dist/tokeniser/bpe.js +1 -1
  281. package/dist/training/Adam.d.ts +2 -1
  282. package/dist/training/Adam.js +12 -28
  283. package/dist/training/AdamExt.d.ts +1 -0
  284. package/dist/training/AdamExt.js +2 -2
  285. package/dist/training/DatasetBuilder.js +3 -20
  286. package/dist/training/FullTrainer.js +55 -48
  287. package/dist/training/Trainer.d.ts +11 -6
  288. package/dist/training/Trainer.js +51 -39
  289. package/dist/training/sparseCrossEntropy.js +3 -3
  290. package/dist/transpose-DKELTqhe.js +38 -0
  291. package/dist/utilities/arrayClose.js +7 -7
  292. package/dist/utilities/dummy.js +35 -27
  293. package/dist/utilities/multinomialCPU.js +2 -2
  294. package/dist/utilities/packed.d.ts +7 -0
  295. package/dist/utilities/packed.js +716 -0
  296. package/dist/utilities/performance.js +1 -1
  297. package/dist/utilities/profile.js +1 -1
  298. package/dist/utilities/safetensors.js +2 -2
  299. package/dist/utilities/sentences.d.ts +5 -0
  300. package/dist/utilities/sentences.js +41 -0
  301. package/dist/utilities/weights.js +2 -2
  302. package/dist/variable-Bhn5bHYv.js +7 -0
  303. package/dist/{webgpu_program-DkQJOJSd.js → webgpu_program-Cigz-7RF.js} +15 -44
  304. package/dist/webgpu_util-BBCnKm2X.js +65 -0
  305. package/dist/zeros-2gldETuK.js +14 -0
  306. package/package.json +4 -3
  307. package/dist/Reshape-Bowtk9BP.js +0 -127
  308. package/dist/Reshape-DUqYftGC.js +0 -30
  309. package/dist/backend_util-CJIiDoV1.js +0 -749
  310. package/dist/broadcast_to-DzlNweb8.js +0 -44
  311. package/dist/concat-B912vBbo.js +0 -33
  312. package/dist/dropout-C-csYCLj.js +0 -193
  313. package/dist/exports_initializers-B8iZMgQ0.js +0 -16
  314. package/dist/gather-Dnpgw-YQ.js +0 -25
  315. package/dist/index-BzFyqcy-.js +0 -4457
  316. package/dist/index-C1rx_Ajs.js +0 -12076
  317. package/dist/kernel_funcs_utils-DKLK0Mg3.js +0 -466
  318. package/dist/log_sum_exp-DO6z8tSE.js +0 -103
  319. package/dist/mat_mul-DzjTFx-u.js +0 -27
  320. package/dist/mod-Dobti4j4.js +0 -27
  321. package/dist/ones-tIJeHlq-.js +0 -29
  322. package/dist/ops/fusedSoftmax.d.ts +0 -2
  323. package/dist/ops/fusedSoftmax.js +0 -10
  324. package/dist/ops/grads/fusedSoftmax.js +0 -22
  325. package/dist/ops-LuCMAnmM.js +0 -1525
  326. package/dist/random_width-CXVRloNK.js +0 -13670
  327. package/dist/range-CWcz7xFA.js +0 -26
  328. package/dist/reciprocal-C4rNcM-S.js +0 -25
  329. package/dist/relu-BjCh_SYb.js +0 -25
  330. package/dist/reshape-CnIwVG1c.js +0 -25
  331. package/dist/selu_util-OtRzVwW5.js +0 -719
  332. package/dist/shared-DmRsFyaJ.js +0 -3134
  333. package/dist/sin-gpDNRxE0.js +0 -47
  334. package/dist/slice-d0Vo9XTN.js +0 -28
  335. package/dist/softmax-D7Jj3p_P.js +0 -28
  336. package/dist/split-DK2k5eHf.js +0 -25
  337. package/dist/stack-DFatutCx.js +0 -27
  338. package/dist/sum-CJ0ULhmt.js +0 -27
  339. package/dist/tensor1d-vML0r3q6.js +0 -27
  340. package/dist/tensor2d-D76QGjF3.js +0 -30
  341. package/dist/tensor4d-Df1WlVDY.js +0 -30
  342. package/dist/webgpu_util-pLEV9tks.js +0 -80
  343. package/dist/zeros-Bj5rMYA7.js +0 -52
@@ -1,25 +1,10 @@
1
- import { f as e, aj as r } from "../../index-BzFyqcy-.js";
2
- import { u as s, l as N } from "../../kernel_funcs_utils-DKLK0Mg3.js";
3
- import { y as l } from "../../shared-DuP7ue-R.js";
4
- /**
5
- * @license
6
- * Copyright 2020 Google LLC. All Rights Reserved.
7
- * Licensed under the Apache License, Version 2.0 (the "License");
8
- * you may not use this file except in compliance with the License.
9
- * You may obtain a copy of the License at
10
- *
11
- * http://www.apache.org/licenses/LICENSE-2.0
12
- *
13
- * Unless required by applicable law or agreed to in writing, software
14
- * distributed under the License is distributed on an "AS IS" BASIS,
15
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
16
- * See the License for the specific language governing permissions and
17
- * limitations under the License.
18
- * =============================================================================
19
- */
20
- const a = N + `
1
+ import "../../index-ZyQhjEPo.js";
2
+ import { u as r, l as e } from "../../kernel_funcs_utils-Dg_-E44D.js";
3
+ import { y as s } from "../../shared-BRksrJb3.js";
4
+ import { r as N, L as l } from "../../tensor_util-DV-FP5Q3.js";
5
+ const t = e + `
21
6
  return x < 0.0 ? NAN : log(x);
22
- `, t = `
7
+ `, a = `
23
8
  vec4 result = log(x);
24
9
  bvec4 isNaN = isnan(x);
25
10
  result.r = isNaN.r ? x.r : (x.r < 0.0 ? NAN : result.r);
@@ -27,13 +12,13 @@ const a = N + `
27
12
  result.b = isNaN.b ? x.b : (x.b < 0.0 ? NAN : result.b);
28
13
  result.a = isNaN.a ? x.a : (x.a < 0.0 ? NAN : result.a);
29
14
  return result;
30
- `, n = s({
31
- opSnippet: a,
32
- packedOpSnippet: t,
33
- cpuKernelImpl: l
15
+ `, n = r({
16
+ opSnippet: t,
17
+ packedOpSnippet: a,
18
+ cpuKernelImpl: s
34
19
  }), o = {
35
- kernelName: r,
20
+ kernelName: l,
36
21
  backendName: "webgl",
37
22
  kernelFunc: n
38
23
  };
39
- e(o);
24
+ N(o);
@@ -0,0 +1 @@
1
+ export {};
@@ -0,0 +1,37 @@
1
+ import { isPackedTensor as k } from "../../utilities/packed.js";
2
+ import { m as M, b as m } from "../../index-ZyQhjEPo.js";
3
+ import { matMulMul as g } from "../matMulMul.js";
4
+ import { matMulGelu as N } from "../matMulGelu.js";
5
+ import { r as U } from "../../tensor_util-DV-FP5Q3.js";
6
+ import { m as G } from "../../mat_mul-DeAh4uTH.js";
7
+ import { r as w } from "../../reshape-DevtBWtf.js";
8
+ import { t as h } from "../../transpose-DKELTqhe.js";
9
+ function P(p) {
10
+ const { A: r, B: o } = p.inputs, { transposeA: l, transposeB: c, scale: u, activation: A, scaleA: f, scaleB: d, forceOutputShape: t, perm: n } = p.attrs, B = !k(r), v = !k(o);
11
+ if (B && v) {
12
+ const a = f !== void 0 ? M(r, m(f)) : r, i = d !== void 0 ? M(o, m(d)) : o;
13
+ let e;
14
+ if (u !== void 0 ? e = g(a, i, m(u), l, c) : A === "gelu" ? e = N(a, i) : e = G(a, i, l, c), n)
15
+ if (t) {
16
+ const s = w(e, t);
17
+ e.dispose();
18
+ const b = h(s, n);
19
+ return s.dispose(), b;
20
+ } else {
21
+ const s = h(e, n);
22
+ return e.dispose(), s;
23
+ }
24
+ else if (t) {
25
+ const s = w(e, t);
26
+ return e.dispose(), s;
27
+ } else
28
+ return e;
29
+ }
30
+ throw new Error("Not implemented: matMul16 with packed inputs");
31
+ }
32
+ const C = {
33
+ kernelName: "MatMul16",
34
+ backendName: "webgl",
35
+ kernelFunc: P
36
+ };
37
+ U(C);
@@ -1,7 +1,9 @@
1
- import { f as E, t as R, e as C, j as $, l as N, n as H, u as O } from "../../index-BzFyqcy-.js";
2
- import { r as f } from "../../Reshape-Bowtk9BP.js";
3
- import { M as U } from "../../mulmat_packed_gpu-BT60jmzP.js";
4
- import { m as A } from "../../mat_mul-DzjTFx-u.js";
1
+ import { t as R, e as C, j as N } from "../../index-ZyQhjEPo.js";
2
+ import { r as f } from "../../Reshape-_kILl6tK.js";
3
+ import { M as H } from "../../mulmat_packed_gpu-BMFhLwta.js";
4
+ import { r as E, u as O } from "../../tensor_util-DV-FP5Q3.js";
5
+ import { m as $ } from "../../mat_mul-DeAh4uTH.js";
6
+ import { s as A, a as U } from "../../tensor-DdQUJZlz.js";
5
7
  const M = 0.7978845608028654, g = 0.044715, j = `
6
8
  vec4 x3 = x * x * x;
7
9
  vec4 inner = x + ${g} * x3;
@@ -29,25 +31,25 @@ const M = 0.7978845608028654, g = 0.044715, j = `
29
31
  vec4 du_dx = ${M} * (1.0 + 3.0 * ${g} * a2);
30
32
  vec4 dgelu = 0.5 * (1.0 + t) + 0.5 * a * sech2 * du_dx;
31
33
  return dgelu * b;
32
- `, ne = 1e3;
34
+ `, se = 1e3;
33
35
  function _({
34
36
  a: e,
35
37
  b: n,
36
38
  transposeA: s,
37
39
  transposeB: t,
38
40
  backend: a,
39
- activationSnippet: c,
41
+ activationSnippet: i,
40
42
  multiplier: o
41
43
  }) {
42
- const r = e.shape.length, i = n.shape.length, u = s ? e.shape[r - 2] : e.shape[r - 1], h = t ? n.shape[i - 1] : n.shape[i - 2], p = s ? e.shape[r - 1] : e.shape[r - 2], l = t ? n.shape[i - 2] : n.shape[i - 1], w = e.shape.slice(0, -2), K = n.shape.slice(0, -2), d = $(w), m = $(K), T = N(e.shape.slice(0, -2), n.shape.slice(0, -2)).concat([p, l]);
43
- H(
44
- u === h,
45
- () => `Error in matMul: inner shapes (${u}) and (${h}) of Tensors with shapes ${e.shape} and ${n.shape} and transposeA=${s} and transposeB=${t} must match.`
44
+ const r = e.shape.length, c = n.shape.length, u = s ? e.shape[r - 2] : e.shape[r - 1], p = t ? n.shape[c - 1] : n.shape[c - 2], h = s ? e.shape[r - 1] : e.shape[r - 2], l = t ? n.shape[c - 2] : n.shape[c - 1], w = e.shape.slice(0, -2), K = n.shape.slice(0, -2), m = A(w), d = A(K), T = N(e.shape.slice(0, -2), n.shape.slice(0, -2)).concat([h, l]);
45
+ U(
46
+ u === p,
47
+ () => `Error in matMul: inner shapes (${u}) and (${p}) of Tensors with shapes ${e.shape} and ${n.shape} and transposeA=${s} and transposeB=${t} must match.`
46
48
  );
47
- const v = s ? [d, u, p] : [d, p, u], x = t ? [m, l, h] : [m, h, l], S = f({ inputs: { x: e }, backend: a, attrs: { shape: v } }), b = f({ inputs: { x: n }, backend: a, attrs: { shape: x } }), D = [S, b], y = Math.max(d, m), L = c, B = O(e.dtype, n.dtype), F = new U(
49
+ const v = s ? [m, u, h] : [m, h, u], x = t ? [d, l, p] : [d, p, l], S = f({ inputs: { x: e }, backend: a, attrs: { shape: v } }), b = f({ inputs: { x: n }, backend: a, attrs: { shape: x } }), D = [S, b], y = Math.max(m, d), L = i, B = O(e.dtype, n.dtype), F = new H(
48
50
  v,
49
51
  x,
50
- [y, p, l],
52
+ [y, h, l],
51
53
  s,
52
54
  t,
53
55
  !1,
@@ -84,7 +86,7 @@ E(W);
84
86
  function J(e) {
85
87
  const { dy: n, x: s, kernel: t } = e.inputs, a = e.backend;
86
88
  return R(() => {
87
- const c = C().makeTensorFromTensorInfo(
89
+ const i = C().makeTensorFromTensorInfo(
88
90
  _({
89
91
  a: s,
90
92
  b: t,
@@ -94,7 +96,7 @@ function J(e) {
94
96
  activationSnippet: q,
95
97
  multiplier: n
96
98
  })
97
- ), o = A(c, t, !1, !0), r = A(s, c, !0, !1);
99
+ ), o = $(i, t, !1, !0), r = $(s, i, !0, !1);
98
100
  return [o, r];
99
101
  });
100
102
  }
@@ -105,7 +107,7 @@ const Q = {
105
107
  };
106
108
  E(Q);
107
109
  export {
108
- ne as MATMUL_SHARED_DIM_THRESHOLD,
110
+ se as MATMUL_SHARED_DIM_THRESHOLD,
109
111
  _ as batchMatMulGeluImpl,
110
112
  z as batchMatMulKernel
111
113
  };
@@ -1,28 +1,29 @@
1
- import { f as u } from "../../index-BzFyqcy-.js";
2
- import { batchMatMulGeluImpl as c } from "./matMulGelu.js";
3
- const M = `
1
+ import "../../index-ZyQhjEPo.js";
2
+ import { batchMatMulGeluImpl as u } from "./matMulGelu.js";
3
+ import { r as c } from "../../tensor_util-DV-FP5Q3.js";
4
+ const p = `
4
5
  return a * b;
5
6
  `;
6
- function p(r) {
7
- const { inputs: n, backend: o, attrs: a } = r, { x: t, kernel: e, y: l } = n, { transposeA: i, transposeB: s } = a;
7
+ function M(r) {
8
+ const { inputs: n, backend: o, attrs: a } = r, { x: t, kernel: e, y: i } = n, { transposeA: l, transposeB: s } = a;
8
9
  if (t === void 0 || e === void 0)
9
10
  throw new Error("BatchMatMul requires two input tensors.");
10
- return c({
11
+ return u({
11
12
  a: t,
12
13
  b: e,
13
- transposeA: i,
14
+ transposeA: l,
14
15
  transposeB: s,
15
16
  backend: o,
16
- activationSnippet: M,
17
- multiplier: l
17
+ activationSnippet: p,
18
+ multiplier: i
18
19
  });
19
20
  }
20
21
  const m = {
21
22
  kernelName: "MatMulMul",
22
23
  backendName: "webgl",
23
- kernelFunc: p
24
+ kernelFunc: M
24
25
  };
25
- u(m);
26
+ c(m);
26
27
  export {
27
- p as batchMatMulKernel
28
+ M as batchMatMulKernel
28
29
  };
@@ -1,4 +1,5 @@
1
- import { f as m } from "../../index-BzFyqcy-.js";
1
+ import "../../index-ZyQhjEPo.js";
2
+ import { r as m } from "../../tensor_util-DV-FP5Q3.js";
2
3
  class f {
3
4
  variableNames = ["a", "b"];
4
5
  outputShape;
@@ -7,8 +8,8 @@ class f {
7
8
  { name: "dropoutRate", type: "float" },
8
9
  { name: "seed", type: "float" }
9
10
  ];
10
- constructor(t, r, o) {
11
- this.outputShape = [t, r, o, o], this.userCode = `
11
+ constructor(r, t, o) {
12
+ this.outputShape = [r, t, o, o], this.userCode = `
12
13
  float random(ivec4 coords) {
13
14
  float x = float(coords.x * 4096 + coords.y * 256 + coords.z * 16 + coords.w);
14
15
  return fract(sin(seed + x) * 43758.5453123);
@@ -26,16 +27,16 @@ class f {
26
27
  `;
27
28
  }
28
29
  }
29
- function b(e) {
30
- const { inputs: t, attrs: r } = e, { a: o, b: s } = t, { dropoutRate: a, seed: c } = r, n = e.backend, d = o.shape[0], u = o.shape[2], p = o.shape[1], l = new f(d, p, u);
30
+ function i(e) {
31
+ const { inputs: r, attrs: t } = e, { a: o, b: s } = r, { dropoutRate: a, seed: c } = t, n = e.backend, d = o.shape[0], u = o.shape[2], p = o.shape[1], l = new f(d, p, u);
31
32
  return n.runWebGLProgram(l, [o, s], "float32", [
32
33
  [a ?? 0],
33
34
  [c ?? Math.random() * 1e4]
34
35
  ]);
35
36
  }
36
- const i = {
37
+ const b = {
37
38
  kernelName: "MulDropout",
38
39
  backendName: "webgl",
39
- kernelFunc: b
40
+ kernelFunc: i
40
41
  };
41
- m(i);
42
+ m(b);
@@ -1,5 +1,6 @@
1
- import { f as p, e as G } from "../../index-BzFyqcy-.js";
2
- import { s as x } from "../../sum-CJ0ULhmt.js";
1
+ import { e as G } from "../../index-ZyQhjEPo.js";
2
+ import { r as g } from "../../tensor_util-DV-FP5Q3.js";
3
+ import { s as x } from "../../sum-_fzj5ZTB.js";
3
4
  class y {
4
5
  variableNames = ["x", "meanSquare", "gamma"];
5
6
  outputShape;
@@ -28,7 +29,7 @@ const C = {
28
29
  backendName: "webgl",
29
30
  kernelFunc: v
30
31
  };
31
- p(C);
32
+ g(C);
32
33
  class b {
33
34
  variableNames = ["x", "meanSquare", "dyGamma", "dyXMean"];
34
35
  outputShape;
@@ -73,14 +74,14 @@ function M(t) {
73
74
  l.dispose();
74
75
  const f = new b(n, m, u), S = r.runWebGLProgram(f, [e, d, s, i], "float32");
75
76
  s.dispose(), i.dispose();
76
- const h = new N(n, m, u), g = r.runWebGLProgram(h, [e, d, a], "float32");
77
+ const h = new N(n, m, u), p = r.runWebGLProgram(h, [e, d, a], "float32");
77
78
  d.dispose();
78
- const q = x(G().makeTensorFromTensorInfo(g), [0, 1]);
79
- return r.disposeIntermediateTensorInfo(g), [S, q];
79
+ const q = x(G().makeTensorFromTensorInfo(p), [0, 1]);
80
+ return r.disposeIntermediateTensorInfo(p), [S, q];
80
81
  }
81
82
  const k = {
82
83
  kernelName: "RMSNormGrad",
83
84
  backendName: "webgl",
84
85
  kernelFunc: M
85
86
  };
86
- p(k);
87
+ g(k);
@@ -1,5 +1,6 @@
1
- import { f as i } from "../../index-BzFyqcy-.js";
2
- class l {
1
+ import "../../index-ZyQhjEPo.js";
2
+ import { r as i } from "../../tensor_util-DV-FP5Q3.js";
3
+ class m {
3
4
  variableNames = ["x", "kernel"];
4
5
  outputShape;
5
6
  userCode;
@@ -30,8 +31,8 @@ class l {
30
31
  `;
31
32
  }
32
33
  }
33
- function m(r) {
34
- const { x: e, kernel: t } = r.inputs, { heads: s } = r.attrs, o = r.backend, n = e.shape[0], c = e.shape[1], u = e.shape[2], a = new l(n, s, c, u);
34
+ function l(r) {
35
+ const { x: e, kernel: t } = r.inputs, { heads: s } = r.attrs, o = r.backend, n = e.shape[0], c = e.shape[1], u = e.shape[2], a = new m(n, s, c, u);
35
36
  return [
36
37
  o.runWebGLProgram(a, [e, t], "float32", [[0]]),
37
38
  o.runWebGLProgram(a, [e, t], "float32", [[1]]),
@@ -41,6 +42,6 @@ function m(r) {
41
42
  const d = {
42
43
  kernelName: "QKV",
43
44
  backendName: "webgl",
44
- kernelFunc: m
45
+ kernelFunc: l
45
46
  };
46
47
  i(d);
@@ -1,12 +1,13 @@
1
- import { f as u } from "../../index-BzFyqcy-.js";
2
- class l {
1
+ import "../../index-ZyQhjEPo.js";
2
+ import { r as h } from "../../tensor_util-DV-FP5Q3.js";
3
+ class g {
3
4
  variableNames = ["x", "sin", "cos"];
4
5
  outputShape;
5
6
  userCode;
6
7
  // enableShapeUniforms = true;
7
8
  customUniforms = [{ name: "pastLen", type: "int" }];
8
- constructor(t, n, s, o) {
9
- this.outputShape = [t, n, s, o], this.userCode = `
9
+ constructor(t, s, e, n) {
10
+ this.outputShape = [t, s, e, n], this.userCode = `
10
11
  void main() {
11
12
  ivec4 coords = getOutputCoords(); // [b, h, t, d]
12
13
  int b = coords.x;
@@ -14,7 +15,7 @@ class l {
14
15
  int t = coords.z;
15
16
  int d = coords.w;
16
17
 
17
- int rotaryDim = ${o};
18
+ int rotaryDim = ${n};
18
19
 
19
20
  float outVal = 0.0;
20
21
 
@@ -44,13 +45,13 @@ class l {
44
45
  `;
45
46
  }
46
47
  }
47
- function h(e) {
48
- const { x: t, sin: n, cos: s } = e.inputs, { pastLen: o } = e.attrs, a = e.backend, r = t.shape[0], d = t.shape[1], i = t.shape[2], c = t.shape[3], p = new l(r, d, i, c);
49
- return a.runWebGLProgram(p, [t, n, s], "float32", [[o]]);
48
+ function m(o) {
49
+ const { x: t } = o.inputs, { pastLen: s, ropeCache: e, negSin: n } = o.attrs, a = n ? e.getNegSin() : e.getSin(), r = e.getCos(), i = o.backend, d = t.shape[0], c = t.shape[1], p = t.shape[2], u = t.shape[3], l = new g(d, c, p, u);
50
+ return i.runWebGLProgram(l, [t, a, r], "float32", [[s]]);
50
51
  }
51
52
  const f = {
52
53
  kernelName: "Rope",
53
54
  backendName: "webgl",
54
- kernelFunc: h
55
+ kernelFunc: m
55
56
  };
56
- u(f);
57
+ h(f);
@@ -1,10 +1,11 @@
1
- import { f as i } from "../../index-BzFyqcy-.js";
1
+ import "../../index-ZyQhjEPo.js";
2
+ import { r as i } from "../../tensor_util-DV-FP5Q3.js";
2
3
  class u {
3
4
  variableNames = ["labels", "softmaxProbs", "dy"];
4
5
  outputShape;
5
6
  userCode;
6
- constructor(e, t) {
7
- this.outputShape = [e, t], this.userCode = `
7
+ constructor(t, e) {
8
+ this.outputShape = [t, e], this.userCode = `
8
9
  void main() {
9
10
  ivec2 coords = getOutputCoords();
10
11
  int index = int(getLabels(coords.x));
@@ -16,8 +17,8 @@ class u {
16
17
  }
17
18
  }
18
19
  function d(o) {
19
- const { logits: e, labels: t, dy: r } = o.inputs, s = o.backend, n = t.shape[0], a = e.shape[1], c = new u(n, a);
20
- return s.runWebGLProgram(c, [t, e, r], "float32");
20
+ const { logits: t, labels: e, dy: r } = o.inputs, s = o.backend, n = e.shape[0], a = t.shape[1], c = new u(n, a);
21
+ return s.runWebGLProgram(c, [e, t, r], "float32");
21
22
  }
22
23
  const b = {
23
24
  kernelName: "EfficientScatterSub",
@@ -1,6 +1,8 @@
1
- import { g as p } from "../../webgpu_program-DkQJOJSd.js";
2
- import { f as d, c as l } from "../../webgpu_util-pLEV9tks.js";
3
- import { f, a4 as c } from "../../index-BzFyqcy-.js";
1
+ import { e as p } from "../../webgpu_program-Cigz-7RF.js";
2
+ import { f as d, c as l } from "../../webgpu_util-BBCnKm2X.js";
3
+ import "../../index-ZyQhjEPo.js";
4
+ import { j as f } from "../../tensor-DdQUJZlz.js";
5
+ import { r as c } from "../../tensor_util-DV-FP5Q3.js";
4
6
  class h {
5
7
  variableNames = ["moments", "value"];
6
8
  outputShape;
@@ -36,19 +38,19 @@ class h {
36
38
  }
37
39
  }
38
40
  function v(t) {
39
- const { moments: e, value: a } = t.inputs, { beta1: n, beta2: s, learningRate: o, epsilon: i } = t.attrs, r = t.backend;
40
- c(e.shape, [...a.shape, 2], "Error in AdamAdjust: ");
41
- const u = new h(a.shape), m = [
41
+ const { moments: e, value: a } = t.inputs, { beta1: n, beta2: o, learningRate: s, epsilon: i } = t.attrs, r = t.backend;
42
+ f(e.shape, [...a.shape, 2], "Error in AdamAdjust: ");
43
+ const m = new h(a.shape), u = [
42
44
  { type: "float32", data: [1 / n] },
43
- { type: "float32", data: [1 / s] },
44
- { type: "float32", data: [o] },
45
+ { type: "float32", data: [1 / o] },
46
+ { type: "float32", data: [s] },
45
47
  { type: "float32", data: [i] }
46
48
  ];
47
- return r.runWebGPUProgram(u, [e, a], "float32", m);
49
+ return r.runWebGPUProgram(m, [e, a], "float32", u);
48
50
  }
49
51
  const b = {
50
52
  kernelName: "AdamAdjust",
51
53
  backendName: "webgpu",
52
54
  kernelFunc: v
53
55
  };
54
- f(b);
56
+ c(b);
@@ -1,7 +1,9 @@
1
- import { g as m } from "../../webgpu_program-DkQJOJSd.js";
2
- import { f as u, c as p } from "../../webgpu_util-pLEV9tks.js";
3
- import { f as d, a4 as c } from "../../index-BzFyqcy-.js";
4
- class h {
1
+ import { e as p } from "../../webgpu_program-Cigz-7RF.js";
2
+ import { f as u, c as d } from "../../webgpu_util-BBCnKm2X.js";
3
+ import "../../index-ZyQhjEPo.js";
4
+ import { j as f } from "../../tensor-DdQUJZlz.js";
5
+ import { r as c } from "../../tensor_util-DV-FP5Q3.js";
6
+ class l {
5
7
  variableNames = ["moments", "gradient"];
6
8
  outputShape;
7
9
  shaderKey = "AdamMoments";
@@ -9,11 +11,11 @@ class h {
9
11
  dispatch;
10
12
  workgroupSize = [64, 1, 1];
11
13
  size = !0;
12
- uniforms = "beta1: f32, beta2: f32";
14
+ uniforms = "beta1: f32, beta2: f32, lossScaling: f32";
13
15
  outputComponent = 2;
14
16
  variableComponents = [2, 1];
15
- constructor(e) {
16
- this.outputShape = e, this.dispatchLayout = u(this.outputShape.slice(0, -1)), this.dispatch = p(
17
+ constructor(t) {
18
+ this.outputShape = t, this.dispatchLayout = u(this.outputShape.slice(0, -1)), this.dispatch = d(
17
19
  this.dispatchLayout,
18
20
  this.outputShape.slice(0, -1),
19
21
  this.workgroupSize,
@@ -22,12 +24,12 @@ class h {
22
24
  }
23
25
  getUserCode() {
24
26
  return `
25
- ${m("index")} {
27
+ ${p("index")} {
26
28
  if (index < uniforms.size) {
27
29
  let m: vec2<f32> = moments[index];
28
30
 
29
31
  // Add gradient clipping here
30
- let g: f32 = clamp(gradient[index], -1.0, 1.0);
32
+ let g: f32 = clamp(gradient[index] * uniforms.lossScaling, -1.0, 1.0);
31
33
 
32
34
  let newM1 = fma(m.x, uniforms.beta1, g * (1.0 - uniforms.beta1));
33
35
  let newM2 = fma(m.y, uniforms.beta2, g * g * (1.0 - uniforms.beta2));
@@ -38,21 +40,24 @@ class h {
38
40
  `;
39
41
  }
40
42
  }
41
- function f(t) {
42
- const { moments: e, gradient: o } = t.inputs, { beta1: a, beta2: n } = t.attrs, r = t.backend;
43
- if (c(e.shape, [...o.shape, 2], "Error in AdamMoments: "), a < 0 || a >= 1)
44
- throw new Error(`Invalid beta1 value: ${a}. Must be in the range [0, 1).`);
45
- if (n < 0 || n >= 1)
46
- throw new Error(`Invalid beta2 value: ${n}. Must be in the range [0, 1).`);
47
- const s = new h(e.shape), i = [
48
- { type: "float32", data: [a] },
49
- { type: "float32", data: [n] }
43
+ function h(e) {
44
+ const { moments: t, gradient: a } = e.inputs, { beta1: n, beta2: o, lossScaling: r } = e.attrs, s = e.backend;
45
+ if (a.dtype !== "float32")
46
+ throw new Error(`Gradient must be float32, but got ${a.dtype}`);
47
+ if (f(t.shape, [...a.shape, 2], "Error in AdamMoments: "), n < 0 || n >= 1)
48
+ throw new Error(`Invalid beta1 value: ${n}. Must be in the range [0, 1).`);
49
+ if (o < 0 || o >= 1)
50
+ throw new Error(`Invalid beta2 value: ${o}. Must be in the range [0, 1).`);
51
+ const i = new l(t.shape), m = [
52
+ { type: "float32", data: [n] },
53
+ { type: "float32", data: [o] },
54
+ { type: "float32", data: [1 / r] }
50
55
  ];
51
- return r.runWebGPUProgram(s, [e, o], "float32", i);
56
+ return s.runWebGPUProgram(i, [t, a], "float32", m);
52
57
  }
53
- const l = {
58
+ const g = {
54
59
  kernelName: "AdamMoments",
55
60
  backendName: "webgpu",
56
- kernelFunc: f
61
+ kernelFunc: h
57
62
  };
58
- d(l);
63
+ c(g);
@@ -0,0 +1 @@
1
+ export {};
@@ -0,0 +1,14 @@
1
+ import "../../index-ZyQhjEPo.js";
2
+ import { BinaryOpProgram as p } from "./utils/binary_op.js";
3
+ import { B as m } from "../../binary_op_util-pKXltfxI.js";
4
+ import { r as c } from "../../tensor_util-DV-FP5Q3.js";
5
+ function i(r) {
6
+ const { a: e, b: n } = r.inputs, a = r.backend, t = new p(m.ADD, e.shape, n.shape), o = a.runWebGPUProgram(t, [e, n], "int32");
7
+ return o.packed = !0, o;
8
+ }
9
+ const s = {
10
+ kernelName: "Add16",
11
+ backendName: "webgpu",
12
+ kernelFunc: i
13
+ };
14
+ c(s);
@@ -1,7 +1,10 @@
1
- import { g as u } from "../../webgpu_program-DkQJOJSd.js";
2
- import { f as d, c as m } from "../../webgpu_util-pLEV9tks.js";
3
- import { f as l, a4 as f } from "../../index-BzFyqcy-.js";
4
- class g {
1
+ import { isPackedTensor as S } from "../../utilities/packed.js";
2
+ import { e as d } from "../../webgpu_program-Cigz-7RF.js";
3
+ import { f as u, c as m } from "../../webgpu_util-BBCnKm2X.js";
4
+ import "../../index-ZyQhjEPo.js";
5
+ import { j as g } from "../../tensor-DdQUJZlz.js";
6
+ import { r as x } from "../../tensor_util-DV-FP5Q3.js";
7
+ class b {
5
8
  variableNames = ["cache", "item"];
6
9
  outputShape;
7
10
  shaderKey = "AppendCache";
@@ -10,14 +13,14 @@ class g {
10
13
  workgroupSize = [64, 1, 1];
11
14
  size = !0;
12
15
  uniforms = "cacheT: i32";
13
- constructor(e, t, s, a, c) {
14
- const o = Math.min(s + 1, c);
15
- this.shaderKey = `AppendCache_${o}`, this.outputShape = [e, t, o, a], this.dispatchLayout = d(this.outputShape), this.dispatch = m(this.dispatchLayout, this.outputShape, this.workgroupSize);
16
+ constructor(e, t, s, o, i) {
17
+ const a = Math.min(s + 1, i);
18
+ this.shaderKey = `AppendCache_${a}`, this.outputShape = [e, t, a, o], this.dispatchLayout = u(this.outputShape), this.dispatch = m(this.dispatchLayout, this.outputShape, this.workgroupSize);
16
19
  }
17
20
  getUserCode() {
18
21
  const e = this.outputShape[2];
19
22
  return `
20
- ${u("index")} {
23
+ ${d("index")} {
21
24
  if (index < uniforms.size) {
22
25
  let coords = getCoordsFromIndex(index); // [b, h, t, d]
23
26
  let b = coords[0];
@@ -45,16 +48,60 @@ class g {
45
48
  `;
46
49
  }
47
50
  }
48
- function S(r) {
49
- const { cache: e, item: t } = r.inputs, { maxSize: s, pastLen: a } = r.attrs, c = r.backend, o = e.shape[0], n = e.shape[2], i = e.shape[1];
50
- if (f(t.shape, [o, i, 1, t.shape[3]], "Error in AppendCache: "), a < 0 || a > s)
51
- throw new Error(`Invalid pastLen value: ${a}. Must be in the range [0, ${s}].`);
52
- const h = new g(o, i, n, t.shape[3], s), p = [{ type: "int32", data: [a] }];
53
- return c.runWebGPUProgram(h, [e, t], "float32", p);
51
+ class C {
52
+ variableNames = ["cache", "item"];
53
+ outputShape;
54
+ shaderKey = "AppendCache";
55
+ dispatchLayout;
56
+ dispatch;
57
+ workgroupSize = [64, 1, 1];
58
+ size = !0;
59
+ uniforms = "cacheT: i32";
60
+ constructor(e, t, s, o, i) {
61
+ const a = Math.min(s + 1, i);
62
+ this.shaderKey = `AppendCache_${a}`, this.outputShape = [e, t, a, o], this.dispatchLayout = u(this.outputShape), this.dispatch = m(this.dispatchLayout, this.outputShape, this.workgroupSize);
63
+ }
64
+ getUserCode() {
65
+ const e = this.outputShape[2];
66
+ return `
67
+ ${d("index")} {
68
+ if (index < uniforms.size) {
69
+ let coords = getCoordsFromIndex(index); // [b, h, t, d]
70
+ let b = coords[0];
71
+ let h = coords[1];
72
+ let t = coords[2];
73
+ let d = coords[3];
74
+
75
+ let itemT = 1;
76
+ let maxSize = ${e};
77
+ let totalT = uniforms.cacheT + itemT;
78
+ let start = select(0, 1, totalT >= maxSize);
79
+
80
+ let srcT = t + start;
81
+ var val: i32 = 0i;
82
+ if (srcT < uniforms.cacheT) {
83
+ val = cache[getIndexFromCoords4D(vec4<i32>(b, h, srcT, d), uniforms.cacheShape)];
84
+ }
85
+ if (srcT == uniforms.cacheT) {
86
+ val = item[getIndexFromCoords4D(vec4<i32>(b, h, 0, d), uniforms.itemShape)];
87
+ }
88
+
89
+ result[index] = val;
90
+ }
91
+ }
92
+ `;
93
+ }
94
+ }
95
+ function v(r) {
96
+ const { cache: e, item: t } = r.inputs, { maxSize: s, pastLen: o } = r.attrs, i = r.backend, a = S(e), c = e.shape[0], n = e.shape[2], h = e.shape[1];
97
+ if (g(t.shape, [c, h, 1, t.shape[3]], "Error in AppendCache: "), o < 0 || o > s)
98
+ throw new Error(`Invalid pastLen value: ${o}. Must be in the range [0, ${s}].`);
99
+ const l = a ? new C(c, h, n, t.shape[3], s) : new b(c, h, n, t.shape[3], s), f = [{ type: "int32", data: [o] }], T = a ? "int32" : e.dtype, p = i.runWebGPUProgram(l, [e, t], T, f);
100
+ return p.packed = a, p;
54
101
  }
55
- const T = {
102
+ const z = {
56
103
  kernelName: "AppendCache",
57
104
  backendName: "webgpu",
58
- kernelFunc: S
105
+ kernelFunc: v
59
106
  };
60
- l(T);
107
+ x(z);