@genai-fi/nanogpt 0.11.0 → 0.12.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/Generator.js +29 -29
- package/dist/{RealDiv-Ds-jvL09.js → RealDiv-C8neBwFi.js} +17 -17
- package/dist/{Reshape-Cd6e-Otn.js → Reshape-Bd4V_4X7.js} +1 -1
- package/dist/{Reshape-Ct266DEk.js → Reshape-Ck29jQSY.js} +7 -7
- package/dist/TeachableLLM.d.ts +2 -1
- package/dist/TeachableLLM.js +9 -9
- package/dist/Trainer.d.ts +4 -2
- package/dist/Trainer.js +12 -9
- package/dist/{axis_util-DofAuy0p.js → axis_util-DGqbT-FX.js} +1 -1
- package/dist/backend.js +2 -2
- package/dist/{backend_util-C7NWHpv7.js → backend_util-DC3rBo_H.js} +18 -18
- package/dist/{backend_webgpu-B0Vls736.js → backend_webgpu-mbhNnlx9.js} +10 -10
- package/dist/{broadcast_to-DDaNMbX7.js → broadcast_to-D1Dmg2Oz.js} +2 -2
- package/dist/checks/appendCache.js +2 -2
- package/dist/checks/attentionMask.js +3 -3
- package/dist/checks/gelu.js +2 -2
- package/dist/checks/matMulGelu.js +2 -2
- package/dist/checks/normRMS.js +4 -4
- package/dist/checks/normRMSGrad.js +3 -3
- package/dist/checks/packUnpack.js +2 -2
- package/dist/checks/qkv.js +2 -2
- package/dist/checks/rope.js +2 -2
- package/dist/clip_by_value-fg2aKzUy.js +12 -0
- package/dist/{complex-DClmWqJt.js → complex-Cyg-eQeZ.js} +1 -1
- package/dist/concat-CSm2rMwe.js +17 -0
- package/dist/{concat_util-CHsJFZJJ.js → concat_util-D0je5Ppu.js} +1 -1
- package/dist/{dataset-DcjWqUVQ.js → dataset-CVIJu7Xa.js} +3 -3
- package/dist/{dropout-OxuaJz6z.js → dropout-DLhSMNTZ.js} +14 -14
- package/dist/expand_dims-ChkuOp6I.js +11 -0
- package/dist/{exports_initializers-eS9QJ6ut.js → exports_initializers-1KWPiStI.js} +1 -1
- package/dist/{floor-DIb-lN_u.js → floor-BRMPgeIs.js} +1 -1
- package/dist/gather-BSULDalH.js +9 -0
- package/dist/{gelu-DqTbCx5x.js → gelu-BK1k-n1i.js} +1 -1
- package/dist/{gpgpu_math-CJcbnKPC.js → gpgpu_math-BJSTk_mW.js} +25 -25
- package/dist/{index-Dj5TkmPY.js → index-BBVLAXZD.js} +14 -14
- package/dist/{index-D0RBWjq8.js → index-Duu1Lvvv.js} +45 -45
- package/dist/{kernel_funcs_utils-CSaumNDs.js → kernel_funcs_utils-BtYrPoJu.js} +8 -8
- package/dist/layers/BaseLayer.js +2 -2
- package/dist/layers/CausalSelfAttention.js +6 -6
- package/dist/layers/MLP.js +4 -4
- package/dist/layers/PositionEmbedding.js +5 -5
- package/dist/layers/RMSNorm.js +3 -3
- package/dist/layers/RoPECache.js +4 -4
- package/dist/layers/TiedEmbedding.js +6 -6
- package/dist/layers/TransformerBlock.js +1 -1
- package/dist/loader/loadTransformers.js +1 -1
- package/dist/loader/oldZipLoad.js +17 -17
- package/dist/{log_sum_exp-VLZgbFAH.js → log_sum_exp-CVqLsVLl.js} +4 -4
- package/dist/main.d.ts +9 -0
- package/dist/main.js +69 -58
- package/dist/{matMul16-cDxwemKj.js → matMul16-xswmhSuF.js} +7 -7
- package/dist/{matMulGelu-B2s_80-H.js → matMulGelu-BpvgnYG8.js} +26 -26
- package/dist/mat_mul-Bn2BDpT4.js +11 -0
- package/dist/{mod-PrOKlFxH.js → mod-B4AUd1Np.js} +1 -1
- package/dist/models/NanoGPTV1.js +2 -2
- package/dist/models/model.js +9 -9
- package/dist/{ones-BX_wEgzB.js → ones-CBI1AQjb.js} +3 -3
- package/dist/ops/adamAdjust.js +1 -1
- package/dist/ops/adamMoments.js +1 -1
- package/dist/ops/add16.js +1 -1
- package/dist/ops/appendCache.js +3 -3
- package/dist/ops/attentionMask.js +1 -1
- package/dist/ops/concat16.js +2 -2
- package/dist/ops/cpu/adamAdjust.js +7 -7
- package/dist/ops/cpu/adamMoments.js +5 -5
- package/dist/ops/cpu/appendCache.js +6 -6
- package/dist/ops/cpu/attentionMask.js +6 -6
- package/dist/ops/cpu/fusedSoftmax.js +5 -5
- package/dist/ops/cpu/gatherSub.js +7 -7
- package/dist/ops/cpu/gelu.js +5 -5
- package/dist/ops/cpu/matMul16.js +2 -2
- package/dist/ops/cpu/matMulGelu.js +3 -3
- package/dist/ops/cpu/matMulMul.js +5 -5
- package/dist/ops/cpu/mulDropout.js +1 -1
- package/dist/ops/cpu/normRMS.js +5 -5
- package/dist/ops/cpu/qkv.js +3 -3
- package/dist/ops/cpu/rope.js +9 -9
- package/dist/ops/cpu/scatterSub.js +5 -5
- package/dist/ops/dot16.js +2 -2
- package/dist/ops/gatherSub.js +1 -1
- package/dist/ops/gelu.js +2 -2
- package/dist/ops/grads/add16.js +1 -1
- package/dist/ops/grads/attentionMask.js +2 -2
- package/dist/ops/grads/gelu.js +2 -2
- package/dist/ops/grads/matMul16.js +3 -3
- package/dist/ops/grads/matMulGelu.js +5 -5
- package/dist/ops/grads/normRMS.js +6 -6
- package/dist/ops/grads/pack16.js +3 -3
- package/dist/ops/grads/qkv.js +9 -9
- package/dist/ops/grads/rope.js +2 -2
- package/dist/ops/grads/softmax16.js +1 -1
- package/dist/ops/grads/unpack16.js +2 -2
- package/dist/ops/matMul16.js +3 -3
- package/dist/ops/matMulGelu.js +2 -2
- package/dist/ops/matMulMul.js +1 -1
- package/dist/ops/mul16.js +1 -1
- package/dist/ops/mulDrop.js +1 -1
- package/dist/ops/normRMS.js +1 -1
- package/dist/ops/pack16.js +2 -2
- package/dist/ops/qkv.js +1 -1
- package/dist/ops/reshape16.js +6 -6
- package/dist/ops/rope.js +2 -2
- package/dist/ops/scatterSub.js +1 -1
- package/dist/ops/slice16.js +2 -2
- package/dist/ops/softmax16.js +1 -1
- package/dist/ops/sub16.js +1 -1
- package/dist/ops/sum16.js +2 -2
- package/dist/ops/transpose16.js +6 -6
- package/dist/ops/unpack16.js +2 -2
- package/dist/ops/webgl/adamAdjust.js +2 -2
- package/dist/ops/webgl/adamMoments.js +1 -1
- package/dist/ops/webgl/appendCache.js +1 -1
- package/dist/ops/webgl/attentionMask.js +4 -4
- package/dist/ops/webgl/fusedSoftmax.js +6 -6
- package/dist/ops/webgl/gatherSub.js +1 -1
- package/dist/ops/webgl/gelu.js +2 -2
- package/dist/ops/webgl/log.js +3 -3
- package/dist/ops/webgl/matMul16.js +10 -10
- package/dist/ops/webgl/matMulGelu.js +4 -4
- package/dist/ops/webgl/matMulMul.js +2 -2
- package/dist/ops/webgl/mulDropout.js +1 -1
- package/dist/ops/webgl/normRMS.js +2 -2
- package/dist/ops/webgl/qkv.js +1 -1
- package/dist/ops/webgl/rope.js +4 -4
- package/dist/ops/webgl/scatterSub.js +1 -1
- package/dist/ops/webgpu/adamAdjust.js +3 -3
- package/dist/ops/webgpu/adamMoments.js +5 -5
- package/dist/ops/webgpu/add16.js +1 -1
- package/dist/ops/webgpu/appendCache.js +3 -3
- package/dist/ops/webgpu/attentionMask.js +5 -5
- package/dist/ops/webgpu/attentionMask32_program.js +2 -2
- package/dist/ops/webgpu/concat16.js +5 -5
- package/dist/ops/webgpu/gatherSub.js +3 -3
- package/dist/ops/webgpu/gelu.js +3 -3
- package/dist/ops/webgpu/matMul16.js +19 -19
- package/dist/ops/webgpu/matMul16_program.js +2 -2
- package/dist/ops/webgpu/mul16.js +1 -1
- package/dist/ops/webgpu/normRMS.js +2 -2
- package/dist/ops/webgpu/normRMSGrad.js +4 -4
- package/dist/ops/webgpu/pack16.js +3 -3
- package/dist/ops/webgpu/pack16_program.js +2 -2
- package/dist/ops/webgpu/qkv.js +4 -4
- package/dist/ops/webgpu/rope.js +3 -3
- package/dist/ops/webgpu/scatterSub.js +3 -3
- package/dist/ops/webgpu/slice16.js +4 -4
- package/dist/ops/webgpu/softmax16.js +4 -4
- package/dist/ops/webgpu/softmax16_program.js +2 -2
- package/dist/ops/webgpu/softmax16_subgroup_program.js +2 -2
- package/dist/ops/webgpu/softmax16grad.js +1 -1
- package/dist/ops/webgpu/sub16.js +1 -1
- package/dist/ops/webgpu/sum16.js +5 -5
- package/dist/ops/webgpu/transpose16.js +2 -2
- package/dist/ops/webgpu/transpose16_program.js +2 -2
- package/dist/ops/webgpu/transpose16_shared_program.js +3 -3
- package/dist/ops/webgpu/unpack16.js +5 -5
- package/dist/ops/webgpu/utils/binary_op.js +3 -3
- package/dist/ops/webgpu/utils/reductions.js +4 -4
- package/dist/{ops-FJapAPfm.js → ops-C2_OXuZ4.js} +35 -35
- package/dist/{pack16-k4jq6aMX.js → pack16-atD0eYRm.js} +6 -6
- package/dist/patches/webgpu_backend.js +8 -8
- package/dist/patches/webgpu_base.js +1 -1
- package/dist/patches/webgpu_program.js +2 -2
- package/dist/{random_width-UGQn4OWb.js → random_width-BN4wGJaW.js} +33 -33
- package/dist/{range-CuGvVN2c.js → range-DKmP1-OQ.js} +1 -1
- package/dist/relu-BsXmGzzu.js +9 -0
- package/dist/{reshape-CkjKPPqB.js → reshape-BI0yzp1T.js} +1 -1
- package/dist/{resize_nearest_neighbor-DB8k9KN_.js → resize_nearest_neighbor-BA_BX-ub.js} +25 -25
- package/dist/{rope-BmZmp9uP.js → rope-DJ7Y7c-u.js} +1 -1
- package/dist/{scatter_nd_util-BY22Cc-C.js → scatter_nd_util-k9MUVUkn.js} +1 -1
- package/dist/{selu_util-BuLbmbrl.js → selu_util-DyW0X1WG.js} +5 -5
- package/dist/{shared-B7USJZgw.js → shared-Q3BS6T03.js} +1 -1
- package/dist/{shared-BQboIImQ.js → shared-nnSWpC3u.js} +6 -6
- package/dist/{slice-Aqy7KbJh.js → slice-wBNvzVyz.js} +3 -3
- package/dist/{slice_util-D8CQRenR.js → slice_util-zN8KFC5I.js} +7 -7
- package/dist/{softmax-faLoUZVT.js → softmax-DfuYyjMh.js} +1 -1
- package/dist/split-BYrLboMq.js +9 -0
- package/dist/squeeze-Bk8Brcct.js +10 -0
- package/dist/{stack-WJK22CFn.js → stack-CDWShFHF.js} +1 -1
- package/dist/{step-dXR33iOg.js → step-BS5JXRR6.js} +14 -14
- package/dist/sum-BPUfDB2X.js +11 -0
- package/dist/{tensor-BQqrDvpx.js → tensor-CEt9Nm2s.js} +1 -1
- package/dist/{tensor1d-LxP9asMm.js → tensor1d-Cc_KCIDg.js} +1 -1
- package/dist/{tensor2d-BN1sSfQO.js → tensor2d-BN97fF71.js} +1 -1
- package/dist/{tensor4d-DVwr7pLF.js → tensor4d-vuDDgdUI.js} +1 -1
- package/dist/{tfjs_backend-Vi4JfLzT.js → tfjs_backend-806hyYve.js} +36 -36
- package/dist/tile-OWUvpIVt.js +11 -0
- package/dist/tokeniser/BaseTokeniser.d.ts +6 -8
- package/dist/tokeniser/BaseTokeniser.js +6 -6
- package/dist/tokeniser/CharTokeniser.d.ts +6 -6
- package/dist/tokeniser/CharTokeniser.js +26 -26
- package/dist/tokeniser/bpe.d.ts +6 -6
- package/dist/tokeniser/bpe.js +9 -9
- package/dist/tokeniser/type.d.ts +6 -8
- package/dist/training/Adam.js +2 -2
- package/dist/training/AdamExt.js +1 -1
- package/dist/training/DatasetBuilder.d.ts +1 -1
- package/dist/training/DatasetBuilder.js +29 -29
- package/dist/training/FullTrainer.js +1 -1
- package/dist/training/Trainer.d.ts +5 -4
- package/dist/training/Trainer.js +37 -40
- package/dist/training/sparseCrossEntropy.js +3 -3
- package/dist/training/tasks/ConversationTask.d.ts +11 -0
- package/dist/training/tasks/ConversationTask.js +26 -0
- package/dist/training/tasks/PretrainingTask.d.ts +11 -0
- package/dist/training/tasks/PretrainingTask.js +34 -0
- package/dist/training/tasks/StartSentenceTask.d.ts +12 -0
- package/dist/training/tasks/StartSentenceTask.js +42 -0
- package/dist/training/tasks/Task.d.ts +8 -0
- package/dist/training/tasks/Task.js +44 -0
- package/dist/{transpose-JawVKyZy.js → transpose-BUkQCJp9.js} +7 -7
- package/dist/{unsorted_segment_sum-LAbmE9G4.js → unsorted_segment_sum-BljxHhCY.js} +78 -78
- package/dist/utilities/dummy.js +3 -3
- package/dist/utilities/multinomialCPU.js +2 -2
- package/dist/utilities/packed.js +1 -1
- package/dist/utilities/performance.js +1 -1
- package/dist/utilities/profile.js +1 -1
- package/dist/utilities/safetensors.js +2 -2
- package/dist/utilities/sentences.d.ts +1 -1
- package/dist/utilities/sentences.js +11 -11
- package/dist/utilities/weights.js +2 -2
- package/dist/{variable-DQ9yYgEU.js → variable-DPt_Iuog.js} +1 -1
- package/dist/{webgpu_program-CAE4RICo.js → webgpu_program-BpWRlghH.js} +1 -1
- package/dist/{webgpu_util-BdovYhXr.js → webgpu_util-DMiKzzQM.js} +7 -7
- package/dist/{zeros-DeiE2zTa.js → zeros-5YROwwUH.js} +2 -2
- package/dist/{zeros_like-BAz3iKru.js → zeros_like-De4n1C3m.js} +57 -57
- package/package.json +1 -1
- package/dist/clip_by_value-Dn5tzexi.js +0 -12
- package/dist/concat-C6X3AAlQ.js +0 -17
- package/dist/expand_dims-BzfJK2uc.js +0 -11
- package/dist/gather-BcO5UQNJ.js +0 -9
- package/dist/mat_mul-DxpNTCRz.js +0 -11
- package/dist/relu-Cf80uA2p.js +0 -9
- package/dist/split-BNz5jcGc.js +0 -9
- package/dist/squeeze--YMgaAAf.js +0 -10
- package/dist/sum-BdplSvq_.js +0 -11
- package/dist/tile-CvN_LyVr.js +0 -11
|
@@ -1,9 +1,9 @@
|
|
|
1
|
-
import { t as c, al as u, b as i } from "../index-
|
|
1
|
+
import { t as c, al as u, b as i } from "../index-Duu1Lvvv.js";
|
|
2
2
|
import f from "./BaseLayer.js";
|
|
3
|
-
import { E as g, D as h } from "../random_width-
|
|
4
|
-
import { r as l } from "../exports_initializers-
|
|
5
|
-
import { m as b } from "../mod-
|
|
6
|
-
import { r as w } from "../range-
|
|
3
|
+
import { E as g, D as h } from "../random_width-BN4wGJaW.js";
|
|
4
|
+
import { r as l } from "../exports_initializers-1KWPiStI.js";
|
|
5
|
+
import { m as b } from "../mod-B4AUd1Np.js";
|
|
6
|
+
import { r as w } from "../range-DKmP1-OQ.js";
|
|
7
7
|
function E(t) {
|
|
8
8
|
return new h(t);
|
|
9
9
|
}
|
package/dist/layers/RMSNorm.js
CHANGED
|
@@ -1,8 +1,8 @@
|
|
|
1
|
-
import { t as s } from "../index-
|
|
1
|
+
import { t as s } from "../index-Duu1Lvvv.js";
|
|
2
2
|
import e from "./BaseLayer.js";
|
|
3
3
|
import { normRMS as a } from "../ops/normRMS.js";
|
|
4
|
-
import { v as i } from "../variable-
|
|
5
|
-
import { o as m } from "../ones-
|
|
4
|
+
import { v as i } from "../variable-DPt_Iuog.js";
|
|
5
|
+
import { o as m } from "../ones-CBI1AQjb.js";
|
|
6
6
|
class l extends e {
|
|
7
7
|
GAMMA;
|
|
8
8
|
constructor(r, t = "", o) {
|
package/dist/layers/RoPECache.js
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
|
-
import {
|
|
2
|
-
import { r as l } from "../zeros_like-
|
|
3
|
-
import { c as f, s as m } from "../unsorted_segment_sum-
|
|
4
|
-
import { r as h } from "../range-
|
|
1
|
+
import { j as a, b as n, p as c, t as p, k as r } from "../index-Duu1Lvvv.js";
|
|
2
|
+
import { r as l } from "../zeros_like-De4n1C3m.js";
|
|
3
|
+
import { c as f, s as m } from "../unsorted_segment_sum-BljxHhCY.js";
|
|
4
|
+
import { r as h } from "../range-DKmP1-OQ.js";
|
|
5
5
|
class D {
|
|
6
6
|
constructor(s) {
|
|
7
7
|
this.config = s;
|
|
@@ -1,13 +1,13 @@
|
|
|
1
|
-
import "../random_width-
|
|
2
|
-
import "../index-
|
|
3
|
-
import { r as s } from "../exports_initializers-
|
|
1
|
+
import "../random_width-BN4wGJaW.js";
|
|
2
|
+
import "../index-Duu1Lvvv.js";
|
|
3
|
+
import { r as s } from "../exports_initializers-1KWPiStI.js";
|
|
4
4
|
import a from "./BaseLayer.js";
|
|
5
5
|
import { dot16 as o } from "../ops/dot16.js";
|
|
6
6
|
import { isPackedTensor as r } from "../utilities/packed.js";
|
|
7
|
-
import { p as m } from "../pack16-
|
|
7
|
+
import { p as m } from "../pack16-atD0eYRm.js";
|
|
8
8
|
import { transpose16 as d } from "../ops/transpose16.js";
|
|
9
|
-
import { v as p } from "../variable-
|
|
10
|
-
import { g as h } from "../gather-
|
|
9
|
+
import { v as p } from "../variable-DPt_Iuog.js";
|
|
10
|
+
import { g as h } from "../gather-BSULDalH.js";
|
|
11
11
|
class g extends a {
|
|
12
12
|
vocabSize;
|
|
13
13
|
embedDim;
|
|
@@ -2,7 +2,7 @@ import p from "./CausalSelfAttention.js";
|
|
|
2
2
|
import m from "./MLP.js";
|
|
3
3
|
import d from "./RMSNorm.js";
|
|
4
4
|
import h from "./BaseLayer.js";
|
|
5
|
-
import { k as n, t as u } from "../index-
|
|
5
|
+
import { k as n, t as u } from "../index-Duu1Lvvv.js";
|
|
6
6
|
import { add16 as l } from "../ops/add16.js";
|
|
7
7
|
class g extends h {
|
|
8
8
|
ln1;
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import l from "../tokeniser/CharTokeniser.js";
|
|
2
2
|
import c from "../tokeniser/bpe.js";
|
|
3
3
|
import { load_safetensors as b } from "../utilities/safetensors.js";
|
|
4
|
-
import {
|
|
4
|
+
import { a9 as y } from "../index-Duu1Lvvv.js";
|
|
5
5
|
import { dummyPassAsync as u } from "../utilities/dummy.js";
|
|
6
6
|
import _ from "../models/factory.js";
|
|
7
7
|
async function L(e, a, r, t) {
|
|
@@ -1,22 +1,22 @@
|
|
|
1
|
-
import {
|
|
2
|
-
import "../random_width-
|
|
3
|
-
import "../zeros_like-
|
|
1
|
+
import { a9 as y } from "../index-Duu1Lvvv.js";
|
|
2
|
+
import "../random_width-BN4wGJaW.js";
|
|
3
|
+
import "../zeros_like-De4n1C3m.js";
|
|
4
4
|
import "../Generator.js";
|
|
5
5
|
import "../index-Cp39cXWe.js";
|
|
6
|
-
import "../dataset-
|
|
6
|
+
import "../dataset-CVIJu7Xa.js";
|
|
7
7
|
import "../ops/cpu/attentionMask.js";
|
|
8
8
|
import "../ops/webgl/attentionMask.js";
|
|
9
9
|
import "../ops/grads/attentionMask.js";
|
|
10
10
|
import "../ops/cpu/rope.js";
|
|
11
11
|
import "../ops/webgl/rope.js";
|
|
12
|
-
import "../rope-
|
|
12
|
+
import "../rope-DJ7Y7c-u.js";
|
|
13
13
|
import "../ops/cpu/appendCache.js";
|
|
14
14
|
import "../ops/webgl/appendCache.js";
|
|
15
15
|
import "../ops/grads/softmax16.js";
|
|
16
|
-
import "../matMul16-
|
|
16
|
+
import "../matMul16-xswmhSuF.js";
|
|
17
17
|
import "../ops/webgl/matMul16.js";
|
|
18
18
|
import "../ops/cpu/matMul16.js";
|
|
19
|
-
import "../pack16-
|
|
19
|
+
import "../pack16-atD0eYRm.js";
|
|
20
20
|
import "../ops/transpose16.js";
|
|
21
21
|
import "../ops/reshape16.js";
|
|
22
22
|
import "../ops/cpu/qkv.js";
|
|
@@ -42,26 +42,26 @@ import "../ops/cpu/adamMoments.js";
|
|
|
42
42
|
import "../ops/webgl/adamMoments.js";
|
|
43
43
|
import "../papaparse.min-C0cScC2i.js";
|
|
44
44
|
import "../ops/cpu/matMulGelu.js";
|
|
45
|
-
import "../matMulGelu-
|
|
45
|
+
import "../matMulGelu-BpvgnYG8.js";
|
|
46
46
|
import "../ops/grads/matMulGelu.js";
|
|
47
47
|
import "../ops/cpu/gelu.js";
|
|
48
48
|
import "../ops/webgl/gelu.js";
|
|
49
|
-
import "../gelu-
|
|
49
|
+
import "../gelu-BK1k-n1i.js";
|
|
50
50
|
import "../ops/webgl/log.js";
|
|
51
51
|
import "../checks/normRMS.js";
|
|
52
52
|
import "../checks/normRMSGrad.js";
|
|
53
53
|
import { importWeights as u } from "../utilities/weights.js";
|
|
54
54
|
async function Mt(r) {
|
|
55
|
-
const e = /* @__PURE__ */ new Map(),
|
|
56
|
-
if (!
|
|
55
|
+
const e = /* @__PURE__ */ new Map(), p = await r.file("manifest.json")?.async("string");
|
|
56
|
+
if (!p)
|
|
57
57
|
throw new Error("Manifest file not found in the zip archive");
|
|
58
|
-
const
|
|
59
|
-
for (const [t, o] of Object.entries(
|
|
58
|
+
const s = JSON.parse(p);
|
|
59
|
+
for (const [t, o] of Object.entries(s.weightSpec))
|
|
60
60
|
e.set(t, { spec: o, data: new Float32Array() });
|
|
61
|
-
const
|
|
62
|
-
if (!
|
|
61
|
+
const a = await r.file("tokeniser.json")?.async("string");
|
|
62
|
+
if (!a)
|
|
63
63
|
throw new Error("Tokeniser file not found in the zip archive");
|
|
64
|
-
const i = JSON.parse(
|
|
64
|
+
const i = JSON.parse(a), f = (i.type ?? "char") === "char" ? new h(i.vocab) : new k(i.vocab, i.merges), c = /* @__PURE__ */ new Map();
|
|
65
65
|
for (const t of Object.keys(r.files))
|
|
66
66
|
if (t.endsWith(".bin")) {
|
|
67
67
|
const o = t.replace(".bin", ""), l = await r.file(t).async("arraybuffer"), w = new Float32Array(l), n = e.get(o) || { spec: [], data: new Float32Array() };
|
|
@@ -70,7 +70,7 @@ async function Mt(r) {
|
|
|
70
70
|
c.set(o, d);
|
|
71
71
|
}
|
|
72
72
|
y();
|
|
73
|
-
const m = b(
|
|
73
|
+
const m = b(s.config);
|
|
74
74
|
return await g(m), m.loadWeights(c), { model: m, tokeniser: f };
|
|
75
75
|
}
|
|
76
76
|
export {
|
|
@@ -1,7 +1,7 @@
|
|
|
1
|
-
import {
|
|
2
|
-
import { e as _ } from "./axis_util-
|
|
3
|
-
import { r as m } from "./reshape-
|
|
4
|
-
import { s as T } from "./sum-
|
|
1
|
+
import { o as r, q as x, E as p, am as E, an as $, ao as d, af as h, c as S, l as K } from "./index-Duu1Lvvv.js";
|
|
2
|
+
import { e as _ } from "./axis_util-DGqbT-FX.js";
|
|
3
|
+
import { r as m } from "./reshape-BI0yzp1T.js";
|
|
4
|
+
import { s as T } from "./sum-BPUfDB2X.js";
|
|
5
5
|
function b(s, o = null, n = !1) {
|
|
6
6
|
const a = { x: x(s, "x", "max") }, e = { reductionIndices: o, keepDims: n };
|
|
7
7
|
return p.runKernel(E, a, e);
|
package/dist/main.d.ts
CHANGED
|
@@ -1,3 +1,6 @@
|
|
|
1
|
+
import { default as PretrainingTask } from './training/tasks/PretrainingTask';
|
|
2
|
+
import { default as StartSentenceTask } from './training/tasks/StartSentenceTask';
|
|
3
|
+
import { default as ConversationTask } from './training/tasks/ConversationTask';
|
|
1
4
|
import { pack16 } from './ops/pack16';
|
|
2
5
|
import { unpack16 } from './ops/unpack16';
|
|
3
6
|
import { default as CausalSelfAttention } from './layers/CausalSelfAttention';
|
|
@@ -19,6 +22,12 @@ export type { TrainingProgress, TrainingLogEntry } from './training/Trainer';
|
|
|
19
22
|
export type { GPTConfig } from './models/config';
|
|
20
23
|
export { estimateParameterCount, estimateMemoryUsage, estimateTrainingMemoryUsage, estimateResources, validateConfig, } from './utilities/parameters';
|
|
21
24
|
export { default as topP } from './utilities/topP';
|
|
25
|
+
export { Task, tokensFromTasks } from './training/tasks/Task';
|
|
26
|
+
export declare const tasks: {
|
|
27
|
+
PretrainingTask: typeof PretrainingTask;
|
|
28
|
+
StartSentenceTask: typeof StartSentenceTask;
|
|
29
|
+
ConversationTask: typeof ConversationTask;
|
|
30
|
+
};
|
|
22
31
|
declare const ops: {
|
|
23
32
|
pack16: typeof pack16;
|
|
24
33
|
unpack16: typeof unpack16;
|
package/dist/main.js
CHANGED
|
@@ -1,18 +1,22 @@
|
|
|
1
|
-
import "./index-
|
|
2
|
-
import "./random_width-
|
|
3
|
-
import "./zeros_like-
|
|
4
|
-
import { default as
|
|
1
|
+
import "./index-Duu1Lvvv.js";
|
|
2
|
+
import "./random_width-BN4wGJaW.js";
|
|
3
|
+
import "./zeros_like-De4n1C3m.js";
|
|
4
|
+
import { default as io } from "./Generator.js";
|
|
5
5
|
import "./index-Cp39cXWe.js";
|
|
6
|
-
import "./dataset-
|
|
7
|
-
import { default as
|
|
8
|
-
import { default as
|
|
9
|
-
import { default as
|
|
10
|
-
import { default as
|
|
11
|
-
import { default as
|
|
12
|
-
import { default as
|
|
13
|
-
import { default as
|
|
14
|
-
import { estimateMemoryUsage as
|
|
15
|
-
import { default as
|
|
6
|
+
import "./dataset-CVIJu7Xa.js";
|
|
7
|
+
import { default as fo } from "./models/NanoGPTV1.js";
|
|
8
|
+
import { default as lo } from "./TeachableLLM.js";
|
|
9
|
+
import { default as co } from "./tokeniser/CharTokeniser.js";
|
|
10
|
+
import { default as ko } from "./tokeniser/bpe.js";
|
|
11
|
+
import { default as go } from "./utilities/waitForModel.js";
|
|
12
|
+
import { default as Co } from "./data/textLoader.js";
|
|
13
|
+
import { default as Eo } from "./models/model.js";
|
|
14
|
+
import { estimateMemoryUsage as Bo, estimateParameterCount as bo, estimateResources as yo, estimateTrainingMemoryUsage as Lo, validateConfig as So } from "./utilities/parameters.js";
|
|
15
|
+
import { default as Ao } from "./utilities/topP.js";
|
|
16
|
+
import { Task as Go, tokensFromTasks as Ro } from "./training/tasks/Task.js";
|
|
17
|
+
import o from "./training/tasks/PretrainingTask.js";
|
|
18
|
+
import r from "./training/tasks/StartSentenceTask.js";
|
|
19
|
+
import t from "./training/tasks/ConversationTask.js";
|
|
16
20
|
import "./ops/cpu/scatterSub.js";
|
|
17
21
|
import "./ops/webgl/scatterSub.js";
|
|
18
22
|
import "./ops/cpu/gatherSub.js";
|
|
@@ -25,15 +29,15 @@ import "./ops/webgl/qkv.js";
|
|
|
25
29
|
import "./ops/grads/qkv.js";
|
|
26
30
|
import "./ops/cpu/rope.js";
|
|
27
31
|
import "./ops/webgl/rope.js";
|
|
28
|
-
import "./rope-
|
|
32
|
+
import "./rope-DJ7Y7c-u.js";
|
|
29
33
|
import "./ops/cpu/appendCache.js";
|
|
30
34
|
import "./ops/webgl/appendCache.js";
|
|
31
35
|
import "./ops/cpu/matMulGelu.js";
|
|
32
|
-
import "./matMulGelu-
|
|
36
|
+
import "./matMulGelu-BpvgnYG8.js";
|
|
33
37
|
import "./ops/grads/matMulGelu.js";
|
|
34
38
|
import "./ops/cpu/gelu.js";
|
|
35
39
|
import "./ops/webgl/gelu.js";
|
|
36
|
-
import "./gelu-
|
|
40
|
+
import "./gelu-BK1k-n1i.js";
|
|
37
41
|
import "./ops/cpu/normRMS.js";
|
|
38
42
|
import "./ops/webgl/normRMS.js";
|
|
39
43
|
import "./ops/grads/normRMS.js";
|
|
@@ -42,51 +46,58 @@ import "./ops/cpu/adamMoments.js";
|
|
|
42
46
|
import "./ops/webgl/adamMoments.js";
|
|
43
47
|
import "./ops/cpu/adamAdjust.js";
|
|
44
48
|
import "./ops/webgl/adamAdjust.js";
|
|
45
|
-
import { u as
|
|
49
|
+
import { u as e, p as m } from "./pack16-atD0eYRm.js";
|
|
46
50
|
import "./ops/grads/softmax16.js";
|
|
47
|
-
import "./matMul16-
|
|
51
|
+
import "./matMul16-xswmhSuF.js";
|
|
48
52
|
import "./ops/webgl/matMul16.js";
|
|
49
53
|
import "./ops/cpu/matMul16.js";
|
|
50
54
|
import "./ops/transpose16.js";
|
|
51
|
-
import { selectBackend as
|
|
52
|
-
import { default as
|
|
53
|
-
import
|
|
54
|
-
import
|
|
55
|
-
import
|
|
56
|
-
import
|
|
57
|
-
import { default as
|
|
58
|
-
import { default as
|
|
59
|
-
import { sentenceEmbeddings as
|
|
60
|
-
const
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
55
|
+
import { selectBackend as wo } from "./backend.js";
|
|
56
|
+
import { default as No } from "./utilities/performance.js";
|
|
57
|
+
import p from "./layers/CausalSelfAttention.js";
|
|
58
|
+
import a from "./layers/MLP.js";
|
|
59
|
+
import i from "./layers/TransformerBlock.js";
|
|
60
|
+
import s from "./layers/RoPECache.js";
|
|
61
|
+
import { default as qo } from "./training/AdamExt.js";
|
|
62
|
+
import { default as Ho } from "./checks/index.js";
|
|
63
|
+
import { sentenceEmbeddings as Jo, sentenceEmbeddingsTensor as Ko } from "./utilities/sentences.js";
|
|
64
|
+
const to = {
|
|
65
|
+
PretrainingTask: o,
|
|
66
|
+
StartSentenceTask: r,
|
|
67
|
+
ConversationTask: t
|
|
68
|
+
}, eo = {
|
|
69
|
+
pack16: m,
|
|
70
|
+
unpack16: e
|
|
71
|
+
}, mo = {
|
|
72
|
+
CausalSelfAttention: p,
|
|
73
|
+
MLP: a,
|
|
74
|
+
TransformerBlock: i,
|
|
75
|
+
RoPECache: s
|
|
68
76
|
};
|
|
69
77
|
export {
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
wo as
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
78
|
+
qo as AdamExt,
|
|
79
|
+
ko as BPETokeniser,
|
|
80
|
+
co as CharTokeniser,
|
|
81
|
+
io as Generator,
|
|
82
|
+
Eo as Model,
|
|
83
|
+
fo as NanoGPT,
|
|
84
|
+
Go as Task,
|
|
85
|
+
lo as TeachableLLM,
|
|
86
|
+
Ho as checks,
|
|
87
|
+
Bo as estimateMemoryUsage,
|
|
88
|
+
bo as estimateParameterCount,
|
|
89
|
+
yo as estimateResources,
|
|
90
|
+
Lo as estimateTrainingMemoryUsage,
|
|
91
|
+
mo as layers,
|
|
92
|
+
Co as loadTextData,
|
|
93
|
+
eo as ops,
|
|
94
|
+
No as performanceTest,
|
|
95
|
+
wo as selectBackend,
|
|
96
|
+
Jo as sentenceEmbeddings,
|
|
97
|
+
Ko as sentenceEmbeddingsTensor,
|
|
98
|
+
to as tasks,
|
|
99
|
+
Ro as tokensFromTasks,
|
|
100
|
+
Ao as topP,
|
|
101
|
+
So as validateConfig,
|
|
102
|
+
go as waitForModel
|
|
92
103
|
};
|
|
@@ -1,9 +1,9 @@
|
|
|
1
|
-
import {
|
|
1
|
+
import { i as y, e as h } from "./index-Duu1Lvvv.js";
|
|
2
2
|
import "./ops/webgl/matMul16.js";
|
|
3
3
|
import "./ops/cpu/matMul16.js";
|
|
4
4
|
import { isPackedTensor as g } from "./utilities/packed.js";
|
|
5
|
-
import { p as v } from "./pack16-
|
|
6
|
-
import { d as k } from "./gelu-
|
|
5
|
+
import { p as v } from "./pack16-atD0eYRm.js";
|
|
6
|
+
import { d as k } from "./gelu-BK1k-n1i.js";
|
|
7
7
|
import { transpose16 as S } from "./ops/transpose16.js";
|
|
8
8
|
import { reshape16 as w } from "./ops/reshape16.js";
|
|
9
9
|
const G = {
|
|
@@ -54,7 +54,7 @@ function l(r, o, n = !1, s = !1, t = {}) {
|
|
|
54
54
|
const e = g(r), f = g(o), i = e || f, a = !i || e ? r : v(r), p = !i || f ? o : v(o), c = h().runKernel("MatMul16", { A: a, B: p }, { transposeA: n, transposeB: s, ...t });
|
|
55
55
|
return i && !e && a.dispose(), i && !f && p.dispose(), c;
|
|
56
56
|
}
|
|
57
|
-
function
|
|
57
|
+
function K(r, o, n, s = !1, t = !1) {
|
|
58
58
|
return l(r, o, s, t, { scale: n });
|
|
59
59
|
}
|
|
60
60
|
function B(r, o, n, s = !1, t = !1) {
|
|
@@ -63,14 +63,14 @@ function B(r, o, n, s = !1, t = !1) {
|
|
|
63
63
|
function M(r, o, n, s = !1, t = !1) {
|
|
64
64
|
return l(r, o, s, t, { scaleB: n });
|
|
65
65
|
}
|
|
66
|
-
function
|
|
66
|
+
function N(r, o, n = !1, s = !1) {
|
|
67
67
|
return l(r, o, n, s, { activation: "gelu" });
|
|
68
68
|
}
|
|
69
69
|
export {
|
|
70
70
|
G as a,
|
|
71
71
|
l as b,
|
|
72
|
-
|
|
72
|
+
N as c,
|
|
73
73
|
B as d,
|
|
74
74
|
M as e,
|
|
75
|
-
|
|
75
|
+
K as m
|
|
76
76
|
};
|
|
@@ -1,12 +1,12 @@
|
|
|
1
|
-
import {
|
|
2
|
-
import { r as M } from "./Reshape-
|
|
3
|
-
import { u as H } from "./gpgpu_math-
|
|
4
|
-
import { m as B } from "./mat_mul-
|
|
1
|
+
import { f as C, t as R, e as I, U as G, _ as L, x as U, ak as F } from "./index-Duu1Lvvv.js";
|
|
2
|
+
import { r as M } from "./Reshape-Ck29jQSY.js";
|
|
3
|
+
import { u as H } from "./gpgpu_math-BJSTk_mW.js";
|
|
4
|
+
import { m as B } from "./mat_mul-Bn2BDpT4.js";
|
|
5
5
|
class W {
|
|
6
6
|
constructor(e, s, a, n = !1, o = !1, r = !1, i = null, u = !1, l = !1) {
|
|
7
7
|
this.variableNames = ["matrixA", "matrixB"], this.packedInputs = !0, this.packedOutput = !0, this.outputShape = a, this.enableShapeUniforms = H(this.outputShape.length);
|
|
8
|
-
const p = n ? e[1] : e[2], h = Math.ceil(p / 2), d = n ? "i * 2, rc.y" : "rc.y, i * 2",
|
|
9
|
-
let c = "",
|
|
8
|
+
const p = n ? e[1] : e[2], h = Math.ceil(p / 2), d = n ? "i * 2, rc.y" : "rc.y, i * 2", b = o ? "rc.z, i * 2" : "i * 2, rc.z", x = n ? ["a.xxyy", "a.zzww"] : ["a.xxzz", "a.yyww"], m = o ? ["b.xzxz", "b.ywyw"] : ["b.xyxy", "b.zwzw"];
|
|
9
|
+
let c = "", g = "";
|
|
10
10
|
i && (u ? c = `vec4 activation(vec4 a) {
|
|
11
11
|
vec4 b = getPreluActivationWeightsAtOutCoords();
|
|
12
12
|
${i}
|
|
@@ -15,8 +15,8 @@ class W {
|
|
|
15
15
|
${i}
|
|
16
16
|
}` : c = `vec4 activation(vec4 x) {
|
|
17
17
|
${i}
|
|
18
|
-
}`,
|
|
19
|
-
const
|
|
18
|
+
}`, g = "result = activation(result);");
|
|
19
|
+
const $ = r ? "result += getBiasAtOutCoords();" : "";
|
|
20
20
|
r && this.variableNames.push("bias"), u && this.variableNames.push("preluActivationWeights"), l && this.variableNames.push("leakyreluAlpha");
|
|
21
21
|
let f = "rc.x", v = "rc.x";
|
|
22
22
|
e[0] < s[0] ? f = `imod(rc.x, ${e[0]})` : s[0] < e[0] && (v = `imod(rc.x, ${s[0]})`), this.userCode = `
|
|
@@ -30,12 +30,12 @@ class W {
|
|
|
30
30
|
int batchB = ${v};
|
|
31
31
|
for (int i = 0; i < ${h}; i++) {
|
|
32
32
|
vec4 a = getMatrixA(batchA, ${d});
|
|
33
|
-
vec4 b = getMatrixB(batchB, ${
|
|
33
|
+
vec4 b = getMatrixB(batchB, ${b});
|
|
34
34
|
|
|
35
35
|
// These swizzled products need to be separately added.
|
|
36
36
|
// See: https://github.com/tensorflow/tfjs/issues/1735
|
|
37
|
-
result += (${
|
|
38
|
-
result += (${
|
|
37
|
+
result += (${x[0]} * ${m[0]});
|
|
38
|
+
result += (${x[1]} * ${m[1]});
|
|
39
39
|
}
|
|
40
40
|
return result;
|
|
41
41
|
}
|
|
@@ -44,10 +44,10 @@ class W {
|
|
|
44
44
|
ivec3 rc = getOutputCoords();
|
|
45
45
|
vec4 result = dot2x2ARowBCol(rc);
|
|
46
46
|
|
|
47
|
-
${g}
|
|
48
|
-
|
|
49
47
|
${$}
|
|
50
48
|
|
|
49
|
+
${g}
|
|
50
|
+
|
|
51
51
|
setOutput(result);
|
|
52
52
|
}
|
|
53
53
|
`;
|
|
@@ -90,30 +90,30 @@ function O({
|
|
|
90
90
|
activationSnippet: o,
|
|
91
91
|
multiplier: r
|
|
92
92
|
}) {
|
|
93
|
-
const i = t.shape.length, u = e.shape.length, l = s ? t.shape[i - 2] : t.shape[i - 1], p = a ? e.shape[u - 1] : e.shape[u - 2], h = s ? t.shape[i - 1] : t.shape[i - 2], d = a ? e.shape[u - 2] : e.shape[u - 1],
|
|
94
|
-
|
|
93
|
+
const i = t.shape.length, u = e.shape.length, l = s ? t.shape[i - 2] : t.shape[i - 1], p = a ? e.shape[u - 1] : e.shape[u - 2], h = s ? t.shape[i - 1] : t.shape[i - 2], d = a ? e.shape[u - 2] : e.shape[u - 1], b = t.shape.slice(0, -2), x = e.shape.slice(0, -2), m = G(b), c = G(x), $ = L(t.shape.slice(0, -2), e.shape.slice(0, -2)).concat([h, d]);
|
|
94
|
+
U(
|
|
95
95
|
l === p,
|
|
96
96
|
() => `Error in matMul: inner shapes (${l}) and (${p}) of Tensors with shapes ${t.shape} and ${e.shape} and transposeA=${s} and transposeB=${a} must match.`
|
|
97
97
|
);
|
|
98
|
-
const f = s ? [m, l, h] : [m, h, l], v = a ? [c, d, p] : [c, p, d], A = M({ inputs: { x: t }, backend: n, attrs: { shape: f } }), y = M({ inputs: { x: e }, backend: n, attrs: { shape: v } }), k = [A, y],
|
|
98
|
+
const f = s ? [m, l, h] : [m, h, l], v = a ? [c, d, p] : [c, p, d], A = M({ inputs: { x: t }, backend: n, attrs: { shape: f } }), y = M({ inputs: { x: e }, backend: n, attrs: { shape: v } }), k = [A, y], _ = Math.max(m, c), E = o, N = F(t.dtype, e.dtype), T = new W(
|
|
99
99
|
f,
|
|
100
100
|
v,
|
|
101
|
-
[
|
|
101
|
+
[_, h, d],
|
|
102
102
|
s,
|
|
103
103
|
a,
|
|
104
104
|
!1,
|
|
105
|
-
|
|
105
|
+
E,
|
|
106
106
|
!!r,
|
|
107
107
|
!1
|
|
108
108
|
), D = [A, y];
|
|
109
109
|
r && D.push(r);
|
|
110
|
-
const z = n.runWebGLProgram(
|
|
110
|
+
const z = n.runWebGLProgram(T, D, N), K = M({ inputs: { x: z }, backend: n, attrs: { shape: $ } });
|
|
111
111
|
k.push(z);
|
|
112
112
|
for (const P of k)
|
|
113
113
|
n.disposeIntermediateTensorInfo(P);
|
|
114
114
|
return K;
|
|
115
115
|
}
|
|
116
|
-
function
|
|
116
|
+
function J(t) {
|
|
117
117
|
const { inputs: e, backend: s } = t, { x: a, kernel: n } = e;
|
|
118
118
|
if (a === void 0 || n === void 0)
|
|
119
119
|
throw new Error("BatchMatMul requires two input tensors.");
|
|
@@ -126,13 +126,13 @@ function V(t) {
|
|
|
126
126
|
activationSnippet: j
|
|
127
127
|
});
|
|
128
128
|
}
|
|
129
|
-
const
|
|
129
|
+
const Q = {
|
|
130
130
|
kernelName: "MatMulGelu",
|
|
131
131
|
backendName: "webgl",
|
|
132
|
-
kernelFunc:
|
|
132
|
+
kernelFunc: J
|
|
133
133
|
};
|
|
134
|
-
C(
|
|
135
|
-
function
|
|
134
|
+
C(Q);
|
|
135
|
+
function V(t) {
|
|
136
136
|
const { dy: e, x: s, kernel: a } = t.inputs, n = t.backend;
|
|
137
137
|
return R(() => {
|
|
138
138
|
const o = I().makeTensorFromTensorInfo(
|
|
@@ -152,12 +152,12 @@ function Q(t) {
|
|
|
152
152
|
const X = {
|
|
153
153
|
kernelName: "MatMulGeluGrad",
|
|
154
154
|
backendName: "webgl",
|
|
155
|
-
kernelFunc:
|
|
155
|
+
kernelFunc: V
|
|
156
156
|
};
|
|
157
157
|
C(X);
|
|
158
158
|
export {
|
|
159
159
|
W as M,
|
|
160
160
|
se as a,
|
|
161
161
|
O as b,
|
|
162
|
-
|
|
162
|
+
J as c
|
|
163
163
|
};
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
import { o as m, q as s, B as c, E as M, D as p } from "./index-Duu1Lvvv.js";
|
|
2
|
+
function f(e, o, n = !1, l = !1) {
|
|
3
|
+
let a = s(e, "a", "matMul"), t = s(o, "b", "matMul");
|
|
4
|
+
[a, t] = c(a, t);
|
|
5
|
+
const r = { a, b: t }, u = { transposeA: n, transposeB: l };
|
|
6
|
+
return M.runKernel(p, r, u);
|
|
7
|
+
}
|
|
8
|
+
const i = /* @__PURE__ */ m({ matMul_: f });
|
|
9
|
+
export {
|
|
10
|
+
i as m
|
|
11
|
+
};
|
package/dist/models/NanoGPTV1.js
CHANGED
|
@@ -3,11 +3,11 @@ import b from "../layers/TransformerBlock.js";
|
|
|
3
3
|
import k from "../layers/TiedEmbedding.js";
|
|
4
4
|
import w from "../layers/RoPECache.js";
|
|
5
5
|
import E from "../layers/RMSNorm.js";
|
|
6
|
-
import { t as l, k as u } from "../index-
|
|
6
|
+
import { t as l, k as u } from "../index-Duu1Lvvv.js";
|
|
7
7
|
import C from "./model.js";
|
|
8
8
|
import P from "../layers/PositionEmbedding.js";
|
|
9
9
|
import { packingSupported as _ } from "../utilities/packed.js";
|
|
10
|
-
import { p as y, u as M } from "../pack16-
|
|
10
|
+
import { p as y, u as M } from "../pack16-atD0eYRm.js";
|
|
11
11
|
class I extends C {
|
|
12
12
|
wte;
|
|
13
13
|
// Token embeddings
|
package/dist/models/model.js
CHANGED
|
@@ -1,23 +1,23 @@
|
|
|
1
1
|
import m from "../layers/BaseLayer.js";
|
|
2
|
-
import "../index-
|
|
3
|
-
import "../random_width-
|
|
4
|
-
import "../zeros_like-
|
|
2
|
+
import "../index-Duu1Lvvv.js";
|
|
3
|
+
import "../random_width-BN4wGJaW.js";
|
|
4
|
+
import "../zeros_like-De4n1C3m.js";
|
|
5
5
|
import "../Generator.js";
|
|
6
6
|
import "../index-Cp39cXWe.js";
|
|
7
|
-
import "../dataset-
|
|
7
|
+
import "../dataset-CVIJu7Xa.js";
|
|
8
8
|
import "../ops/cpu/attentionMask.js";
|
|
9
9
|
import "../ops/webgl/attentionMask.js";
|
|
10
10
|
import "../ops/grads/attentionMask.js";
|
|
11
11
|
import "../ops/cpu/rope.js";
|
|
12
12
|
import "../ops/webgl/rope.js";
|
|
13
|
-
import "../rope-
|
|
13
|
+
import "../rope-DJ7Y7c-u.js";
|
|
14
14
|
import "../ops/cpu/appendCache.js";
|
|
15
15
|
import "../ops/webgl/appendCache.js";
|
|
16
16
|
import "../ops/grads/softmax16.js";
|
|
17
|
-
import "../matMul16-
|
|
17
|
+
import "../matMul16-xswmhSuF.js";
|
|
18
18
|
import "../ops/webgl/matMul16.js";
|
|
19
19
|
import "../ops/cpu/matMul16.js";
|
|
20
|
-
import "../pack16-
|
|
20
|
+
import "../pack16-atD0eYRm.js";
|
|
21
21
|
import "../ops/transpose16.js";
|
|
22
22
|
import "../ops/reshape16.js";
|
|
23
23
|
import "../ops/cpu/qkv.js";
|
|
@@ -40,11 +40,11 @@ import "../ops/webgl/scatterSub.js";
|
|
|
40
40
|
import "../ops/cpu/gatherSub.js";
|
|
41
41
|
import "../ops/webgl/gatherSub.js";
|
|
42
42
|
import "../ops/cpu/matMulGelu.js";
|
|
43
|
-
import "../matMulGelu-
|
|
43
|
+
import "../matMulGelu-BpvgnYG8.js";
|
|
44
44
|
import "../ops/grads/matMulGelu.js";
|
|
45
45
|
import "../ops/cpu/gelu.js";
|
|
46
46
|
import "../ops/webgl/gelu.js";
|
|
47
|
-
import "../gelu-
|
|
47
|
+
import "../gelu-BK1k-n1i.js";
|
|
48
48
|
import "../ops/webgl/log.js";
|
|
49
49
|
import "../checks/normRMS.js";
|
|
50
50
|
import "../checks/normRMSGrad.js";
|
|
@@ -1,6 +1,6 @@
|
|
|
1
|
-
import {
|
|
2
|
-
import { c as f } from "./complex-
|
|
3
|
-
import { z as c } from "./zeros-
|
|
1
|
+
import { u as n, V as t, U as m, E as i } from "./index-Duu1Lvvv.js";
|
|
2
|
+
import { c as f } from "./complex-Cyg-eQeZ.js";
|
|
3
|
+
import { z as c } from "./zeros-5YROwwUH.js";
|
|
4
4
|
function l(o, r = "float32") {
|
|
5
5
|
if (n(o), r === "complex64") {
|
|
6
6
|
const s = l(o, "float32"), a = c(o, "float32");
|
package/dist/ops/adamAdjust.js
CHANGED
package/dist/ops/adamMoments.js
CHANGED
package/dist/ops/add16.js
CHANGED