@fugood/llama.node 0.6.3 → 1.0.0-beta.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CMakeLists.txt +40 -30
- package/README.md +4 -1
- package/lib/binding.js +41 -29
- package/lib/binding.ts +26 -25
- package/package.json +40 -7
- package/scripts/build.js +47 -0
- package/scripts/llama.cpp.patch +109 -0
- package/src/anyascii.c +22223 -0
- package/src/anyascii.h +42 -0
- package/src/tts_utils.cpp +20 -7
- package/src/tts_utils.h +2 -0
- package/bin/darwin/arm64/llama-node.node +0 -0
- package/bin/darwin/x64/llama-node.node +0 -0
- package/bin/linux/arm64/llama-node.node +0 -0
- package/bin/linux/x64/llama-node.node +0 -0
- package/bin/linux-cuda/arm64/llama-node.node +0 -0
- package/bin/linux-cuda/x64/llama-node.node +0 -0
- package/bin/linux-vulkan/arm64/llama-node.node +0 -0
- package/bin/linux-vulkan/x64/llama-node.node +0 -0
- package/bin/win32/x64/llama-node.node +0 -0
- package/bin/win32/x64/node.lib +0 -0
- package/bin/win32-vulkan/arm64/llama-node.node +0 -0
- package/bin/win32-vulkan/arm64/node.lib +0 -0
- package/bin/win32-vulkan/x64/llama-node.node +0 -0
- package/bin/win32-vulkan/x64/node.lib +0 -0
- package/src/llama.cpp/.github/workflows/build-linux-cross.yml +0 -233
- package/src/llama.cpp/.github/workflows/build.yml +0 -1078
- package/src/llama.cpp/.github/workflows/close-issue.yml +0 -28
- package/src/llama.cpp/.github/workflows/docker.yml +0 -178
- package/src/llama.cpp/.github/workflows/editorconfig.yml +0 -29
- package/src/llama.cpp/.github/workflows/gguf-publish.yml +0 -44
- package/src/llama.cpp/.github/workflows/labeler.yml +0 -17
- package/src/llama.cpp/.github/workflows/python-check-requirements.yml +0 -33
- package/src/llama.cpp/.github/workflows/python-lint.yml +0 -30
- package/src/llama.cpp/.github/workflows/python-type-check.yml +0 -40
- package/src/llama.cpp/.github/workflows/release.yml +0 -739
- package/src/llama.cpp/.github/workflows/server.yml +0 -237
- package/src/llama.cpp/.github/workflows/winget.yml +0 -42
- package/src/llama.cpp/cmake/arm64-apple-clang.cmake +0 -16
- package/src/llama.cpp/cmake/arm64-windows-llvm.cmake +0 -16
- package/src/llama.cpp/cmake/build-info.cmake +0 -64
- package/src/llama.cpp/cmake/common.cmake +0 -35
- package/src/llama.cpp/cmake/git-vars.cmake +0 -22
- package/src/llama.cpp/cmake/x64-windows-llvm.cmake +0 -5
- package/src/llama.cpp/common/build-info.cpp.in +0 -4
- package/src/llama.cpp/docs/build.md +0 -561
- package/src/llama.cpp/examples/CMakeLists.txt +0 -43
- package/src/llama.cpp/examples/batched/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/batched/batched.cpp +0 -246
- package/src/llama.cpp/examples/chat-13B.bat +0 -57
- package/src/llama.cpp/examples/convert-llama2c-to-ggml/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/convert-llama2c-to-ggml/convert-llama2c-to-ggml.cpp +0 -941
- package/src/llama.cpp/examples/deprecation-warning/deprecation-warning.cpp +0 -35
- package/src/llama.cpp/examples/embedding/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/embedding/embedding.cpp +0 -323
- package/src/llama.cpp/examples/eval-callback/CMakeLists.txt +0 -10
- package/src/llama.cpp/examples/eval-callback/eval-callback.cpp +0 -194
- package/src/llama.cpp/examples/gen-docs/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/gen-docs/gen-docs.cpp +0 -83
- package/src/llama.cpp/examples/gguf/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/gguf/gguf.cpp +0 -265
- package/src/llama.cpp/examples/gguf-hash/CMakeLists.txt +0 -22
- package/src/llama.cpp/examples/gguf-hash/deps/rotate-bits/rotate-bits.h +0 -46
- package/src/llama.cpp/examples/gguf-hash/deps/sha1/sha1.c +0 -295
- package/src/llama.cpp/examples/gguf-hash/deps/sha1/sha1.h +0 -52
- package/src/llama.cpp/examples/gguf-hash/deps/sha256/sha256.c +0 -221
- package/src/llama.cpp/examples/gguf-hash/deps/sha256/sha256.h +0 -24
- package/src/llama.cpp/examples/gguf-hash/deps/xxhash/xxhash.c +0 -42
- package/src/llama.cpp/examples/gguf-hash/deps/xxhash/xxhash.h +0 -7093
- package/src/llama.cpp/examples/gguf-hash/gguf-hash.cpp +0 -694
- package/src/llama.cpp/examples/gritlm/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/gritlm/gritlm.cpp +0 -229
- package/src/llama.cpp/examples/jeopardy/questions.txt +0 -100
- package/src/llama.cpp/examples/llama.android/app/build.gradle.kts +0 -65
- package/src/llama.cpp/examples/llama.android/build.gradle.kts +0 -6
- package/src/llama.cpp/examples/llama.android/llama/build.gradle.kts +0 -71
- package/src/llama.cpp/examples/llama.android/llama/src/main/cpp/CMakeLists.txt +0 -53
- package/src/llama.cpp/examples/llama.android/llama/src/main/cpp/llama-android.cpp +0 -452
- package/src/llama.cpp/examples/llama.android/settings.gradle.kts +0 -18
- package/src/llama.cpp/examples/lookahead/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/lookahead/lookahead.cpp +0 -472
- package/src/llama.cpp/examples/lookup/CMakeLists.txt +0 -23
- package/src/llama.cpp/examples/lookup/lookup-create.cpp +0 -40
- package/src/llama.cpp/examples/lookup/lookup-merge.cpp +0 -47
- package/src/llama.cpp/examples/lookup/lookup-stats.cpp +0 -157
- package/src/llama.cpp/examples/lookup/lookup.cpp +0 -242
- package/src/llama.cpp/examples/parallel/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/parallel/parallel.cpp +0 -492
- package/src/llama.cpp/examples/passkey/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/passkey/passkey.cpp +0 -277
- package/src/llama.cpp/examples/retrieval/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/retrieval/retrieval.cpp +0 -304
- package/src/llama.cpp/examples/save-load-state/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/save-load-state/save-load-state.cpp +0 -246
- package/src/llama.cpp/examples/simple/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/simple/simple.cpp +0 -206
- package/src/llama.cpp/examples/simple-chat/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/simple-chat/simple-chat.cpp +0 -206
- package/src/llama.cpp/examples/simple-cmake-pkg/CMakeLists.txt +0 -11
- package/src/llama.cpp/examples/speculative/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/speculative/speculative.cpp +0 -644
- package/src/llama.cpp/examples/speculative-simple/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/speculative-simple/speculative-simple.cpp +0 -261
- package/src/llama.cpp/examples/sycl/CMakeLists.txt +0 -9
- package/src/llama.cpp/examples/sycl/build.sh +0 -23
- package/src/llama.cpp/examples/sycl/ls-sycl-device.cpp +0 -13
- package/src/llama.cpp/examples/sycl/run-llama2.sh +0 -27
- package/src/llama.cpp/examples/sycl/run-llama3.sh +0 -28
- package/src/llama.cpp/examples/sycl/win-build-sycl.bat +0 -33
- package/src/llama.cpp/examples/sycl/win-run-llama2.bat +0 -9
- package/src/llama.cpp/examples/sycl/win-run-llama3.bat +0 -9
- package/src/llama.cpp/examples/training/CMakeLists.txt +0 -5
- package/src/llama.cpp/examples/training/finetune.cpp +0 -96
- package/src/llama.cpp/ggml/cmake/GitVars.cmake +0 -22
- package/src/llama.cpp/ggml/cmake/common.cmake +0 -26
- package/src/llama.cpp/ggml/src/ggml-alloc.c +0 -1042
- package/src/llama.cpp/ggml/src/ggml-backend-impl.h +0 -255
- package/src/llama.cpp/ggml/src/ggml-backend-reg.cpp +0 -586
- package/src/llama.cpp/ggml/src/ggml-backend.cpp +0 -2008
- package/src/llama.cpp/ggml/src/ggml-blas/CMakeLists.txt +0 -87
- package/src/llama.cpp/ggml/src/ggml-blas/ggml-blas.cpp +0 -517
- package/src/llama.cpp/ggml/src/ggml-cann/CMakeLists.txt +0 -74
- package/src/llama.cpp/ggml/src/ggml-cann/acl_tensor.cpp +0 -179
- package/src/llama.cpp/ggml/src/ggml-cann/acl_tensor.h +0 -258
- package/src/llama.cpp/ggml/src/ggml-cann/aclnn_ops.cpp +0 -2863
- package/src/llama.cpp/ggml/src/ggml-cann/aclnn_ops.h +0 -1110
- package/src/llama.cpp/ggml/src/ggml-cann/common.h +0 -420
- package/src/llama.cpp/ggml/src/ggml-cann/ggml-cann.cpp +0 -2570
- package/src/llama.cpp/ggml/src/ggml-common.h +0 -1857
- package/src/llama.cpp/ggml/src/ggml-cpu/cmake/FindSIMD.cmake +0 -100
- package/src/llama.cpp/ggml/src/ggml-cuda/CMakeLists.txt +0 -184
- package/src/llama.cpp/ggml/src/ggml-cuda/vendors/cuda.h +0 -15
- package/src/llama.cpp/ggml/src/ggml-cuda/vendors/hip.h +0 -243
- package/src/llama.cpp/ggml/src/ggml-cuda/vendors/musa.h +0 -140
- package/src/llama.cpp/ggml/src/ggml-hip/CMakeLists.txt +0 -131
- package/src/llama.cpp/ggml/src/ggml-impl.h +0 -601
- package/src/llama.cpp/ggml/src/ggml-kompute/CMakeLists.txt +0 -166
- package/src/llama.cpp/ggml/src/ggml-kompute/ggml-kompute.cpp +0 -2251
- package/src/llama.cpp/ggml/src/ggml-metal/CMakeLists.txt +0 -120
- package/src/llama.cpp/ggml/src/ggml-metal/ggml-metal-impl.h +0 -622
- package/src/llama.cpp/ggml/src/ggml-musa/CMakeLists.txt +0 -113
- package/src/llama.cpp/ggml/src/ggml-opencl/CMakeLists.txt +0 -96
- package/src/llama.cpp/ggml/src/ggml-opencl/ggml-opencl.cpp +0 -5124
- package/src/llama.cpp/ggml/src/ggml-opt.cpp +0 -1037
- package/src/llama.cpp/ggml/src/ggml-quants.c +0 -5232
- package/src/llama.cpp/ggml/src/ggml-quants.h +0 -100
- package/src/llama.cpp/ggml/src/ggml-rpc/CMakeLists.txt +0 -9
- package/src/llama.cpp/ggml/src/ggml-rpc/ggml-rpc.cpp +0 -1813
- package/src/llama.cpp/ggml/src/ggml-sycl/CMakeLists.txt +0 -189
- package/src/llama.cpp/ggml/src/ggml-sycl/backend.hpp +0 -37
- package/src/llama.cpp/ggml/src/ggml-sycl/binbcast.cpp +0 -239
- package/src/llama.cpp/ggml/src/ggml-sycl/binbcast.hpp +0 -39
- package/src/llama.cpp/ggml/src/ggml-sycl/common.cpp +0 -83
- package/src/llama.cpp/ggml/src/ggml-sycl/common.hpp +0 -493
- package/src/llama.cpp/ggml/src/ggml-sycl/concat.cpp +0 -197
- package/src/llama.cpp/ggml/src/ggml-sycl/concat.hpp +0 -20
- package/src/llama.cpp/ggml/src/ggml-sycl/conv.cpp +0 -100
- package/src/llama.cpp/ggml/src/ggml-sycl/conv.hpp +0 -20
- package/src/llama.cpp/ggml/src/ggml-sycl/convert.cpp +0 -623
- package/src/llama.cpp/ggml/src/ggml-sycl/convert.hpp +0 -34
- package/src/llama.cpp/ggml/src/ggml-sycl/cpy.cpp +0 -701
- package/src/llama.cpp/ggml/src/ggml-sycl/cpy.hpp +0 -11
- package/src/llama.cpp/ggml/src/ggml-sycl/dequantize.hpp +0 -791
- package/src/llama.cpp/ggml/src/ggml-sycl/dmmv.cpp +0 -1160
- package/src/llama.cpp/ggml/src/ggml-sycl/dmmv.hpp +0 -27
- package/src/llama.cpp/ggml/src/ggml-sycl/dpct/helper.hpp +0 -2957
- package/src/llama.cpp/ggml/src/ggml-sycl/element_wise.cpp +0 -1536
- package/src/llama.cpp/ggml/src/ggml-sycl/element_wise.hpp +0 -75
- package/src/llama.cpp/ggml/src/ggml-sycl/gemm.hpp +0 -99
- package/src/llama.cpp/ggml/src/ggml-sycl/getrows.cpp +0 -311
- package/src/llama.cpp/ggml/src/ggml-sycl/getrows.hpp +0 -20
- package/src/llama.cpp/ggml/src/ggml-sycl/ggml-sycl.cpp +0 -4443
- package/src/llama.cpp/ggml/src/ggml-sycl/gla.cpp +0 -105
- package/src/llama.cpp/ggml/src/ggml-sycl/gla.hpp +0 -8
- package/src/llama.cpp/ggml/src/ggml-sycl/im2col.cpp +0 -136
- package/src/llama.cpp/ggml/src/ggml-sycl/im2col.hpp +0 -21
- package/src/llama.cpp/ggml/src/ggml-sycl/mmq.cpp +0 -3030
- package/src/llama.cpp/ggml/src/ggml-sycl/mmq.hpp +0 -33
- package/src/llama.cpp/ggml/src/ggml-sycl/mmvq.cpp +0 -1108
- package/src/llama.cpp/ggml/src/ggml-sycl/mmvq.hpp +0 -27
- package/src/llama.cpp/ggml/src/ggml-sycl/norm.cpp +0 -474
- package/src/llama.cpp/ggml/src/ggml-sycl/norm.hpp +0 -26
- package/src/llama.cpp/ggml/src/ggml-sycl/outprod.cpp +0 -46
- package/src/llama.cpp/ggml/src/ggml-sycl/outprod.hpp +0 -10
- package/src/llama.cpp/ggml/src/ggml-sycl/presets.hpp +0 -74
- package/src/llama.cpp/ggml/src/ggml-sycl/quants.hpp +0 -83
- package/src/llama.cpp/ggml/src/ggml-sycl/rope.cpp +0 -362
- package/src/llama.cpp/ggml/src/ggml-sycl/rope.hpp +0 -20
- package/src/llama.cpp/ggml/src/ggml-sycl/softmax.cpp +0 -264
- package/src/llama.cpp/ggml/src/ggml-sycl/softmax.hpp +0 -20
- package/src/llama.cpp/ggml/src/ggml-sycl/sycl_hw.cpp +0 -13
- package/src/llama.cpp/ggml/src/ggml-sycl/sycl_hw.hpp +0 -23
- package/src/llama.cpp/ggml/src/ggml-sycl/tsembd.cpp +0 -73
- package/src/llama.cpp/ggml/src/ggml-sycl/tsembd.hpp +0 -20
- package/src/llama.cpp/ggml/src/ggml-sycl/vecdotq.hpp +0 -1215
- package/src/llama.cpp/ggml/src/ggml-sycl/wkv.cpp +0 -305
- package/src/llama.cpp/ggml/src/ggml-sycl/wkv.hpp +0 -10
- package/src/llama.cpp/ggml/src/ggml-threading.cpp +0 -12
- package/src/llama.cpp/ggml/src/ggml-threading.h +0 -14
- package/src/llama.cpp/ggml/src/ggml-vulkan/CMakeLists.txt +0 -196
- package/src/llama.cpp/ggml/src/ggml-vulkan/ggml-vulkan.cpp +0 -10699
- package/src/llama.cpp/ggml/src/ggml-vulkan/vulkan-shaders/CMakeLists.txt +0 -39
- package/src/llama.cpp/ggml/src/ggml-vulkan/vulkan-shaders/vulkan-shaders-gen.cpp +0 -751
- package/src/llama.cpp/ggml/src/ggml.c +0 -6550
- package/src/llama.cpp/ggml/src/gguf.cpp +0 -1330
- package/src/llama.cpp/models/.editorconfig +0 -1
- package/src/llama.cpp/models/ggml-vocab-aquila.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-baichuan.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-bert-bge.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-bert-bge.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-bert-bge.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-chameleon.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-chameleon.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-command-r.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-command-r.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-command-r.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-deepseek-coder.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-deepseek-coder.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-deepseek-coder.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-deepseek-llm.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-deepseek-llm.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-deepseek-llm.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-deepseek-r1-qwen.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-deepseek-r1-qwen.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-falcon.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-falcon.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-falcon.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-gpt-2.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-gpt-2.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-gpt-2.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-gpt-4o.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-gpt-4o.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-gpt-neox.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-llama-bpe.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-llama-bpe.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-llama-bpe.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-llama-spm.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-llama-spm.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-llama-spm.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-llama4.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-llama4.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-mpt.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-mpt.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-mpt.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-phi-3.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-phi-3.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-phi-3.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-pixtral.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-pixtral.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-qwen2.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-qwen2.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-qwen2.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-refact.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-refact.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-refact.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-roberta-bpe.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-roberta-bpe.gguf.out +0 -46
- package/src/llama.cpp/models/ggml-vocab-starcoder.gguf +0 -0
- package/src/llama.cpp/models/ggml-vocab-starcoder.gguf.inp +0 -112
- package/src/llama.cpp/models/ggml-vocab-starcoder.gguf.out +0 -46
- package/src/llama.cpp/pocs/CMakeLists.txt +0 -14
- package/src/llama.cpp/pocs/vdot/CMakeLists.txt +0 -9
- package/src/llama.cpp/pocs/vdot/q8dot.cpp +0 -173
- package/src/llama.cpp/pocs/vdot/vdot.cpp +0 -311
- package/src/llama.cpp/prompts/LLM-questions.txt +0 -49
- package/src/llama.cpp/prompts/alpaca.txt +0 -1
- package/src/llama.cpp/prompts/assistant.txt +0 -31
- package/src/llama.cpp/prompts/chat-with-baichuan.txt +0 -4
- package/src/llama.cpp/prompts/chat-with-bob.txt +0 -7
- package/src/llama.cpp/prompts/chat-with-qwen.txt +0 -1
- package/src/llama.cpp/prompts/chat-with-vicuna-v0.txt +0 -7
- package/src/llama.cpp/prompts/chat-with-vicuna-v1.txt +0 -7
- package/src/llama.cpp/prompts/chat.txt +0 -28
- package/src/llama.cpp/prompts/dan-modified.txt +0 -1
- package/src/llama.cpp/prompts/dan.txt +0 -1
- package/src/llama.cpp/prompts/mnemonics.txt +0 -93
- package/src/llama.cpp/prompts/parallel-questions.txt +0 -43
- package/src/llama.cpp/prompts/reason-act.txt +0 -18
- package/src/llama.cpp/requirements/requirements-all.txt +0 -15
- package/src/llama.cpp/requirements/requirements-compare-llama-bench.txt +0 -2
- package/src/llama.cpp/requirements/requirements-convert_hf_to_gguf.txt +0 -7
- package/src/llama.cpp/requirements/requirements-convert_hf_to_gguf_update.txt +0 -7
- package/src/llama.cpp/requirements/requirements-convert_legacy_llama.txt +0 -5
- package/src/llama.cpp/requirements/requirements-convert_llama_ggml_to_gguf.txt +0 -1
- package/src/llama.cpp/requirements/requirements-convert_lora_to_gguf.txt +0 -4
- package/src/llama.cpp/requirements/requirements-gguf_editor_gui.txt +0 -3
- package/src/llama.cpp/requirements/requirements-pydantic.txt +0 -3
- package/src/llama.cpp/requirements/requirements-test-tokenizer-random.txt +0 -1
- package/src/llama.cpp/requirements/requirements-tool_bench.txt +0 -12
- package/src/llama.cpp/requirements.txt +0 -13
- package/src/llama.cpp/scripts/build-info.sh +0 -30
- package/src/llama.cpp/scripts/install-oneapi.bat +0 -19
- package/src/llama.cpp/scripts/xxd.cmake +0 -16
- package/src/llama.cpp/tests/CMakeLists.txt +0 -177
- package/src/llama.cpp/tests/get-model.cpp +0 -21
- package/src/llama.cpp/tests/get-model.h +0 -2
- package/src/llama.cpp/tests/test-arg-parser.cpp +0 -178
- package/src/llama.cpp/tests/test-autorelease.cpp +0 -24
- package/src/llama.cpp/tests/test-backend-ops.cpp +0 -4793
- package/src/llama.cpp/tests/test-barrier.cpp +0 -94
- package/src/llama.cpp/tests/test-c.c +0 -7
- package/src/llama.cpp/tests/test-chat-template.cpp +0 -417
- package/src/llama.cpp/tests/test-chat.cpp +0 -985
- package/src/llama.cpp/tests/test-double-float.cpp +0 -57
- package/src/llama.cpp/tests/test-gbnf-validator.cpp +0 -109
- package/src/llama.cpp/tests/test-gguf.cpp +0 -1338
- package/src/llama.cpp/tests/test-grammar-integration.cpp +0 -1308
- package/src/llama.cpp/tests/test-grammar-llguidance.cpp +0 -1201
- package/src/llama.cpp/tests/test-grammar-parser.cpp +0 -519
- package/src/llama.cpp/tests/test-json-schema-to-grammar.cpp +0 -1304
- package/src/llama.cpp/tests/test-llama-grammar.cpp +0 -408
- package/src/llama.cpp/tests/test-log.cpp +0 -39
- package/src/llama.cpp/tests/test-model-load-cancel.cpp +0 -27
- package/src/llama.cpp/tests/test-mtmd-c-api.c +0 -63
- package/src/llama.cpp/tests/test-opt.cpp +0 -904
- package/src/llama.cpp/tests/test-quantize-fns.cpp +0 -186
- package/src/llama.cpp/tests/test-quantize-perf.cpp +0 -365
- package/src/llama.cpp/tests/test-quantize-stats.cpp +0 -424
- package/src/llama.cpp/tests/test-regex-partial.cpp +0 -288
- package/src/llama.cpp/tests/test-rope.cpp +0 -262
- package/src/llama.cpp/tests/test-sampling.cpp +0 -399
- package/src/llama.cpp/tests/test-tokenizer-0.cpp +0 -312
- package/src/llama.cpp/tests/test-tokenizer-1-bpe.cpp +0 -155
- package/src/llama.cpp/tests/test-tokenizer-1-spm.cpp +0 -125
- package/src/llama.cpp/tools/CMakeLists.txt +0 -39
- package/src/llama.cpp/tools/batched-bench/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/batched-bench/batched-bench.cpp +0 -204
- package/src/llama.cpp/tools/cvector-generator/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/cvector-generator/completions.txt +0 -582
- package/src/llama.cpp/tools/cvector-generator/cvector-generator.cpp +0 -508
- package/src/llama.cpp/tools/cvector-generator/mean.hpp +0 -48
- package/src/llama.cpp/tools/cvector-generator/negative.txt +0 -4
- package/src/llama.cpp/tools/cvector-generator/pca.hpp +0 -315
- package/src/llama.cpp/tools/cvector-generator/positive.txt +0 -4
- package/src/llama.cpp/tools/export-lora/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/export-lora/export-lora.cpp +0 -434
- package/src/llama.cpp/tools/gguf-split/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/gguf-split/gguf-split.cpp +0 -583
- package/src/llama.cpp/tools/imatrix/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/imatrix/imatrix.cpp +0 -667
- package/src/llama.cpp/tools/llama-bench/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/llama-bench/llama-bench.cpp +0 -2024
- package/src/llama.cpp/tools/main/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/main/main.cpp +0 -977
- package/src/llama.cpp/tools/mtmd/CMakeLists.txt +0 -58
- package/src/llama.cpp/tools/mtmd/clip-impl.h +0 -462
- package/src/llama.cpp/tools/mtmd/clip.cpp +0 -4024
- package/src/llama.cpp/tools/mtmd/clip.h +0 -101
- package/src/llama.cpp/tools/mtmd/deprecation-warning.cpp +0 -22
- package/src/llama.cpp/tools/mtmd/miniaudio.h +0 -93468
- package/src/llama.cpp/tools/mtmd/mtmd-audio.cpp +0 -855
- package/src/llama.cpp/tools/mtmd/mtmd-audio.h +0 -62
- package/src/llama.cpp/tools/mtmd/mtmd-cli.cpp +0 -377
- package/src/llama.cpp/tools/mtmd/mtmd-helper.cpp +0 -297
- package/src/llama.cpp/tools/mtmd/mtmd.cpp +0 -942
- package/src/llama.cpp/tools/mtmd/mtmd.h +0 -362
- package/src/llama.cpp/tools/mtmd/requirements.txt +0 -5
- package/src/llama.cpp/tools/perplexity/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/perplexity/perplexity.cpp +0 -2063
- package/src/llama.cpp/tools/quantize/CMakeLists.txt +0 -6
- package/src/llama.cpp/tools/quantize/quantize.cpp +0 -519
- package/src/llama.cpp/tools/rpc/CMakeLists.txt +0 -4
- package/src/llama.cpp/tools/rpc/rpc-server.cpp +0 -322
- package/src/llama.cpp/tools/run/CMakeLists.txt +0 -16
- package/src/llama.cpp/tools/run/linenoise.cpp/linenoise.cpp +0 -1995
- package/src/llama.cpp/tools/run/linenoise.cpp/linenoise.h +0 -137
- package/src/llama.cpp/tools/run/run.cpp +0 -1261
- package/src/llama.cpp/tools/server/CMakeLists.txt +0 -51
- package/src/llama.cpp/tools/server/bench/requirements.txt +0 -2
- package/src/llama.cpp/tools/server/httplib.h +0 -10506
- package/src/llama.cpp/tools/server/server.cpp +0 -4966
- package/src/llama.cpp/tools/server/tests/requirements.txt +0 -8
- package/src/llama.cpp/tools/server/utils.hpp +0 -1337
- package/src/llama.cpp/tools/tokenize/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/tokenize/tokenize.cpp +0 -416
- package/src/llama.cpp/tools/tts/CMakeLists.txt +0 -5
- package/src/llama.cpp/tools/tts/tts.cpp +0 -1092
|
@@ -1,75 +0,0 @@
|
|
|
1
|
-
#ifndef GGML_SYCL_ELEMENTWISE_HPP
|
|
2
|
-
#define GGML_SYCL_ELEMENTWISE_HPP
|
|
3
|
-
|
|
4
|
-
#include "common.hpp"
|
|
5
|
-
#include "ggml.h"
|
|
6
|
-
#include <limits.h>
|
|
7
|
-
|
|
8
|
-
template <typename T>
|
|
9
|
-
T neg_infinity() {
|
|
10
|
-
return -std::numeric_limits<T>::infinity();
|
|
11
|
-
}
|
|
12
|
-
|
|
13
|
-
template<typename T>
|
|
14
|
-
struct typed_data {
|
|
15
|
-
const T * src;
|
|
16
|
-
T * dst;
|
|
17
|
-
};
|
|
18
|
-
|
|
19
|
-
template<typename T>
|
|
20
|
-
typed_data<T> cast_data(ggml_tensor * dst) {
|
|
21
|
-
return {
|
|
22
|
-
/* .src = */ static_cast<const T *>(dst->src[0]->data),
|
|
23
|
-
/* .dst = */ static_cast<T *>(dst->data)
|
|
24
|
-
};
|
|
25
|
-
}
|
|
26
|
-
|
|
27
|
-
void ggml_sycl_sqrt(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
28
|
-
|
|
29
|
-
void ggml_sycl_sin(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
30
|
-
|
|
31
|
-
void ggml_sycl_cos(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
32
|
-
|
|
33
|
-
void ggml_sycl_acc(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
34
|
-
|
|
35
|
-
void ggml_sycl_gelu(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
36
|
-
|
|
37
|
-
void ggml_sycl_silu(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
38
|
-
|
|
39
|
-
void ggml_sycl_gelu_quick(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
40
|
-
|
|
41
|
-
void ggml_sycl_tanh(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
42
|
-
|
|
43
|
-
void ggml_sycl_relu(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
44
|
-
|
|
45
|
-
void ggml_sycl_sigmoid(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
46
|
-
|
|
47
|
-
void ggml_sycl_hardsigmoid(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
48
|
-
|
|
49
|
-
void ggml_sycl_hardswish(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
50
|
-
|
|
51
|
-
void ggml_sycl_exp(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
52
|
-
|
|
53
|
-
void ggml_sycl_log(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
54
|
-
|
|
55
|
-
void ggml_sycl_neg(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
56
|
-
|
|
57
|
-
void ggml_sycl_step(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
58
|
-
|
|
59
|
-
void ggml_sycl_leaky_relu(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
60
|
-
|
|
61
|
-
void ggml_sycl_sqr(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
62
|
-
|
|
63
|
-
void ggml_sycl_upscale(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
64
|
-
|
|
65
|
-
void ggml_sycl_pad(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
66
|
-
|
|
67
|
-
void ggml_sycl_clamp(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
68
|
-
|
|
69
|
-
void ggml_sycl_sgn(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
70
|
-
|
|
71
|
-
void ggml_sycl_abs(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
72
|
-
|
|
73
|
-
void ggml_sycl_elu(ggml_backend_sycl_context & ctx, ggml_tensor * dst);
|
|
74
|
-
#endif // GGML_SYCL_ELEMENTWISE_HPP
|
|
75
|
-
|
|
@@ -1,99 +0,0 @@
|
|
|
1
|
-
//
|
|
2
|
-
// MIT license
|
|
3
|
-
// Copyright (C) 2024 Intel Corporation
|
|
4
|
-
// SPDX-License-Identifier: MIT
|
|
5
|
-
//
|
|
6
|
-
|
|
7
|
-
//
|
|
8
|
-
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
9
|
-
// See https://llvm.org/LICENSE.txt for license information.
|
|
10
|
-
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
11
|
-
//
|
|
12
|
-
|
|
13
|
-
#ifndef GGML_SYCL_GEMM_HPP
|
|
14
|
-
#define GGML_SYCL_GEMM_HPP
|
|
15
|
-
|
|
16
|
-
#include "ggml-sycl.h"
|
|
17
|
-
|
|
18
|
-
#if GGML_SYCL_DNNL
|
|
19
|
-
|
|
20
|
-
#include "dnnl.hpp"
|
|
21
|
-
#include "dnnl_sycl.hpp"
|
|
22
|
-
|
|
23
|
-
class DnnlGemmWrapper {
|
|
24
|
-
public:
|
|
25
|
-
using dt = dnnl::memory::data_type;
|
|
26
|
-
using tag = dnnl::memory::format_tag;
|
|
27
|
-
|
|
28
|
-
template<typename T>
|
|
29
|
-
static constexpr dt to_dt() {
|
|
30
|
-
if constexpr (std::is_same_v<T, float>) return dt::f32;
|
|
31
|
-
else if constexpr (std::is_same_v<T, sycl::half>) return dt::f16;
|
|
32
|
-
else static_assert(0);
|
|
33
|
-
}
|
|
34
|
-
|
|
35
|
-
// matrix A has m rows, k columns
|
|
36
|
-
// matrix B has k rows, n columns
|
|
37
|
-
// nra - number of elements to skip when moving into next row in A
|
|
38
|
-
// nrb - number of elements to skip when moving into next row in B
|
|
39
|
-
// nca - number of elements to skip when moving into next column in A
|
|
40
|
-
// ncb - number of elements to skip when moving into next column in B
|
|
41
|
-
// stride_a - number of elements to skip when moving to next A matrix
|
|
42
|
-
// stride_b - number of elements to skip when moving to next B matrix
|
|
43
|
-
// batches_a - number of A matrices
|
|
44
|
-
// batches_b - number of B matrices
|
|
45
|
-
static void gemm(ggml_backend_sycl_context & ctx, int m, int n, int k,
|
|
46
|
-
const void * a, dt at, dnnl_dim_t nra, dnnl_dim_t nca, dnnl_dim_t stride_a,
|
|
47
|
-
const void * b, dt bt, dnnl_dim_t nrb, dnnl_dim_t ncb, dnnl_dim_t stride_b,
|
|
48
|
-
void * c, dt ct, const queue_ptr & q, dnnl_dim_t batches_a, dnnl_dim_t batches_b) {
|
|
49
|
-
|
|
50
|
-
auto stream = ctx.stream_dnnl(q);
|
|
51
|
-
auto eng = ctx.engine_dnnl(q);
|
|
52
|
-
|
|
53
|
-
// { # strides, # rows, # columns }
|
|
54
|
-
dnnl::memory::dims a_dims = { batches_a, m, k };
|
|
55
|
-
dnnl::memory::dims b_dims = { batches_b, k, n };
|
|
56
|
-
dnnl::memory::dims c_dims = { std::max(batches_a, batches_b), m, n };
|
|
57
|
-
|
|
58
|
-
// { # elements to skip to next stride, # elements to skip to next row, # elements to skip to next column }
|
|
59
|
-
dnnl::memory::dims a_strides = { stride_a, nra, nca };
|
|
60
|
-
dnnl::memory::dims b_strides = { stride_b, nrb, ncb };
|
|
61
|
-
|
|
62
|
-
const auto a_in_md = dnnl::memory::desc(a_dims, at, a_strides);
|
|
63
|
-
const auto b_in_md = dnnl::memory::desc(b_dims, bt, b_strides);
|
|
64
|
-
const auto c_md = dnnl::memory::desc(c_dims, ct, tag::abc);
|
|
65
|
-
|
|
66
|
-
dnnl::primitive_attr primitive_attr;
|
|
67
|
-
primitive_attr.set_scratchpad_mode(dnnl::scratchpad_mode::user);
|
|
68
|
-
|
|
69
|
-
auto a_mem = dnnl::memory(a_in_md, eng, const_cast<void*>(a));
|
|
70
|
-
auto b_mem = dnnl::memory(b_in_md, eng, const_cast<void*>(b));
|
|
71
|
-
auto matmul_pd = dnnl::matmul::primitive_desc(eng, a_in_md, b_in_md, c_md, primitive_attr);
|
|
72
|
-
auto c_mem = dnnl::memory(matmul_pd.dst_desc(), eng, c);
|
|
73
|
-
|
|
74
|
-
auto scratchpad_md = matmul_pd.scratchpad_desc();
|
|
75
|
-
auto scratchpad_mem = ctx.get_scratchpad_mem(scratchpad_md, eng, q);
|
|
76
|
-
auto matmul_prim = dnnl::matmul(matmul_pd);
|
|
77
|
-
|
|
78
|
-
std::unordered_map<int, dnnl::memory> matmul_args;
|
|
79
|
-
matmul_args.insert({ DNNL_ARG_SRC, a_mem });
|
|
80
|
-
matmul_args.insert({ DNNL_ARG_WEIGHTS, b_mem });
|
|
81
|
-
matmul_args.insert({ DNNL_ARG_DST, c_mem });
|
|
82
|
-
matmul_args.insert({ DNNL_ARG_SCRATCHPAD, scratchpad_mem });
|
|
83
|
-
|
|
84
|
-
matmul_prim.execute(stream, matmul_args);
|
|
85
|
-
}
|
|
86
|
-
|
|
87
|
-
// matrices A and B are column major, both having k rows
|
|
88
|
-
// matrix A has m column, matrix B has n columns
|
|
89
|
-
// output: column major matrix C = A transposed * B
|
|
90
|
-
static void row_gemm(ggml_backend_sycl_context & ctx, int m, int n, int k,
|
|
91
|
-
const void * a, dt at, const void * b, dt bt, void * c, dt ct, const queue_ptr & q) {
|
|
92
|
-
|
|
93
|
-
gemm(ctx, m, n, k, a, at, k, 1, k * m, b, bt, 1, k, n * k, c, ct, q, 1, 1);
|
|
94
|
-
}
|
|
95
|
-
};
|
|
96
|
-
|
|
97
|
-
#endif
|
|
98
|
-
|
|
99
|
-
#endif // GGML_SYCL_GEMM_HPP
|
|
@@ -1,311 +0,0 @@
|
|
|
1
|
-
//
|
|
2
|
-
// MIT license
|
|
3
|
-
// Copyright (C) 2024 Intel Corporation
|
|
4
|
-
// SPDX-License-Identifier: MIT
|
|
5
|
-
//
|
|
6
|
-
|
|
7
|
-
//
|
|
8
|
-
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
9
|
-
// See https://llvm.org/LICENSE.txt for license information.
|
|
10
|
-
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
11
|
-
//
|
|
12
|
-
|
|
13
|
-
#include "ggml-impl.h"
|
|
14
|
-
#include "common.hpp"
|
|
15
|
-
#include "dequantize.hpp"
|
|
16
|
-
#include "getrows.hpp"
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
template<int qk, int qr, dequantize_kernel_t dequantize_kernel, typename dst_t>
|
|
20
|
-
static void k_get_rows(
|
|
21
|
-
const void * src0, const int32_t * src1, dst_t * dst,
|
|
22
|
-
int64_t ne00, /*int64_t ne01, int64_t ne02, int64_t ne03,*/
|
|
23
|
-
/*int64_t ne10, int64_t ne11,*/ int64_t ne12, /*int64_t ne13,*/
|
|
24
|
-
/*size_t s0,*/ size_t s1, size_t s2, size_t s3,
|
|
25
|
-
/*size_t nb00,*/ size_t nb01, size_t nb02, size_t nb03,
|
|
26
|
-
size_t s10, size_t s11, size_t s12,
|
|
27
|
-
const sycl::nd_item<3> &item_ct1/*, size_t s13*/) {
|
|
28
|
-
|
|
29
|
-
const int i00 = (item_ct1.get_group(2) * item_ct1.get_local_range(2) +
|
|
30
|
-
item_ct1.get_local_id(2)) *
|
|
31
|
-
2;
|
|
32
|
-
const int i10 = item_ct1.get_local_range(1) * item_ct1.get_group(1) +
|
|
33
|
-
item_ct1.get_local_id(1);
|
|
34
|
-
const int i11 = (item_ct1.get_group(0) * item_ct1.get_local_range(0) +
|
|
35
|
-
item_ct1.get_local_id(0)) /
|
|
36
|
-
ne12;
|
|
37
|
-
const int i12 = (item_ct1.get_group(0) * item_ct1.get_local_range(0) +
|
|
38
|
-
item_ct1.get_local_id(0)) %
|
|
39
|
-
ne12;
|
|
40
|
-
|
|
41
|
-
if (i00 >= ne00) {
|
|
42
|
-
return;
|
|
43
|
-
}
|
|
44
|
-
|
|
45
|
-
const int i01 = src1[i10*s10 + i11*s11 + i12*s12];
|
|
46
|
-
|
|
47
|
-
dst_t * dst_row = dst + i10*s1 + i11*s2 + i12*s3;
|
|
48
|
-
const void * src0_row = (const char *)src0 + i01*nb01 + i11*nb02 + i12*nb03;
|
|
49
|
-
|
|
50
|
-
const int ib = i00/qk; // block index
|
|
51
|
-
const int iqs = (i00%qk)/qr; // quant index
|
|
52
|
-
const int iybs = i00 - i00%qk; // dst block start index
|
|
53
|
-
const int y_offset = qr == 1 ? 1 : qk/2;
|
|
54
|
-
|
|
55
|
-
// dequantize
|
|
56
|
-
dfloat2 v;
|
|
57
|
-
dequantize_kernel(src0_row, ib, iqs, v);
|
|
58
|
-
|
|
59
|
-
dst_row[iybs + iqs + 0] = v.x();
|
|
60
|
-
dst_row[iybs + iqs + y_offset] = v.y();
|
|
61
|
-
}
|
|
62
|
-
|
|
63
|
-
template<int qk, int qr, dequantize_kernel_t_reorder dequantize_kernel_recorder, typename dst_t>
|
|
64
|
-
static void k_get_rows_reorder(
|
|
65
|
-
const void * src0, const void *src0_dq, const int32_t * src1, dst_t * dst,
|
|
66
|
-
int64_t ne00, /*int64_t ne01, int64_t ne02, int64_t ne03,*/
|
|
67
|
-
/*int64_t ne10, int64_t ne11,*/ int64_t ne12, /*int64_t ne13,*/
|
|
68
|
-
/*size_t s0,*/ size_t s1, size_t s2, size_t s3,
|
|
69
|
-
/*size_t nb00,*/ size_t nb01, size_t nb02, size_t nb03,
|
|
70
|
-
size_t s10, size_t s11, size_t s12,
|
|
71
|
-
const sycl::nd_item<3> &item_ct1/*, size_t s13*/) {
|
|
72
|
-
|
|
73
|
-
const int i00 = (item_ct1.get_group(2) * item_ct1.get_local_range(2) +
|
|
74
|
-
item_ct1.get_local_id(2)) *
|
|
75
|
-
2;
|
|
76
|
-
const int i10 = item_ct1.get_local_range(1) * item_ct1.get_group(1) +
|
|
77
|
-
item_ct1.get_local_id(1);
|
|
78
|
-
const int i11 = (item_ct1.get_group(0) * item_ct1.get_local_range(0) +
|
|
79
|
-
item_ct1.get_local_id(0)) /
|
|
80
|
-
ne12;
|
|
81
|
-
const int i12 = (item_ct1.get_group(0) * item_ct1.get_local_range(0) +
|
|
82
|
-
item_ct1.get_local_id(0)) %
|
|
83
|
-
ne12;
|
|
84
|
-
|
|
85
|
-
if (i00 >= ne00) {
|
|
86
|
-
return;
|
|
87
|
-
}
|
|
88
|
-
auto ncols = ne00;
|
|
89
|
-
const int i01 = src1[i10*s10 + i11*s11 + i12*s12];
|
|
90
|
-
|
|
91
|
-
dst_t * dst_row = dst + i10*s1 + i11*s2 + i12*s3;
|
|
92
|
-
|
|
93
|
-
const int src0_off = i01 * ncols + i00;
|
|
94
|
-
const int ib = src0_off / QK4_0; // block index
|
|
95
|
-
const int iqs = (i00%qk)/qr; // x quant index
|
|
96
|
-
const int iybs = i00 - i00%qk; // dst block start index
|
|
97
|
-
const int y_offset = qr == 1 ? 1 : qk/2;
|
|
98
|
-
|
|
99
|
-
// dequantize
|
|
100
|
-
dfloat2 v;
|
|
101
|
-
dequantize_kernel_recorder((const void *)src0_dq, ib, (const void *)src0, src0_off/2, v);
|
|
102
|
-
|
|
103
|
-
dst_row[iybs + iqs + 0] = v.x();
|
|
104
|
-
dst_row[iybs + iqs + y_offset] = v.y();
|
|
105
|
-
|
|
106
|
-
GGML_UNUSED(nb01);
|
|
107
|
-
GGML_UNUSED(nb02);
|
|
108
|
-
GGML_UNUSED(nb03);
|
|
109
|
-
}
|
|
110
|
-
|
|
111
|
-
template<typename src0_t, typename dst_t>
|
|
112
|
-
static void k_get_rows_float(
|
|
113
|
-
const src0_t * src0, const int32_t * src1, dst_t * dst,
|
|
114
|
-
int64_t ne00, /*int64_t ne01, int64_t ne02, int64_t ne03,*/
|
|
115
|
-
/*int64_t ne10, int64_t ne11,*/ int64_t ne12, /*int64_t ne13,*/
|
|
116
|
-
/*size_t s0,*/ size_t s1, size_t s2, size_t s3,
|
|
117
|
-
/*size_t nb00,*/ size_t nb01, size_t nb02, size_t nb03,
|
|
118
|
-
size_t s10, size_t s11, size_t s12,
|
|
119
|
-
const sycl::nd_item<3> &item_ct1/*, size_t s13*/) {
|
|
120
|
-
|
|
121
|
-
const int i00 = item_ct1.get_group(2) * item_ct1.get_local_range(2) +
|
|
122
|
-
item_ct1.get_local_id(2);
|
|
123
|
-
const int i10 = item_ct1.get_local_range(1) * item_ct1.get_group(1) +
|
|
124
|
-
item_ct1.get_local_id(1);
|
|
125
|
-
const int i11 = (item_ct1.get_group(0) * item_ct1.get_local_range(0) +
|
|
126
|
-
item_ct1.get_local_id(0)) /
|
|
127
|
-
ne12;
|
|
128
|
-
const int i12 = (item_ct1.get_group(0) * item_ct1.get_local_range(0) +
|
|
129
|
-
item_ct1.get_local_id(0)) %
|
|
130
|
-
ne12;
|
|
131
|
-
|
|
132
|
-
if (i00 >= ne00) {
|
|
133
|
-
return;
|
|
134
|
-
}
|
|
135
|
-
|
|
136
|
-
const int i01 = src1[i10*s10 + i11*s11 + i12*s12];
|
|
137
|
-
|
|
138
|
-
dst_t * dst_row = dst + i10*s1 + i11*s2 + i12*s3;
|
|
139
|
-
const src0_t * src0_row = (const src0_t *)((const char *)src0 + i01*nb01 + i11*nb02 + i12*nb03);
|
|
140
|
-
|
|
141
|
-
dst_row[i00] = src0_row[i00];
|
|
142
|
-
}
|
|
143
|
-
|
|
144
|
-
template <int qk, int qr, dequantize_kernel_t dq>
|
|
145
|
-
static void get_rows_sycl(ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
|
|
146
|
-
ggml_tensor *dst, const void *src0_dd,
|
|
147
|
-
const int32_t *src1_dd, float *dst_dd,
|
|
148
|
-
queue_ptr stream) {
|
|
149
|
-
|
|
150
|
-
GGML_TENSOR_BINARY_OP_LOCALS
|
|
151
|
-
|
|
152
|
-
const sycl::range<3> block_dims(1, 1, SYCL_GET_ROWS_BLOCK_SIZE);
|
|
153
|
-
const int block_num_x = (ne00 + 2*SYCL_GET_ROWS_BLOCK_SIZE - 1) / (2*SYCL_GET_ROWS_BLOCK_SIZE);
|
|
154
|
-
const sycl::range<3> block_nums(ne11 * ne12, ne10, block_num_x);
|
|
155
|
-
|
|
156
|
-
// strides in elements
|
|
157
|
-
//const size_t s0 = nb0 / ggml_element_size(dst);
|
|
158
|
-
const size_t s1 = nb1 / ggml_element_size(dst);
|
|
159
|
-
const size_t s2 = nb2 / ggml_element_size(dst);
|
|
160
|
-
const size_t s3 = nb3 / ggml_element_size(dst);
|
|
161
|
-
|
|
162
|
-
const size_t s10 = nb10 / ggml_element_size(src1);
|
|
163
|
-
const size_t s11 = nb11 / ggml_element_size(src1);
|
|
164
|
-
const size_t s12 = nb12 / ggml_element_size(src1);
|
|
165
|
-
//const size_t s13 = nb13 / ggml_element_size(src1);
|
|
166
|
-
|
|
167
|
-
GGML_ASSERT(ne00 % 2 == 0);
|
|
168
|
-
|
|
169
|
-
stream->parallel_for(sycl::nd_range<3>(block_nums * block_dims, block_dims),
|
|
170
|
-
[=](sycl::nd_item<3> item_ct1) {
|
|
171
|
-
k_get_rows<qk, qr, dq>(
|
|
172
|
-
src0_dd, src1_dd, dst_dd, ne00, ne12, s1, s2,
|
|
173
|
-
s3, nb01, nb02, nb03, s10, s11, s12, item_ct1);
|
|
174
|
-
});
|
|
175
|
-
|
|
176
|
-
GGML_UNUSED(dst);
|
|
177
|
-
GGML_UNUSED(ctx);
|
|
178
|
-
}
|
|
179
|
-
|
|
180
|
-
template <int qk, int qr, dequantize_kernel_t_reorder dq_reorder>
|
|
181
|
-
static void get_rows_sycl_reorder(ggml_backend_sycl_context & ctx, const ggml_tensor *src0, const ggml_tensor *src1,
|
|
182
|
-
ggml_tensor *dst, const void *src0_dd,
|
|
183
|
-
const int32_t *src1_dd, float *dst_dd,
|
|
184
|
-
queue_ptr stream) {
|
|
185
|
-
|
|
186
|
-
GGML_TENSOR_BINARY_OP_LOCALS
|
|
187
|
-
|
|
188
|
-
const sycl::range<3> block_dims(1, 1, SYCL_GET_ROWS_BLOCK_SIZE);
|
|
189
|
-
const int block_num_x = (ne00 + 2*SYCL_GET_ROWS_BLOCK_SIZE - 1) / (2*SYCL_GET_ROWS_BLOCK_SIZE);
|
|
190
|
-
const sycl::range<3> block_nums(ne11 * ne12, ne10, block_num_x);
|
|
191
|
-
|
|
192
|
-
// strides in elements
|
|
193
|
-
//const size_t s0 = nb0 / ggml_element_size(dst);
|
|
194
|
-
const size_t s1 = nb1 / ggml_element_size(dst);
|
|
195
|
-
const size_t s2 = nb2 / ggml_element_size(dst);
|
|
196
|
-
const size_t s3 = nb3 / ggml_element_size(dst);
|
|
197
|
-
|
|
198
|
-
const size_t s10 = nb10 / ggml_element_size(src1);
|
|
199
|
-
const size_t s11 = nb11 / ggml_element_size(src1);
|
|
200
|
-
const size_t s12 = nb12 / ggml_element_size(src1);
|
|
201
|
-
//const size_t s13 = nb13 / ggml_element_size(src1);
|
|
202
|
-
|
|
203
|
-
GGML_ASSERT(ne00 % 2 == 0);
|
|
204
|
-
|
|
205
|
-
const uint8_t* src0_q = (const uint8_t*)src0_dd;
|
|
206
|
-
const size_t ncols = ne00;
|
|
207
|
-
const size_t nrows = ne01;
|
|
208
|
-
const sycl::half* src0_dq = (const sycl::half*)(src0_q + nrows * ncols / 2);
|
|
209
|
-
stream->parallel_for(sycl::nd_range<3>(block_nums * block_dims, block_dims),
|
|
210
|
-
[=](sycl::nd_item<3> item_ct1) [[sycl::reqd_sub_group_size(WARP_SIZE)]]{
|
|
211
|
-
k_get_rows_reorder<qk, qr, dq_reorder>(
|
|
212
|
-
src0_dd, src0_dq, src1_dd, dst_dd, ne00, ne12, s1, s2,
|
|
213
|
-
s3, nb01, nb02, nb03, s10, s11, s12, item_ct1);
|
|
214
|
-
});
|
|
215
|
-
|
|
216
|
-
GGML_UNUSED(dst);
|
|
217
|
-
GGML_UNUSED(ctx);
|
|
218
|
-
}
|
|
219
|
-
|
|
220
|
-
|
|
221
|
-
template <typename src0_t>
|
|
222
|
-
static void get_rows_sycl_float(ggml_backend_sycl_context & ctx, const ggml_tensor *src0,
|
|
223
|
-
const ggml_tensor *src1, ggml_tensor *dst,
|
|
224
|
-
const src0_t *src0_dd, const int32_t *src1_dd,
|
|
225
|
-
float *dst_dd, queue_ptr stream) {
|
|
226
|
-
|
|
227
|
-
GGML_TENSOR_BINARY_OP_LOCALS
|
|
228
|
-
|
|
229
|
-
const sycl::range<3> block_dims(1, 1, SYCL_GET_ROWS_BLOCK_SIZE);
|
|
230
|
-
const int block_num_x = (ne00 + SYCL_GET_ROWS_BLOCK_SIZE - 1) / SYCL_GET_ROWS_BLOCK_SIZE;
|
|
231
|
-
const sycl::range<3> block_nums(ne11 * ne12, ne10, block_num_x);
|
|
232
|
-
|
|
233
|
-
// strides in elements
|
|
234
|
-
//const size_t s0 = nb0 / ggml_element_size(dst);
|
|
235
|
-
const size_t s1 = nb1 / ggml_element_size(dst);
|
|
236
|
-
const size_t s2 = nb2 / ggml_element_size(dst);
|
|
237
|
-
const size_t s3 = nb3 / ggml_element_size(dst);
|
|
238
|
-
|
|
239
|
-
const size_t s10 = nb10 / ggml_element_size(src1);
|
|
240
|
-
const size_t s11 = nb11 / ggml_element_size(src1);
|
|
241
|
-
const size_t s12 = nb12 / ggml_element_size(src1);
|
|
242
|
-
//const size_t s13 = nb13 / ggml_element_size(src1);
|
|
243
|
-
|
|
244
|
-
{
|
|
245
|
-
dpct::has_capability_or_fail(stream->get_device(),
|
|
246
|
-
{sycl::aspect::fp16});
|
|
247
|
-
|
|
248
|
-
stream->parallel_for(
|
|
249
|
-
sycl::nd_range<3>(block_nums * block_dims, block_dims),
|
|
250
|
-
[=](sycl::nd_item<3> item_ct1) {
|
|
251
|
-
k_get_rows_float(src0_dd, src1_dd, dst_dd, ne00, ne12, s1, s2,
|
|
252
|
-
s3, nb01, nb02, nb03, s10, s11, s12, item_ct1);
|
|
253
|
-
});
|
|
254
|
-
}
|
|
255
|
-
|
|
256
|
-
GGML_UNUSED(dst);
|
|
257
|
-
GGML_UNUSED(ctx);
|
|
258
|
-
}
|
|
259
|
-
|
|
260
|
-
void ggml_sycl_op_get_rows(ggml_backend_sycl_context & ctx, ggml_tensor *dst) {
|
|
261
|
-
|
|
262
|
-
GGML_ASSERT(dst->src[1]->type == GGML_TYPE_I32);
|
|
263
|
-
GGML_ASSERT(dst->type == GGML_TYPE_F32);
|
|
264
|
-
|
|
265
|
-
GGML_ASSERT(dst->src[0]->nb[0] == ggml_type_size(dst->src[0]->type));
|
|
266
|
-
GGML_ASSERT(dst->src[1]->nb[0] == ggml_type_size(dst->src[1]->type));
|
|
267
|
-
GGML_ASSERT(dst->nb[0] == ggml_type_size(dst->type));
|
|
268
|
-
|
|
269
|
-
const int32_t * src1_i32 = (const int32_t *) dst->src[1]->data;
|
|
270
|
-
/* TODO: Refactor and remove duplicates */
|
|
271
|
-
switch (dst->src[0]->type) {
|
|
272
|
-
case GGML_TYPE_F16:
|
|
273
|
-
get_rows_sycl_float(ctx, dst->src[0], dst->src[1], dst, (const sycl::half *)dst->src[0]->data,
|
|
274
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
275
|
-
break;
|
|
276
|
-
case GGML_TYPE_F32:
|
|
277
|
-
get_rows_sycl_float(ctx, dst->src[0], dst->src[1], dst, (const float *)dst->src[0]->data,
|
|
278
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
279
|
-
break;
|
|
280
|
-
case GGML_TYPE_Q4_0:
|
|
281
|
-
if (ctx.opt_feature.reorder && dst->op == GGML_OP_MUL_MAT) {
|
|
282
|
-
get_rows_sycl_reorder<QK4_0, QR4_0, dequantize_q4_0_reorder>(ctx, dst->src[0], dst->src[1], dst, (const float *)dst->src[0]->data,
|
|
283
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
284
|
-
} else {
|
|
285
|
-
get_rows_sycl<QK4_0, QR4_0, dequantize_q4_0>(ctx, dst->src[0], dst->src[1], dst, (const float *)dst->src[0]->data,
|
|
286
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
287
|
-
}
|
|
288
|
-
break;
|
|
289
|
-
case GGML_TYPE_Q4_1:
|
|
290
|
-
get_rows_sycl<QK4_1, QR4_1, dequantize_q4_1>(ctx, dst->src[0], dst->src[1], dst, (const float *)dst->src[0]->data,
|
|
291
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
292
|
-
break;
|
|
293
|
-
case GGML_TYPE_Q5_0:
|
|
294
|
-
get_rows_sycl<QK5_0, QR5_0, dequantize_q5_0>(ctx, dst->src[0], dst->src[1], dst, (const float *)dst->src[0]->data,
|
|
295
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
296
|
-
break;
|
|
297
|
-
case GGML_TYPE_Q5_1:
|
|
298
|
-
get_rows_sycl<QK5_1, QR5_1, dequantize_q5_1>(ctx, dst->src[0], dst->src[1], dst, (const float *)dst->src[0]->data,
|
|
299
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
300
|
-
break;
|
|
301
|
-
case GGML_TYPE_Q8_0:
|
|
302
|
-
get_rows_sycl<QK8_0, QR8_0, dequantize_q8_0>(ctx, dst->src[0], dst->src[1], dst, (const float *)dst->src[0]->data,
|
|
303
|
-
src1_i32, (float *)dst->data, ctx.stream());
|
|
304
|
-
break;
|
|
305
|
-
default:
|
|
306
|
-
// TODO: k-quants
|
|
307
|
-
GGML_LOG_ERROR("%s: unsupported type: %s\n", __func__, ggml_type_name(dst->src[0]->type));
|
|
308
|
-
GGML_ABORT("fatal error");
|
|
309
|
-
}
|
|
310
|
-
}
|
|
311
|
-
|
|
@@ -1,20 +0,0 @@
|
|
|
1
|
-
//
|
|
2
|
-
// MIT license
|
|
3
|
-
// Copyright (C) 2024 Intel Corporation
|
|
4
|
-
// SPDX-License-Identifier: MIT
|
|
5
|
-
//
|
|
6
|
-
|
|
7
|
-
//
|
|
8
|
-
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
9
|
-
// See https://llvm.org/LICENSE.txt for license information.
|
|
10
|
-
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
11
|
-
//
|
|
12
|
-
|
|
13
|
-
#ifndef GGML_SYCL_GETROWS_HPP
|
|
14
|
-
#define GGML_SYCL_GETROWS_HPP
|
|
15
|
-
|
|
16
|
-
#include "common.hpp"
|
|
17
|
-
|
|
18
|
-
void ggml_sycl_op_get_rows(ggml_backend_sycl_context & ctx, ggml_tensor *dst);
|
|
19
|
-
|
|
20
|
-
#endif // GGML_SYCL_GETROWS_HPP
|