@blank-utils/llm 0.4.17 → 0.4.18

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1063,7 +1063,11 @@ ${newText}` : newText);
1063
1063
  try {
1064
1064
  const captioner = await captionerPromiseRef.current;
1065
1065
  if (captioner) {
1066
- const out = await captioner(dataUrl);
1066
+ const out = await captioner(dataUrl, {
1067
+ max_new_tokens: 100,
1068
+ num_beams: 4,
1069
+ repetition_penalty: 1.5
1070
+ });
1067
1071
  console.log("[ImagePipeline] Raw captioner output:", out);
1068
1072
  if (Array.isArray(out) && out[0] && out[0].generated_text) {
1069
1073
  extractedText = out[0].generated_text;
package/dist/index.js CHANGED
@@ -26,7 +26,7 @@ import {
26
26
  useCompletion,
27
27
  useLLM,
28
28
  useStream
29
- } from "./chunk-MV5JXFWL.js";
29
+ } from "./chunk-SB7Q5JMU.js";
30
30
  export {
31
31
  Chat,
32
32
  ChatInput,
@@ -9,7 +9,7 @@ import {
9
9
  useCompletion,
10
10
  useLLM,
11
11
  useStream
12
- } from "../chunk-MV5JXFWL.js";
12
+ } from "../chunk-SB7Q5JMU.js";
13
13
  export {
14
14
  Chat,
15
15
  ChatApp,
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@blank-utils/llm",
3
- "version": "0.4.17",
3
+ "version": "0.4.18",
4
4
  "description": "Run LLMs directly in your browser with WebGPU acceleration. Supports React hooks and eager background loading.",
5
5
  "type": "module",
6
6
  "main": "./dist/index.js",