@fugood/bricks-project 2.22.0-beta.13 → 2.22.0-beta.14

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -599,7 +599,7 @@ export const templateActionNameMap = {
599
599
  GENERATOR_LLM_TOKENIZE: {
600
600
  mode: 'GENERATOR_LLM_MODE',
601
601
  prompt: 'GENERATOR_LLM_PROMPT',
602
- promptImagePaths: 'GENERATOR_LLM_PROMPT_IMAGE_PATHS',
602
+ promptMediaPaths: 'GENERATOR_LLM_PROMPT_MEDIA_PATHS',
603
603
  messages: 'GENERATOR_LLM_MESSAGES',
604
604
  },
605
605
  GENERATOR_LLM_DETOKENIZE: {
@@ -613,7 +613,7 @@ export const templateActionNameMap = {
613
613
  parallelToolCalls: 'GENERATOR_LLM_PARALLEL_TOOL_CALLS',
614
614
  toolChoice: 'GENERATOR_LLM_TOOL_CHOICE',
615
615
  prompt: 'GENERATOR_LLM_PROMPT',
616
- promptImagePaths: 'GENERATOR_LLM_PROMPT_IMAGE_PATHS',
616
+ promptMediaPaths: 'GENERATOR_LLM_PROMPT_MEDIA_PATHS',
617
617
  promptTemplateData: 'GENERATOR_LLM_PROMPT_TEMPLATE_DATA',
618
618
  promptTemplateType: 'GENERATOR_LLM_PROMPT_TEMPLATE_TYPE',
619
619
  responseFormat: 'GENERATOR_LLM_RESPONSE_FORMAT',
@@ -626,7 +626,7 @@ export const templateActionNameMap = {
626
626
  parallelToolCalls: 'GENERATOR_LLM_PARALLEL_TOOL_CALLS',
627
627
  toolChoice: 'GENERATOR_LLM_TOOL_CHOICE',
628
628
  prompt: 'GENERATOR_LLM_PROMPT',
629
- promptImagePaths: 'GENERATOR_LLM_PROMPT_IMAGE_PATHS',
629
+ promptMediaPaths: 'GENERATOR_LLM_PROMPT_MEDIA_PATHS',
630
630
  promptTemplateData: 'GENERATOR_LLM_PROMPT_TEMPLATE_DATA',
631
631
  promptTemplateType: 'GENERATOR_LLM_PROMPT_TEMPLATE_TYPE',
632
632
  responseFormat: 'GENERATOR_LLM_RESPONSE_FORMAT',
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@fugood/bricks-project",
3
- "version": "2.22.0-beta.13",
3
+ "version": "2.22.0-beta.14",
4
4
  "main": "index.ts",
5
5
  "scripts": {
6
6
  "build": "node scripts/build.js"
@@ -14,5 +14,5 @@
14
14
  "lodash": "^4.17.4",
15
15
  "uuid": "^8.3.1"
16
16
  },
17
- "gitHead": "0039a8fa08e89411caf8608a98ae80d5cca64cbe"
17
+ "gitHead": "4b62185a9ec82561db90d85b8281de9e67f978ae"
18
18
  }
@@ -5094,24 +5094,19 @@ Default property:
5094
5094
  /* Model type */
5095
5095
  modelType?:
5096
5096
  | 'auto'
5097
- | 'gpt2'
5098
- | 'gptj'
5099
- | 'gpt_bigcode'
5100
- | 'gpt_neo'
5101
- | 'gpt_neox'
5102
- | 'bloom'
5103
- | 'mpt'
5104
- | 'opt'
5105
- | 'llama'
5106
- | 'falcon'
5107
- | 'mistral'
5097
+ | 'text-generation'
5098
+ | 'qwen2-vl'
5099
+ | 'paligemma'
5100
+ | 'llava'
5101
+ | 'llava_onevision'
5102
+ | 'moondream1'
5103
+ | 'florence2'
5104
+ | 'idefics3'
5105
+ | 'smolvlm'
5106
+ | 'phi3_v'
5108
5107
  | 't5'
5109
5108
  | 'mt5'
5110
5109
  | 'longt5'
5111
- | 'phi'
5112
- | 'qwen2'
5113
- | 'stablelm'
5114
- | 'gemma'
5115
5110
  | DataLink
5116
5111
  /* Load quantized model (deprecated, use `quantizeType` instead) */
5117
5112
  quantized?: boolean | DataLink
@@ -5924,7 +5919,7 @@ export type GeneratorLLMActionTokenize = ActionWithParams & {
5924
5919
  mapping?: string
5925
5920
  }
5926
5921
  | {
5927
- input: 'promptImagePaths'
5922
+ input: 'promptMediaPaths'
5928
5923
  value?: Array<any> | DataLink | EventProperty
5929
5924
  mapping?: string
5930
5925
  }
@@ -5986,7 +5981,7 @@ export type GeneratorLLMActionProcessPrompt = ActionWithParams & {
5986
5981
  mapping?: string
5987
5982
  }
5988
5983
  | {
5989
- input: 'promptImagePaths'
5984
+ input: 'promptMediaPaths'
5990
5985
  value?: Array<any> | DataLink | EventProperty
5991
5986
  mapping?: string
5992
5987
  }
@@ -6048,7 +6043,7 @@ export type GeneratorLLMActionCompletion = ActionWithParams & {
6048
6043
  mapping?: string
6049
6044
  }
6050
6045
  | {
6051
- input: 'promptImagePaths'
6046
+ input: 'promptMediaPaths'
6052
6047
  value?: Array<any> | DataLink | EventProperty
6053
6048
  mapping?: string
6054
6049
  }
@@ -6391,9 +6386,9 @@ Default property:
6391
6386
  | DataLink
6392
6387
  /* Prompt (text mode) */
6393
6388
  completionPrompt?: string | DataLink
6394
- /* Image paths to be used in the prompt template (PREVIEW FEATURE)
6395
- In prompt, use `<__image__>` for position of image content */
6396
- completionPromptImagePaths?: Array<string | DataLink> | DataLink
6389
+ /* Media paths to be used in the prompt template (PREVIEW FEATURE)
6390
+ In prompt, use `<__media__>` for position of media content */
6391
+ completionPromptMediaPaths?: Array<string | DataLink> | DataLink
6397
6392
  /* Data to be used in the prompt template (e.g. `Hello ${name}`). Supports nested data, such as `Hello ${user.name}`. */
6398
6393
  completionPromptTemplateData?: {} | DataLink
6399
6394
  /* The prompt template type */
@@ -6798,7 +6793,7 @@ interface GeneratorOpenAILLMDef {
6798
6793
  Default property:
6799
6794
  {
6800
6795
  "apiEndpoint": "https://api.openai.com/v1",
6801
- "model": "gpt-4o-mini",
6796
+ "model": "gpt-4o",
6802
6797
  "completionMessages": [
6803
6798
  {
6804
6799
  "role": "system",
@@ -6808,8 +6803,6 @@ Default property:
6808
6803
  "completionMaxTokens": 1024,
6809
6804
  "completionTemperature": 1,
6810
6805
  "completionTopP": 1,
6811
- "completionFrequencyPenalty": 0,
6812
- "completionPresencePenalty": 0,
6813
6806
  "completionStop": []
6814
6807
  }
6815
6808
  */
@@ -6898,7 +6891,11 @@ Default property:
6898
6891
  - Compatible with OpenAI API format
6899
6892
  - Supports function calling
6900
6893
  - Streaming responses
6901
- - Custom API endpoints */
6894
+ - Custom API endpoints, like
6895
+ - OpenAI API: https://platform.openai.com/docs/guides/text?api-mode=chat
6896
+ - Anthropic API: https://docs.anthropic.com/en/api/openai-sdk
6897
+ - Gemini API: https://ai.google.dev/gemini-api/docs/openai
6898
+ - llama.cpp server: https://github.com/ggml-org/llama.cpp/tree/master/tools/server */
6902
6899
  export type GeneratorOpenAILLM = Generator &
6903
6900
  GeneratorOpenAILLMDef & {
6904
6901
  templateKey: 'GENERATOR_OPENAI_LLM'