@fugood/bricks-project 2.22.0-beta.13 → 2.22.0-beta.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/compile/action-name-map.ts +3 -3
- package/package.json +2 -2
- package/types/generators.ts +22 -25
|
@@ -599,7 +599,7 @@ export const templateActionNameMap = {
|
|
|
599
599
|
GENERATOR_LLM_TOKENIZE: {
|
|
600
600
|
mode: 'GENERATOR_LLM_MODE',
|
|
601
601
|
prompt: 'GENERATOR_LLM_PROMPT',
|
|
602
|
-
|
|
602
|
+
promptMediaPaths: 'GENERATOR_LLM_PROMPT_MEDIA_PATHS',
|
|
603
603
|
messages: 'GENERATOR_LLM_MESSAGES',
|
|
604
604
|
},
|
|
605
605
|
GENERATOR_LLM_DETOKENIZE: {
|
|
@@ -613,7 +613,7 @@ export const templateActionNameMap = {
|
|
|
613
613
|
parallelToolCalls: 'GENERATOR_LLM_PARALLEL_TOOL_CALLS',
|
|
614
614
|
toolChoice: 'GENERATOR_LLM_TOOL_CHOICE',
|
|
615
615
|
prompt: 'GENERATOR_LLM_PROMPT',
|
|
616
|
-
|
|
616
|
+
promptMediaPaths: 'GENERATOR_LLM_PROMPT_MEDIA_PATHS',
|
|
617
617
|
promptTemplateData: 'GENERATOR_LLM_PROMPT_TEMPLATE_DATA',
|
|
618
618
|
promptTemplateType: 'GENERATOR_LLM_PROMPT_TEMPLATE_TYPE',
|
|
619
619
|
responseFormat: 'GENERATOR_LLM_RESPONSE_FORMAT',
|
|
@@ -626,7 +626,7 @@ export const templateActionNameMap = {
|
|
|
626
626
|
parallelToolCalls: 'GENERATOR_LLM_PARALLEL_TOOL_CALLS',
|
|
627
627
|
toolChoice: 'GENERATOR_LLM_TOOL_CHOICE',
|
|
628
628
|
prompt: 'GENERATOR_LLM_PROMPT',
|
|
629
|
-
|
|
629
|
+
promptMediaPaths: 'GENERATOR_LLM_PROMPT_MEDIA_PATHS',
|
|
630
630
|
promptTemplateData: 'GENERATOR_LLM_PROMPT_TEMPLATE_DATA',
|
|
631
631
|
promptTemplateType: 'GENERATOR_LLM_PROMPT_TEMPLATE_TYPE',
|
|
632
632
|
responseFormat: 'GENERATOR_LLM_RESPONSE_FORMAT',
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@fugood/bricks-project",
|
|
3
|
-
"version": "2.22.0-beta.
|
|
3
|
+
"version": "2.22.0-beta.14",
|
|
4
4
|
"main": "index.ts",
|
|
5
5
|
"scripts": {
|
|
6
6
|
"build": "node scripts/build.js"
|
|
@@ -14,5 +14,5 @@
|
|
|
14
14
|
"lodash": "^4.17.4",
|
|
15
15
|
"uuid": "^8.3.1"
|
|
16
16
|
},
|
|
17
|
-
"gitHead": "
|
|
17
|
+
"gitHead": "4b62185a9ec82561db90d85b8281de9e67f978ae"
|
|
18
18
|
}
|
package/types/generators.ts
CHANGED
|
@@ -5094,24 +5094,19 @@ Default property:
|
|
|
5094
5094
|
/* Model type */
|
|
5095
5095
|
modelType?:
|
|
5096
5096
|
| 'auto'
|
|
5097
|
-
| '
|
|
5098
|
-
| '
|
|
5099
|
-
| '
|
|
5100
|
-
| '
|
|
5101
|
-
| '
|
|
5102
|
-
| '
|
|
5103
|
-
| '
|
|
5104
|
-
| '
|
|
5105
|
-
| '
|
|
5106
|
-
| '
|
|
5107
|
-
| 'mistral'
|
|
5097
|
+
| 'text-generation'
|
|
5098
|
+
| 'qwen2-vl'
|
|
5099
|
+
| 'paligemma'
|
|
5100
|
+
| 'llava'
|
|
5101
|
+
| 'llava_onevision'
|
|
5102
|
+
| 'moondream1'
|
|
5103
|
+
| 'florence2'
|
|
5104
|
+
| 'idefics3'
|
|
5105
|
+
| 'smolvlm'
|
|
5106
|
+
| 'phi3_v'
|
|
5108
5107
|
| 't5'
|
|
5109
5108
|
| 'mt5'
|
|
5110
5109
|
| 'longt5'
|
|
5111
|
-
| 'phi'
|
|
5112
|
-
| 'qwen2'
|
|
5113
|
-
| 'stablelm'
|
|
5114
|
-
| 'gemma'
|
|
5115
5110
|
| DataLink
|
|
5116
5111
|
/* Load quantized model (deprecated, use `quantizeType` instead) */
|
|
5117
5112
|
quantized?: boolean | DataLink
|
|
@@ -5924,7 +5919,7 @@ export type GeneratorLLMActionTokenize = ActionWithParams & {
|
|
|
5924
5919
|
mapping?: string
|
|
5925
5920
|
}
|
|
5926
5921
|
| {
|
|
5927
|
-
input: '
|
|
5922
|
+
input: 'promptMediaPaths'
|
|
5928
5923
|
value?: Array<any> | DataLink | EventProperty
|
|
5929
5924
|
mapping?: string
|
|
5930
5925
|
}
|
|
@@ -5986,7 +5981,7 @@ export type GeneratorLLMActionProcessPrompt = ActionWithParams & {
|
|
|
5986
5981
|
mapping?: string
|
|
5987
5982
|
}
|
|
5988
5983
|
| {
|
|
5989
|
-
input: '
|
|
5984
|
+
input: 'promptMediaPaths'
|
|
5990
5985
|
value?: Array<any> | DataLink | EventProperty
|
|
5991
5986
|
mapping?: string
|
|
5992
5987
|
}
|
|
@@ -6048,7 +6043,7 @@ export type GeneratorLLMActionCompletion = ActionWithParams & {
|
|
|
6048
6043
|
mapping?: string
|
|
6049
6044
|
}
|
|
6050
6045
|
| {
|
|
6051
|
-
input: '
|
|
6046
|
+
input: 'promptMediaPaths'
|
|
6052
6047
|
value?: Array<any> | DataLink | EventProperty
|
|
6053
6048
|
mapping?: string
|
|
6054
6049
|
}
|
|
@@ -6391,9 +6386,9 @@ Default property:
|
|
|
6391
6386
|
| DataLink
|
|
6392
6387
|
/* Prompt (text mode) */
|
|
6393
6388
|
completionPrompt?: string | DataLink
|
|
6394
|
-
/*
|
|
6395
|
-
In prompt, use `<
|
|
6396
|
-
|
|
6389
|
+
/* Media paths to be used in the prompt template (PREVIEW FEATURE)
|
|
6390
|
+
In prompt, use `<__media__>` for position of media content */
|
|
6391
|
+
completionPromptMediaPaths?: Array<string | DataLink> | DataLink
|
|
6397
6392
|
/* Data to be used in the prompt template (e.g. `Hello ${name}`). Supports nested data, such as `Hello ${user.name}`. */
|
|
6398
6393
|
completionPromptTemplateData?: {} | DataLink
|
|
6399
6394
|
/* The prompt template type */
|
|
@@ -6798,7 +6793,7 @@ interface GeneratorOpenAILLMDef {
|
|
|
6798
6793
|
Default property:
|
|
6799
6794
|
{
|
|
6800
6795
|
"apiEndpoint": "https://api.openai.com/v1",
|
|
6801
|
-
"model": "gpt-4o
|
|
6796
|
+
"model": "gpt-4o",
|
|
6802
6797
|
"completionMessages": [
|
|
6803
6798
|
{
|
|
6804
6799
|
"role": "system",
|
|
@@ -6808,8 +6803,6 @@ Default property:
|
|
|
6808
6803
|
"completionMaxTokens": 1024,
|
|
6809
6804
|
"completionTemperature": 1,
|
|
6810
6805
|
"completionTopP": 1,
|
|
6811
|
-
"completionFrequencyPenalty": 0,
|
|
6812
|
-
"completionPresencePenalty": 0,
|
|
6813
6806
|
"completionStop": []
|
|
6814
6807
|
}
|
|
6815
6808
|
*/
|
|
@@ -6898,7 +6891,11 @@ Default property:
|
|
|
6898
6891
|
- Compatible with OpenAI API format
|
|
6899
6892
|
- Supports function calling
|
|
6900
6893
|
- Streaming responses
|
|
6901
|
-
- Custom API endpoints
|
|
6894
|
+
- Custom API endpoints, like
|
|
6895
|
+
- OpenAI API: https://platform.openai.com/docs/guides/text?api-mode=chat
|
|
6896
|
+
- Anthropic API: https://docs.anthropic.com/en/api/openai-sdk
|
|
6897
|
+
- Gemini API: https://ai.google.dev/gemini-api/docs/openai
|
|
6898
|
+
- llama.cpp server: https://github.com/ggml-org/llama.cpp/tree/master/tools/server */
|
|
6902
6899
|
export type GeneratorOpenAILLM = Generator &
|
|
6903
6900
|
GeneratorOpenAILLMDef & {
|
|
6904
6901
|
templateKey: 'GENERATOR_OPENAI_LLM'
|