@letta-ai/letta-client 1.0.0-alpha.16 → 1.0.0-alpha.17
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +15 -0
- package/client.d.mts +2 -2
- package/client.d.mts.map +1 -1
- package/client.d.ts +2 -2
- package/client.d.ts.map +1 -1
- package/client.js.map +1 -1
- package/client.mjs.map +1 -1
- package/package.json +1 -1
- package/resources/agents/agents.d.mts +962 -111
- package/resources/agents/agents.d.mts.map +1 -1
- package/resources/agents/agents.d.ts +962 -111
- package/resources/agents/agents.d.ts.map +1 -1
- package/resources/agents/agents.js.map +1 -1
- package/resources/agents/agents.mjs.map +1 -1
- package/resources/agents/blocks.d.mts +1 -15
- package/resources/agents/blocks.d.mts.map +1 -1
- package/resources/agents/blocks.d.ts +1 -15
- package/resources/agents/blocks.d.ts.map +1 -1
- package/resources/agents/folders.d.mts +2 -7
- package/resources/agents/folders.d.mts.map +1 -1
- package/resources/agents/folders.d.ts +2 -7
- package/resources/agents/folders.d.ts.map +1 -1
- package/resources/agents/messages.d.mts +0 -10
- package/resources/agents/messages.d.mts.map +1 -1
- package/resources/agents/messages.d.ts +0 -10
- package/resources/agents/messages.d.ts.map +1 -1
- package/resources/archives.d.mts +0 -6
- package/resources/archives.d.mts.map +1 -1
- package/resources/archives.d.ts +0 -6
- package/resources/archives.d.ts.map +1 -1
- package/resources/folders/folders.d.mts +0 -6
- package/resources/folders/folders.d.mts.map +1 -1
- package/resources/folders/folders.d.ts +0 -6
- package/resources/folders/folders.d.ts.map +1 -1
- package/resources/folders/folders.js.map +1 -1
- package/resources/folders/folders.mjs.map +1 -1
- package/resources/index.d.mts +1 -1
- package/resources/index.d.mts.map +1 -1
- package/resources/index.d.ts +1 -1
- package/resources/index.d.ts.map +1 -1
- package/resources/index.js.map +1 -1
- package/resources/index.mjs.map +1 -1
- package/resources/mcp-servers/mcp-servers.d.mts +0 -10
- package/resources/mcp-servers/mcp-servers.d.mts.map +1 -1
- package/resources/mcp-servers/mcp-servers.d.ts +0 -10
- package/resources/mcp-servers/mcp-servers.d.ts.map +1 -1
- package/resources/mcp-servers/mcp-servers.js.map +1 -1
- package/resources/mcp-servers/mcp-servers.mjs.map +1 -1
- package/resources/models/embeddings.d.mts +1 -66
- package/resources/models/embeddings.d.mts.map +1 -1
- package/resources/models/embeddings.d.ts +1 -66
- package/resources/models/embeddings.d.ts.map +1 -1
- package/resources/models/index.d.mts +1 -1
- package/resources/models/index.d.mts.map +1 -1
- package/resources/models/index.d.ts +1 -1
- package/resources/models/index.d.ts.map +1 -1
- package/resources/models/index.js.map +1 -1
- package/resources/models/index.mjs.map +1 -1
- package/resources/models/models.d.mts +169 -109
- package/resources/models/models.d.mts.map +1 -1
- package/resources/models/models.d.ts +169 -109
- package/resources/models/models.d.ts.map +1 -1
- package/resources/models/models.js.map +1 -1
- package/resources/models/models.mjs.map +1 -1
- package/resources/runs/runs.d.mts +0 -5
- package/resources/runs/runs.d.mts.map +1 -1
- package/resources/runs/runs.d.ts +0 -5
- package/resources/runs/runs.d.ts.map +1 -1
- package/resources/runs/runs.js.map +1 -1
- package/resources/runs/runs.mjs.map +1 -1
- package/resources/tools.d.mts +0 -4
- package/resources/tools.d.mts.map +1 -1
- package/resources/tools.d.ts +0 -4
- package/resources/tools.d.ts.map +1 -1
- package/src/client.ts +4 -0
- package/src/resources/agents/agents.ts +1235 -69
- package/src/resources/agents/blocks.ts +1 -15
- package/src/resources/agents/folders.ts +2 -7
- package/src/resources/agents/messages.ts +0 -10
- package/src/resources/archives.ts +0 -6
- package/src/resources/folders/folders.ts +0 -6
- package/src/resources/index.ts +2 -0
- package/src/resources/mcp-servers/mcp-servers.ts +0 -10
- package/src/resources/models/embeddings.ts +1 -100
- package/src/resources/models/index.ts +2 -0
- package/src/resources/models/models.ts +249 -153
- package/src/resources/runs/runs.ts +0 -5
- package/src/resources/tools.ts +0 -4
- package/src/version.ts +1 -1
- package/version.d.mts +1 -1
- package/version.d.ts +1 -1
- package/version.js +1 -1
- package/version.mjs +1 -1
|
@@ -254,16 +254,6 @@ export interface AgentEnvironmentVariable {
|
|
|
254
254
|
* Representation of an agent's state. This is the state of the agent at a given
|
|
255
255
|
* time, and is persisted in the DB backend. The state has all the information
|
|
256
256
|
* needed to recreate a persisted agent.
|
|
257
|
-
*
|
|
258
|
-
* Parameters: id (str): The unique identifier of the agent. name (str): The name
|
|
259
|
-
* of the agent (must be unique to the user). created_at (datetime): The datetime
|
|
260
|
-
* the agent was created. message_ids (List[str]): The ids of the messages in the
|
|
261
|
-
* agent's in-context memory. memory (Memory): The in-context memory of the agent.
|
|
262
|
-
* tools (List[str]): The tools used by the agent. This includes any memory editing
|
|
263
|
-
* functions specified in `memory`. system (str): The system prompt used by the
|
|
264
|
-
* agent. llm_config (LLMConfig): The LLM configuration used by the agent.
|
|
265
|
-
* embedding_config (EmbeddingConfig): The embedding configuration used by the
|
|
266
|
-
* agent.
|
|
267
257
|
*/
|
|
268
258
|
export interface AgentState {
|
|
269
259
|
/**
|
|
@@ -350,9 +340,9 @@ export interface AgentState {
|
|
|
350
340
|
description?: string | null;
|
|
351
341
|
|
|
352
342
|
/**
|
|
353
|
-
* The embedding model used by the agent.
|
|
343
|
+
* The embedding model handle used by the agent (format: provider/model-name).
|
|
354
344
|
*/
|
|
355
|
-
embedding?:
|
|
345
|
+
embedding?: string | null;
|
|
356
346
|
|
|
357
347
|
/**
|
|
358
348
|
* If set to True, memory management will move to a background agent thread.
|
|
@@ -429,9 +419,25 @@ export interface AgentState {
|
|
|
429
419
|
metadata?: { [key: string]: unknown } | null;
|
|
430
420
|
|
|
431
421
|
/**
|
|
432
|
-
*
|
|
422
|
+
* The model handle used by the agent (format: provider/model-name).
|
|
433
423
|
*/
|
|
434
|
-
model?:
|
|
424
|
+
model?: string | null;
|
|
425
|
+
|
|
426
|
+
/**
|
|
427
|
+
* The model settings used by the agent.
|
|
428
|
+
*/
|
|
429
|
+
model_settings?:
|
|
430
|
+
| AgentState.OpenAIModelSettings
|
|
431
|
+
| AgentState.AnthropicModelSettings
|
|
432
|
+
| AgentState.GoogleAIModelSettings
|
|
433
|
+
| AgentState.GoogleVertexModelSettings
|
|
434
|
+
| AgentState.AzureModelSettings
|
|
435
|
+
| AgentState.XaiModelSettings
|
|
436
|
+
| AgentState.GroqModelSettings
|
|
437
|
+
| AgentState.DeepseekModelSettings
|
|
438
|
+
| AgentState.TogetherModelSettings
|
|
439
|
+
| AgentState.BedrockModelSettings
|
|
440
|
+
| null;
|
|
435
441
|
|
|
436
442
|
/**
|
|
437
443
|
* @deprecated Deprecated: Use `managed_group` field instead. The multi-agent group
|
|
@@ -639,13 +645,8 @@ export namespace AgentState {
|
|
|
639
645
|
}
|
|
640
646
|
|
|
641
647
|
/**
|
|
642
|
-
* Representation of a source, which is a collection of
|
|
643
|
-
*
|
|
644
|
-
* Parameters: id (str): The ID of the source name (str): The name of the source.
|
|
645
|
-
* embedding_config (EmbeddingConfig): The embedding configuration used by the
|
|
646
|
-
* source. user_id (str): The ID of the user that created the source. metadata
|
|
647
|
-
* (dict): Metadata associated with the source. description (str): The description
|
|
648
|
-
* of the source.
|
|
648
|
+
* (Deprecated: Use Folder) Representation of a source, which is a collection of
|
|
649
|
+
* files and passages.
|
|
649
650
|
*/
|
|
650
651
|
export interface Source {
|
|
651
652
|
/**
|
|
@@ -704,34 +705,411 @@ export namespace AgentState {
|
|
|
704
705
|
vector_db_provider?: ArchivesAPI.VectorDBProvider;
|
|
705
706
|
}
|
|
706
707
|
|
|
708
|
+
export interface OpenAIModelSettings {
|
|
709
|
+
/**
|
|
710
|
+
* The maximum number of tokens the model can generate.
|
|
711
|
+
*/
|
|
712
|
+
max_output_tokens?: number;
|
|
713
|
+
|
|
714
|
+
/**
|
|
715
|
+
* Whether to enable parallel tool calling.
|
|
716
|
+
*/
|
|
717
|
+
parallel_tool_calls?: boolean;
|
|
718
|
+
|
|
719
|
+
/**
|
|
720
|
+
* The provider of the model.
|
|
721
|
+
*/
|
|
722
|
+
provider?: 'openai';
|
|
723
|
+
|
|
724
|
+
/**
|
|
725
|
+
* The reasoning configuration for the model.
|
|
726
|
+
*/
|
|
727
|
+
reasoning?: OpenAIModelSettings.Reasoning;
|
|
728
|
+
|
|
729
|
+
/**
|
|
730
|
+
* The response format for the model.
|
|
731
|
+
*/
|
|
732
|
+
response_format?:
|
|
733
|
+
| AgentsAPI.TextResponseFormat
|
|
734
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
735
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
736
|
+
| null;
|
|
737
|
+
|
|
738
|
+
/**
|
|
739
|
+
* The temperature of the model.
|
|
740
|
+
*/
|
|
741
|
+
temperature?: number;
|
|
742
|
+
}
|
|
743
|
+
|
|
744
|
+
export namespace OpenAIModelSettings {
|
|
745
|
+
/**
|
|
746
|
+
* The reasoning configuration for the model.
|
|
747
|
+
*/
|
|
748
|
+
export interface Reasoning {
|
|
749
|
+
/**
|
|
750
|
+
* The reasoning effort to use when generating text reasoning models
|
|
751
|
+
*/
|
|
752
|
+
reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
|
|
753
|
+
}
|
|
754
|
+
}
|
|
755
|
+
|
|
756
|
+
export interface AnthropicModelSettings {
|
|
757
|
+
/**
|
|
758
|
+
* The maximum number of tokens the model can generate.
|
|
759
|
+
*/
|
|
760
|
+
max_output_tokens?: number;
|
|
761
|
+
|
|
762
|
+
/**
|
|
763
|
+
* Whether to enable parallel tool calling.
|
|
764
|
+
*/
|
|
765
|
+
parallel_tool_calls?: boolean;
|
|
766
|
+
|
|
767
|
+
/**
|
|
768
|
+
* The provider of the model.
|
|
769
|
+
*/
|
|
770
|
+
provider?: 'anthropic';
|
|
771
|
+
|
|
772
|
+
/**
|
|
773
|
+
* The temperature of the model.
|
|
774
|
+
*/
|
|
775
|
+
temperature?: number;
|
|
776
|
+
|
|
777
|
+
/**
|
|
778
|
+
* The thinking configuration for the model.
|
|
779
|
+
*/
|
|
780
|
+
thinking?: AnthropicModelSettings.Thinking;
|
|
781
|
+
|
|
782
|
+
/**
|
|
783
|
+
* Soft control for how verbose model output should be, used for GPT-5 models.
|
|
784
|
+
*/
|
|
785
|
+
verbosity?: 'low' | 'medium' | 'high' | null;
|
|
786
|
+
}
|
|
787
|
+
|
|
788
|
+
export namespace AnthropicModelSettings {
|
|
789
|
+
/**
|
|
790
|
+
* The thinking configuration for the model.
|
|
791
|
+
*/
|
|
792
|
+
export interface Thinking {
|
|
793
|
+
/**
|
|
794
|
+
* The maximum number of tokens the model can use for extended thinking.
|
|
795
|
+
*/
|
|
796
|
+
budget_tokens?: number;
|
|
797
|
+
|
|
798
|
+
/**
|
|
799
|
+
* The type of thinking to use.
|
|
800
|
+
*/
|
|
801
|
+
type?: 'enabled' | 'disabled';
|
|
802
|
+
}
|
|
803
|
+
}
|
|
804
|
+
|
|
805
|
+
export interface GoogleAIModelSettings {
|
|
806
|
+
/**
|
|
807
|
+
* The maximum number of tokens the model can generate.
|
|
808
|
+
*/
|
|
809
|
+
max_output_tokens?: number;
|
|
810
|
+
|
|
811
|
+
/**
|
|
812
|
+
* Whether to enable parallel tool calling.
|
|
813
|
+
*/
|
|
814
|
+
parallel_tool_calls?: boolean;
|
|
815
|
+
|
|
816
|
+
/**
|
|
817
|
+
* The provider of the model.
|
|
818
|
+
*/
|
|
819
|
+
provider?: 'google_ai';
|
|
820
|
+
|
|
821
|
+
/**
|
|
822
|
+
* The response schema for the model.
|
|
823
|
+
*/
|
|
824
|
+
response_schema?:
|
|
825
|
+
| AgentsAPI.TextResponseFormat
|
|
826
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
827
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
828
|
+
| null;
|
|
829
|
+
|
|
830
|
+
/**
|
|
831
|
+
* The temperature of the model.
|
|
832
|
+
*/
|
|
833
|
+
temperature?: number;
|
|
834
|
+
|
|
835
|
+
/**
|
|
836
|
+
* The thinking configuration for the model.
|
|
837
|
+
*/
|
|
838
|
+
thinking_config?: GoogleAIModelSettings.ThinkingConfig;
|
|
839
|
+
}
|
|
840
|
+
|
|
841
|
+
export namespace GoogleAIModelSettings {
|
|
842
|
+
/**
|
|
843
|
+
* The thinking configuration for the model.
|
|
844
|
+
*/
|
|
845
|
+
export interface ThinkingConfig {
|
|
846
|
+
/**
|
|
847
|
+
* Whether to include thoughts in the model's response.
|
|
848
|
+
*/
|
|
849
|
+
include_thoughts?: boolean;
|
|
850
|
+
|
|
851
|
+
/**
|
|
852
|
+
* The thinking budget for the model.
|
|
853
|
+
*/
|
|
854
|
+
thinking_budget?: number;
|
|
855
|
+
}
|
|
856
|
+
}
|
|
857
|
+
|
|
858
|
+
export interface GoogleVertexModelSettings {
|
|
859
|
+
/**
|
|
860
|
+
* The maximum number of tokens the model can generate.
|
|
861
|
+
*/
|
|
862
|
+
max_output_tokens?: number;
|
|
863
|
+
|
|
864
|
+
/**
|
|
865
|
+
* Whether to enable parallel tool calling.
|
|
866
|
+
*/
|
|
867
|
+
parallel_tool_calls?: boolean;
|
|
868
|
+
|
|
869
|
+
/**
|
|
870
|
+
* The provider of the model.
|
|
871
|
+
*/
|
|
872
|
+
provider?: 'google_vertex';
|
|
873
|
+
|
|
874
|
+
/**
|
|
875
|
+
* The response schema for the model.
|
|
876
|
+
*/
|
|
877
|
+
response_schema?:
|
|
878
|
+
| AgentsAPI.TextResponseFormat
|
|
879
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
880
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
881
|
+
| null;
|
|
882
|
+
|
|
883
|
+
/**
|
|
884
|
+
* The temperature of the model.
|
|
885
|
+
*/
|
|
886
|
+
temperature?: number;
|
|
887
|
+
|
|
888
|
+
/**
|
|
889
|
+
* The thinking configuration for the model.
|
|
890
|
+
*/
|
|
891
|
+
thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
|
|
892
|
+
}
|
|
893
|
+
|
|
894
|
+
export namespace GoogleVertexModelSettings {
|
|
895
|
+
/**
|
|
896
|
+
* The thinking configuration for the model.
|
|
897
|
+
*/
|
|
898
|
+
export interface ThinkingConfig {
|
|
899
|
+
/**
|
|
900
|
+
* Whether to include thoughts in the model's response.
|
|
901
|
+
*/
|
|
902
|
+
include_thoughts?: boolean;
|
|
903
|
+
|
|
904
|
+
/**
|
|
905
|
+
* The thinking budget for the model.
|
|
906
|
+
*/
|
|
907
|
+
thinking_budget?: number;
|
|
908
|
+
}
|
|
909
|
+
}
|
|
910
|
+
|
|
911
|
+
/**
|
|
912
|
+
* Azure OpenAI model configuration (OpenAI-compatible).
|
|
913
|
+
*/
|
|
914
|
+
export interface AzureModelSettings {
|
|
915
|
+
/**
|
|
916
|
+
* The maximum number of tokens the model can generate.
|
|
917
|
+
*/
|
|
918
|
+
max_output_tokens?: number;
|
|
919
|
+
|
|
920
|
+
/**
|
|
921
|
+
* Whether to enable parallel tool calling.
|
|
922
|
+
*/
|
|
923
|
+
parallel_tool_calls?: boolean;
|
|
924
|
+
|
|
925
|
+
/**
|
|
926
|
+
* The provider of the model.
|
|
927
|
+
*/
|
|
928
|
+
provider?: 'azure';
|
|
929
|
+
|
|
930
|
+
/**
|
|
931
|
+
* The response format for the model.
|
|
932
|
+
*/
|
|
933
|
+
response_format?:
|
|
934
|
+
| AgentsAPI.TextResponseFormat
|
|
935
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
936
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
937
|
+
| null;
|
|
938
|
+
|
|
939
|
+
/**
|
|
940
|
+
* The temperature of the model.
|
|
941
|
+
*/
|
|
942
|
+
temperature?: number;
|
|
943
|
+
}
|
|
944
|
+
|
|
945
|
+
/**
|
|
946
|
+
* xAI model configuration (OpenAI-compatible).
|
|
947
|
+
*/
|
|
948
|
+
export interface XaiModelSettings {
|
|
949
|
+
/**
|
|
950
|
+
* The maximum number of tokens the model can generate.
|
|
951
|
+
*/
|
|
952
|
+
max_output_tokens?: number;
|
|
953
|
+
|
|
954
|
+
/**
|
|
955
|
+
* Whether to enable parallel tool calling.
|
|
956
|
+
*/
|
|
957
|
+
parallel_tool_calls?: boolean;
|
|
958
|
+
|
|
959
|
+
/**
|
|
960
|
+
* The provider of the model.
|
|
961
|
+
*/
|
|
962
|
+
provider?: 'xai';
|
|
963
|
+
|
|
964
|
+
/**
|
|
965
|
+
* The response format for the model.
|
|
966
|
+
*/
|
|
967
|
+
response_format?:
|
|
968
|
+
| AgentsAPI.TextResponseFormat
|
|
969
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
970
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
971
|
+
| null;
|
|
972
|
+
|
|
973
|
+
/**
|
|
974
|
+
* The temperature of the model.
|
|
975
|
+
*/
|
|
976
|
+
temperature?: number;
|
|
977
|
+
}
|
|
978
|
+
|
|
979
|
+
/**
|
|
980
|
+
* Groq model configuration (OpenAI-compatible).
|
|
981
|
+
*/
|
|
982
|
+
export interface GroqModelSettings {
|
|
983
|
+
/**
|
|
984
|
+
* The maximum number of tokens the model can generate.
|
|
985
|
+
*/
|
|
986
|
+
max_output_tokens?: number;
|
|
987
|
+
|
|
988
|
+
/**
|
|
989
|
+
* Whether to enable parallel tool calling.
|
|
990
|
+
*/
|
|
991
|
+
parallel_tool_calls?: boolean;
|
|
992
|
+
|
|
993
|
+
/**
|
|
994
|
+
* The provider of the model.
|
|
995
|
+
*/
|
|
996
|
+
provider?: 'groq';
|
|
997
|
+
|
|
998
|
+
/**
|
|
999
|
+
* The response format for the model.
|
|
1000
|
+
*/
|
|
1001
|
+
response_format?:
|
|
1002
|
+
| AgentsAPI.TextResponseFormat
|
|
1003
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1004
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1005
|
+
| null;
|
|
1006
|
+
|
|
1007
|
+
/**
|
|
1008
|
+
* The temperature of the model.
|
|
1009
|
+
*/
|
|
1010
|
+
temperature?: number;
|
|
1011
|
+
}
|
|
1012
|
+
|
|
707
1013
|
/**
|
|
708
|
-
*
|
|
1014
|
+
* Deepseek model configuration (OpenAI-compatible).
|
|
709
1015
|
*/
|
|
710
|
-
export interface
|
|
1016
|
+
export interface DeepseekModelSettings {
|
|
711
1017
|
/**
|
|
712
|
-
* The
|
|
1018
|
+
* The maximum number of tokens the model can generate.
|
|
1019
|
+
*/
|
|
1020
|
+
max_output_tokens?: number;
|
|
1021
|
+
|
|
1022
|
+
/**
|
|
1023
|
+
* Whether to enable parallel tool calling.
|
|
713
1024
|
*/
|
|
714
|
-
|
|
1025
|
+
parallel_tool_calls?: boolean;
|
|
715
1026
|
|
|
716
1027
|
/**
|
|
717
1028
|
* The provider of the model.
|
|
718
1029
|
*/
|
|
719
|
-
provider
|
|
1030
|
+
provider?: 'deepseek';
|
|
1031
|
+
|
|
1032
|
+
/**
|
|
1033
|
+
* The response format for the model.
|
|
1034
|
+
*/
|
|
1035
|
+
response_format?:
|
|
1036
|
+
| AgentsAPI.TextResponseFormat
|
|
1037
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1038
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1039
|
+
| null;
|
|
1040
|
+
|
|
1041
|
+
/**
|
|
1042
|
+
* The temperature of the model.
|
|
1043
|
+
*/
|
|
1044
|
+
temperature?: number;
|
|
720
1045
|
}
|
|
721
1046
|
|
|
722
1047
|
/**
|
|
723
|
-
*
|
|
1048
|
+
* Together AI model configuration (OpenAI-compatible).
|
|
724
1049
|
*/
|
|
725
|
-
export interface
|
|
1050
|
+
export interface TogetherModelSettings {
|
|
1051
|
+
/**
|
|
1052
|
+
* The maximum number of tokens the model can generate.
|
|
1053
|
+
*/
|
|
1054
|
+
max_output_tokens?: number;
|
|
1055
|
+
|
|
726
1056
|
/**
|
|
727
|
-
*
|
|
1057
|
+
* Whether to enable parallel tool calling.
|
|
728
1058
|
*/
|
|
729
|
-
|
|
1059
|
+
parallel_tool_calls?: boolean;
|
|
730
1060
|
|
|
1061
|
+
/**
|
|
1062
|
+
* The provider of the model.
|
|
1063
|
+
*/
|
|
1064
|
+
provider?: 'together';
|
|
1065
|
+
|
|
1066
|
+
/**
|
|
1067
|
+
* The response format for the model.
|
|
1068
|
+
*/
|
|
1069
|
+
response_format?:
|
|
1070
|
+
| AgentsAPI.TextResponseFormat
|
|
1071
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1072
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1073
|
+
| null;
|
|
1074
|
+
|
|
1075
|
+
/**
|
|
1076
|
+
* The temperature of the model.
|
|
1077
|
+
*/
|
|
1078
|
+
temperature?: number;
|
|
1079
|
+
}
|
|
1080
|
+
|
|
1081
|
+
/**
|
|
1082
|
+
* AWS Bedrock model configuration.
|
|
1083
|
+
*/
|
|
1084
|
+
export interface BedrockModelSettings {
|
|
731
1085
|
/**
|
|
732
1086
|
* The maximum number of tokens the model can generate.
|
|
733
1087
|
*/
|
|
734
1088
|
max_output_tokens?: number;
|
|
1089
|
+
|
|
1090
|
+
/**
|
|
1091
|
+
* Whether to enable parallel tool calling.
|
|
1092
|
+
*/
|
|
1093
|
+
parallel_tool_calls?: boolean;
|
|
1094
|
+
|
|
1095
|
+
/**
|
|
1096
|
+
* The provider of the model.
|
|
1097
|
+
*/
|
|
1098
|
+
provider?: 'bedrock';
|
|
1099
|
+
|
|
1100
|
+
/**
|
|
1101
|
+
* The response format for the model.
|
|
1102
|
+
*/
|
|
1103
|
+
response_format?:
|
|
1104
|
+
| AgentsAPI.TextResponseFormat
|
|
1105
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1106
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1107
|
+
| null;
|
|
1108
|
+
|
|
1109
|
+
/**
|
|
1110
|
+
* The temperature of the model.
|
|
1111
|
+
*/
|
|
1112
|
+
temperature?: number;
|
|
735
1113
|
}
|
|
736
1114
|
}
|
|
737
1115
|
|
|
@@ -1111,10 +1489,9 @@ export interface AgentCreateParams {
|
|
|
1111
1489
|
description?: string | null;
|
|
1112
1490
|
|
|
1113
1491
|
/**
|
|
1114
|
-
* The embedding
|
|
1115
|
-
* provider/model-name.
|
|
1492
|
+
* The embedding model handle used by the agent (format: provider/model-name).
|
|
1116
1493
|
*/
|
|
1117
|
-
embedding?: string |
|
|
1494
|
+
embedding?: string | null;
|
|
1118
1495
|
|
|
1119
1496
|
/**
|
|
1120
1497
|
* @deprecated Deprecated: No longer used. The embedding chunk size used by the
|
|
@@ -1230,10 +1607,25 @@ export interface AgentCreateParams {
|
|
|
1230
1607
|
metadata?: { [key: string]: unknown } | null;
|
|
1231
1608
|
|
|
1232
1609
|
/**
|
|
1233
|
-
* The model handle
|
|
1234
|
-
|
|
1610
|
+
* The model handle for the agent to use (format: provider/model-name).
|
|
1611
|
+
*/
|
|
1612
|
+
model?: string | null;
|
|
1613
|
+
|
|
1614
|
+
/**
|
|
1615
|
+
* The model settings for the agent.
|
|
1235
1616
|
*/
|
|
1236
|
-
|
|
1617
|
+
model_settings?:
|
|
1618
|
+
| AgentCreateParams.OpenAIModelSettings
|
|
1619
|
+
| AgentCreateParams.AnthropicModelSettings
|
|
1620
|
+
| AgentCreateParams.GoogleAIModelSettings
|
|
1621
|
+
| AgentCreateParams.GoogleVertexModelSettings
|
|
1622
|
+
| AgentCreateParams.AzureModelSettings
|
|
1623
|
+
| AgentCreateParams.XaiModelSettings
|
|
1624
|
+
| AgentCreateParams.GroqModelSettings
|
|
1625
|
+
| AgentCreateParams.DeepseekModelSettings
|
|
1626
|
+
| AgentCreateParams.TogetherModelSettings
|
|
1627
|
+
| AgentCreateParams.BedrockModelSettings
|
|
1628
|
+
| null;
|
|
1237
1629
|
|
|
1238
1630
|
/**
|
|
1239
1631
|
* The name of the agent.
|
|
@@ -1345,39 +1737,419 @@ export interface AgentCreateParams {
|
|
|
1345
1737
|
}
|
|
1346
1738
|
|
|
1347
1739
|
export namespace AgentCreateParams {
|
|
1348
|
-
export interface
|
|
1740
|
+
export interface OpenAIModelSettings {
|
|
1741
|
+
/**
|
|
1742
|
+
* The maximum number of tokens the model can generate.
|
|
1743
|
+
*/
|
|
1744
|
+
max_output_tokens?: number;
|
|
1745
|
+
|
|
1349
1746
|
/**
|
|
1350
|
-
*
|
|
1747
|
+
* Whether to enable parallel tool calling.
|
|
1351
1748
|
*/
|
|
1352
|
-
|
|
1749
|
+
parallel_tool_calls?: boolean;
|
|
1353
1750
|
|
|
1354
1751
|
/**
|
|
1355
1752
|
* The provider of the model.
|
|
1356
1753
|
*/
|
|
1357
|
-
provider
|
|
1358
|
-
}
|
|
1754
|
+
provider?: 'openai';
|
|
1359
1755
|
|
|
1360
|
-
/**
|
|
1361
|
-
* Schema for defining settings for a model
|
|
1362
|
-
*/
|
|
1363
|
-
export interface ModelSettings {
|
|
1364
1756
|
/**
|
|
1365
|
-
* The
|
|
1757
|
+
* The reasoning configuration for the model.
|
|
1366
1758
|
*/
|
|
1367
|
-
|
|
1759
|
+
reasoning?: OpenAIModelSettings.Reasoning;
|
|
1368
1760
|
|
|
1369
1761
|
/**
|
|
1370
|
-
* The
|
|
1762
|
+
* The response format for the model.
|
|
1371
1763
|
*/
|
|
1372
|
-
|
|
1764
|
+
response_format?:
|
|
1765
|
+
| AgentsAPI.TextResponseFormat
|
|
1766
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1767
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1768
|
+
| null;
|
|
1769
|
+
|
|
1770
|
+
/**
|
|
1771
|
+
* The temperature of the model.
|
|
1772
|
+
*/
|
|
1773
|
+
temperature?: number;
|
|
1373
1774
|
}
|
|
1374
|
-
}
|
|
1375
1775
|
|
|
1376
|
-
export
|
|
1377
|
-
|
|
1378
|
-
|
|
1379
|
-
|
|
1380
|
-
|
|
1776
|
+
export namespace OpenAIModelSettings {
|
|
1777
|
+
/**
|
|
1778
|
+
* The reasoning configuration for the model.
|
|
1779
|
+
*/
|
|
1780
|
+
export interface Reasoning {
|
|
1781
|
+
/**
|
|
1782
|
+
* The reasoning effort to use when generating text reasoning models
|
|
1783
|
+
*/
|
|
1784
|
+
reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
|
|
1785
|
+
}
|
|
1786
|
+
}
|
|
1787
|
+
|
|
1788
|
+
export interface AnthropicModelSettings {
|
|
1789
|
+
/**
|
|
1790
|
+
* The maximum number of tokens the model can generate.
|
|
1791
|
+
*/
|
|
1792
|
+
max_output_tokens?: number;
|
|
1793
|
+
|
|
1794
|
+
/**
|
|
1795
|
+
* Whether to enable parallel tool calling.
|
|
1796
|
+
*/
|
|
1797
|
+
parallel_tool_calls?: boolean;
|
|
1798
|
+
|
|
1799
|
+
/**
|
|
1800
|
+
* The provider of the model.
|
|
1801
|
+
*/
|
|
1802
|
+
provider?: 'anthropic';
|
|
1803
|
+
|
|
1804
|
+
/**
|
|
1805
|
+
* The temperature of the model.
|
|
1806
|
+
*/
|
|
1807
|
+
temperature?: number;
|
|
1808
|
+
|
|
1809
|
+
/**
|
|
1810
|
+
* The thinking configuration for the model.
|
|
1811
|
+
*/
|
|
1812
|
+
thinking?: AnthropicModelSettings.Thinking;
|
|
1813
|
+
|
|
1814
|
+
/**
|
|
1815
|
+
* Soft control for how verbose model output should be, used for GPT-5 models.
|
|
1816
|
+
*/
|
|
1817
|
+
verbosity?: 'low' | 'medium' | 'high' | null;
|
|
1818
|
+
}
|
|
1819
|
+
|
|
1820
|
+
export namespace AnthropicModelSettings {
|
|
1821
|
+
/**
|
|
1822
|
+
* The thinking configuration for the model.
|
|
1823
|
+
*/
|
|
1824
|
+
export interface Thinking {
|
|
1825
|
+
/**
|
|
1826
|
+
* The maximum number of tokens the model can use for extended thinking.
|
|
1827
|
+
*/
|
|
1828
|
+
budget_tokens?: number;
|
|
1829
|
+
|
|
1830
|
+
/**
|
|
1831
|
+
* The type of thinking to use.
|
|
1832
|
+
*/
|
|
1833
|
+
type?: 'enabled' | 'disabled';
|
|
1834
|
+
}
|
|
1835
|
+
}
|
|
1836
|
+
|
|
1837
|
+
export interface GoogleAIModelSettings {
|
|
1838
|
+
/**
|
|
1839
|
+
* The maximum number of tokens the model can generate.
|
|
1840
|
+
*/
|
|
1841
|
+
max_output_tokens?: number;
|
|
1842
|
+
|
|
1843
|
+
/**
|
|
1844
|
+
* Whether to enable parallel tool calling.
|
|
1845
|
+
*/
|
|
1846
|
+
parallel_tool_calls?: boolean;
|
|
1847
|
+
|
|
1848
|
+
/**
|
|
1849
|
+
* The provider of the model.
|
|
1850
|
+
*/
|
|
1851
|
+
provider?: 'google_ai';
|
|
1852
|
+
|
|
1853
|
+
/**
|
|
1854
|
+
* The response schema for the model.
|
|
1855
|
+
*/
|
|
1856
|
+
response_schema?:
|
|
1857
|
+
| AgentsAPI.TextResponseFormat
|
|
1858
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1859
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1860
|
+
| null;
|
|
1861
|
+
|
|
1862
|
+
/**
|
|
1863
|
+
* The temperature of the model.
|
|
1864
|
+
*/
|
|
1865
|
+
temperature?: number;
|
|
1866
|
+
|
|
1867
|
+
/**
|
|
1868
|
+
* The thinking configuration for the model.
|
|
1869
|
+
*/
|
|
1870
|
+
thinking_config?: GoogleAIModelSettings.ThinkingConfig;
|
|
1871
|
+
}
|
|
1872
|
+
|
|
1873
|
+
export namespace GoogleAIModelSettings {
|
|
1874
|
+
/**
|
|
1875
|
+
* The thinking configuration for the model.
|
|
1876
|
+
*/
|
|
1877
|
+
export interface ThinkingConfig {
|
|
1878
|
+
/**
|
|
1879
|
+
* Whether to include thoughts in the model's response.
|
|
1880
|
+
*/
|
|
1881
|
+
include_thoughts?: boolean;
|
|
1882
|
+
|
|
1883
|
+
/**
|
|
1884
|
+
* The thinking budget for the model.
|
|
1885
|
+
*/
|
|
1886
|
+
thinking_budget?: number;
|
|
1887
|
+
}
|
|
1888
|
+
}
|
|
1889
|
+
|
|
1890
|
+
export interface GoogleVertexModelSettings {
|
|
1891
|
+
/**
|
|
1892
|
+
* The maximum number of tokens the model can generate.
|
|
1893
|
+
*/
|
|
1894
|
+
max_output_tokens?: number;
|
|
1895
|
+
|
|
1896
|
+
/**
|
|
1897
|
+
* Whether to enable parallel tool calling.
|
|
1898
|
+
*/
|
|
1899
|
+
parallel_tool_calls?: boolean;
|
|
1900
|
+
|
|
1901
|
+
/**
|
|
1902
|
+
* The provider of the model.
|
|
1903
|
+
*/
|
|
1904
|
+
provider?: 'google_vertex';
|
|
1905
|
+
|
|
1906
|
+
/**
|
|
1907
|
+
* The response schema for the model.
|
|
1908
|
+
*/
|
|
1909
|
+
response_schema?:
|
|
1910
|
+
| AgentsAPI.TextResponseFormat
|
|
1911
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1912
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1913
|
+
| null;
|
|
1914
|
+
|
|
1915
|
+
/**
|
|
1916
|
+
* The temperature of the model.
|
|
1917
|
+
*/
|
|
1918
|
+
temperature?: number;
|
|
1919
|
+
|
|
1920
|
+
/**
|
|
1921
|
+
* The thinking configuration for the model.
|
|
1922
|
+
*/
|
|
1923
|
+
thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
|
|
1924
|
+
}
|
|
1925
|
+
|
|
1926
|
+
export namespace GoogleVertexModelSettings {
|
|
1927
|
+
/**
|
|
1928
|
+
* The thinking configuration for the model.
|
|
1929
|
+
*/
|
|
1930
|
+
export interface ThinkingConfig {
|
|
1931
|
+
/**
|
|
1932
|
+
* Whether to include thoughts in the model's response.
|
|
1933
|
+
*/
|
|
1934
|
+
include_thoughts?: boolean;
|
|
1935
|
+
|
|
1936
|
+
/**
|
|
1937
|
+
* The thinking budget for the model.
|
|
1938
|
+
*/
|
|
1939
|
+
thinking_budget?: number;
|
|
1940
|
+
}
|
|
1941
|
+
}
|
|
1942
|
+
|
|
1943
|
+
/**
|
|
1944
|
+
* Azure OpenAI model configuration (OpenAI-compatible).
|
|
1945
|
+
*/
|
|
1946
|
+
export interface AzureModelSettings {
|
|
1947
|
+
/**
|
|
1948
|
+
* The maximum number of tokens the model can generate.
|
|
1949
|
+
*/
|
|
1950
|
+
max_output_tokens?: number;
|
|
1951
|
+
|
|
1952
|
+
/**
|
|
1953
|
+
* Whether to enable parallel tool calling.
|
|
1954
|
+
*/
|
|
1955
|
+
parallel_tool_calls?: boolean;
|
|
1956
|
+
|
|
1957
|
+
/**
|
|
1958
|
+
* The provider of the model.
|
|
1959
|
+
*/
|
|
1960
|
+
provider?: 'azure';
|
|
1961
|
+
|
|
1962
|
+
/**
|
|
1963
|
+
* The response format for the model.
|
|
1964
|
+
*/
|
|
1965
|
+
response_format?:
|
|
1966
|
+
| AgentsAPI.TextResponseFormat
|
|
1967
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1968
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1969
|
+
| null;
|
|
1970
|
+
|
|
1971
|
+
/**
|
|
1972
|
+
* The temperature of the model.
|
|
1973
|
+
*/
|
|
1974
|
+
temperature?: number;
|
|
1975
|
+
}
|
|
1976
|
+
|
|
1977
|
+
/**
|
|
1978
|
+
* xAI model configuration (OpenAI-compatible).
|
|
1979
|
+
*/
|
|
1980
|
+
export interface XaiModelSettings {
|
|
1981
|
+
/**
|
|
1982
|
+
* The maximum number of tokens the model can generate.
|
|
1983
|
+
*/
|
|
1984
|
+
max_output_tokens?: number;
|
|
1985
|
+
|
|
1986
|
+
/**
|
|
1987
|
+
* Whether to enable parallel tool calling.
|
|
1988
|
+
*/
|
|
1989
|
+
parallel_tool_calls?: boolean;
|
|
1990
|
+
|
|
1991
|
+
/**
|
|
1992
|
+
* The provider of the model.
|
|
1993
|
+
*/
|
|
1994
|
+
provider?: 'xai';
|
|
1995
|
+
|
|
1996
|
+
/**
|
|
1997
|
+
* The response format for the model.
|
|
1998
|
+
*/
|
|
1999
|
+
response_format?:
|
|
2000
|
+
| AgentsAPI.TextResponseFormat
|
|
2001
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2002
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2003
|
+
| null;
|
|
2004
|
+
|
|
2005
|
+
/**
|
|
2006
|
+
* The temperature of the model.
|
|
2007
|
+
*/
|
|
2008
|
+
temperature?: number;
|
|
2009
|
+
}
|
|
2010
|
+
|
|
2011
|
+
/**
|
|
2012
|
+
* Groq model configuration (OpenAI-compatible).
|
|
2013
|
+
*/
|
|
2014
|
+
export interface GroqModelSettings {
|
|
2015
|
+
/**
|
|
2016
|
+
* The maximum number of tokens the model can generate.
|
|
2017
|
+
*/
|
|
2018
|
+
max_output_tokens?: number;
|
|
2019
|
+
|
|
2020
|
+
/**
|
|
2021
|
+
* Whether to enable parallel tool calling.
|
|
2022
|
+
*/
|
|
2023
|
+
parallel_tool_calls?: boolean;
|
|
2024
|
+
|
|
2025
|
+
/**
|
|
2026
|
+
* The provider of the model.
|
|
2027
|
+
*/
|
|
2028
|
+
provider?: 'groq';
|
|
2029
|
+
|
|
2030
|
+
/**
|
|
2031
|
+
* The response format for the model.
|
|
2032
|
+
*/
|
|
2033
|
+
response_format?:
|
|
2034
|
+
| AgentsAPI.TextResponseFormat
|
|
2035
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2036
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2037
|
+
| null;
|
|
2038
|
+
|
|
2039
|
+
/**
|
|
2040
|
+
* The temperature of the model.
|
|
2041
|
+
*/
|
|
2042
|
+
temperature?: number;
|
|
2043
|
+
}
|
|
2044
|
+
|
|
2045
|
+
/**
|
|
2046
|
+
* Deepseek model configuration (OpenAI-compatible).
|
|
2047
|
+
*/
|
|
2048
|
+
export interface DeepseekModelSettings {
|
|
2049
|
+
/**
|
|
2050
|
+
* The maximum number of tokens the model can generate.
|
|
2051
|
+
*/
|
|
2052
|
+
max_output_tokens?: number;
|
|
2053
|
+
|
|
2054
|
+
/**
|
|
2055
|
+
* Whether to enable parallel tool calling.
|
|
2056
|
+
*/
|
|
2057
|
+
parallel_tool_calls?: boolean;
|
|
2058
|
+
|
|
2059
|
+
/**
|
|
2060
|
+
* The provider of the model.
|
|
2061
|
+
*/
|
|
2062
|
+
provider?: 'deepseek';
|
|
2063
|
+
|
|
2064
|
+
/**
|
|
2065
|
+
* The response format for the model.
|
|
2066
|
+
*/
|
|
2067
|
+
response_format?:
|
|
2068
|
+
| AgentsAPI.TextResponseFormat
|
|
2069
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2070
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2071
|
+
| null;
|
|
2072
|
+
|
|
2073
|
+
/**
|
|
2074
|
+
* The temperature of the model.
|
|
2075
|
+
*/
|
|
2076
|
+
temperature?: number;
|
|
2077
|
+
}
|
|
2078
|
+
|
|
2079
|
+
/**
|
|
2080
|
+
* Together AI model configuration (OpenAI-compatible).
|
|
2081
|
+
*/
|
|
2082
|
+
export interface TogetherModelSettings {
|
|
2083
|
+
/**
|
|
2084
|
+
* The maximum number of tokens the model can generate.
|
|
2085
|
+
*/
|
|
2086
|
+
max_output_tokens?: number;
|
|
2087
|
+
|
|
2088
|
+
/**
|
|
2089
|
+
* Whether to enable parallel tool calling.
|
|
2090
|
+
*/
|
|
2091
|
+
parallel_tool_calls?: boolean;
|
|
2092
|
+
|
|
2093
|
+
/**
|
|
2094
|
+
* The provider of the model.
|
|
2095
|
+
*/
|
|
2096
|
+
provider?: 'together';
|
|
2097
|
+
|
|
2098
|
+
/**
|
|
2099
|
+
* The response format for the model.
|
|
2100
|
+
*/
|
|
2101
|
+
response_format?:
|
|
2102
|
+
| AgentsAPI.TextResponseFormat
|
|
2103
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2104
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2105
|
+
| null;
|
|
2106
|
+
|
|
2107
|
+
/**
|
|
2108
|
+
* The temperature of the model.
|
|
2109
|
+
*/
|
|
2110
|
+
temperature?: number;
|
|
2111
|
+
}
|
|
2112
|
+
|
|
2113
|
+
/**
|
|
2114
|
+
* AWS Bedrock model configuration.
|
|
2115
|
+
*/
|
|
2116
|
+
export interface BedrockModelSettings {
|
|
2117
|
+
/**
|
|
2118
|
+
* The maximum number of tokens the model can generate.
|
|
2119
|
+
*/
|
|
2120
|
+
max_output_tokens?: number;
|
|
2121
|
+
|
|
2122
|
+
/**
|
|
2123
|
+
* Whether to enable parallel tool calling.
|
|
2124
|
+
*/
|
|
2125
|
+
parallel_tool_calls?: boolean;
|
|
2126
|
+
|
|
2127
|
+
/**
|
|
2128
|
+
* The provider of the model.
|
|
2129
|
+
*/
|
|
2130
|
+
provider?: 'bedrock';
|
|
2131
|
+
|
|
2132
|
+
/**
|
|
2133
|
+
* The response format for the model.
|
|
2134
|
+
*/
|
|
2135
|
+
response_format?:
|
|
2136
|
+
| AgentsAPI.TextResponseFormat
|
|
2137
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2138
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2139
|
+
| null;
|
|
2140
|
+
|
|
2141
|
+
/**
|
|
2142
|
+
* The temperature of the model.
|
|
2143
|
+
*/
|
|
2144
|
+
temperature?: number;
|
|
2145
|
+
}
|
|
2146
|
+
}
|
|
2147
|
+
|
|
2148
|
+
export interface AgentRetrieveParams {
|
|
2149
|
+
/**
|
|
2150
|
+
* Specify which relational fields to include in the response. No relationships are
|
|
2151
|
+
* included by default.
|
|
2152
|
+
*/
|
|
1381
2153
|
include?: Array<
|
|
1382
2154
|
| 'agent.blocks'
|
|
1383
2155
|
| 'agent.identities'
|
|
@@ -1571,10 +2343,9 @@ export interface AgentModifyParams {
|
|
|
1571
2343
|
description?: string | null;
|
|
1572
2344
|
|
|
1573
2345
|
/**
|
|
1574
|
-
* The embedding
|
|
1575
|
-
* provider/model-name.
|
|
2346
|
+
* The embedding model handle used by the agent (format: provider/model-name).
|
|
1576
2347
|
*/
|
|
1577
|
-
embedding?: string |
|
|
2348
|
+
embedding?: string | null;
|
|
1578
2349
|
|
|
1579
2350
|
/**
|
|
1580
2351
|
* Configuration for embedding model connection and processing parameters.
|
|
@@ -1646,10 +2417,25 @@ export interface AgentModifyParams {
|
|
|
1646
2417
|
metadata?: { [key: string]: unknown } | null;
|
|
1647
2418
|
|
|
1648
2419
|
/**
|
|
1649
|
-
* The model used by the agent
|
|
1650
|
-
* model schema for more information.
|
|
2420
|
+
* The model handle used by the agent (format: provider/model-name).
|
|
1651
2421
|
*/
|
|
1652
|
-
model?: string |
|
|
2422
|
+
model?: string | null;
|
|
2423
|
+
|
|
2424
|
+
/**
|
|
2425
|
+
* The model settings for the agent.
|
|
2426
|
+
*/
|
|
2427
|
+
model_settings?:
|
|
2428
|
+
| AgentModifyParams.OpenAIModelSettings
|
|
2429
|
+
| AgentModifyParams.AnthropicModelSettings
|
|
2430
|
+
| AgentModifyParams.GoogleAIModelSettings
|
|
2431
|
+
| AgentModifyParams.GoogleVertexModelSettings
|
|
2432
|
+
| AgentModifyParams.AzureModelSettings
|
|
2433
|
+
| AgentModifyParams.XaiModelSettings
|
|
2434
|
+
| AgentModifyParams.GroqModelSettings
|
|
2435
|
+
| AgentModifyParams.DeepseekModelSettings
|
|
2436
|
+
| AgentModifyParams.TogetherModelSettings
|
|
2437
|
+
| AgentModifyParams.BedrockModelSettings
|
|
2438
|
+
| null;
|
|
1653
2439
|
|
|
1654
2440
|
/**
|
|
1655
2441
|
* The name of the agent.
|
|
@@ -1742,31 +2528,411 @@ export interface AgentModifyParams {
|
|
|
1742
2528
|
}
|
|
1743
2529
|
|
|
1744
2530
|
export namespace AgentModifyParams {
|
|
1745
|
-
export interface
|
|
2531
|
+
export interface OpenAIModelSettings {
|
|
2532
|
+
/**
|
|
2533
|
+
* The maximum number of tokens the model can generate.
|
|
2534
|
+
*/
|
|
2535
|
+
max_output_tokens?: number;
|
|
2536
|
+
|
|
2537
|
+
/**
|
|
2538
|
+
* Whether to enable parallel tool calling.
|
|
2539
|
+
*/
|
|
2540
|
+
parallel_tool_calls?: boolean;
|
|
2541
|
+
|
|
2542
|
+
/**
|
|
2543
|
+
* The provider of the model.
|
|
2544
|
+
*/
|
|
2545
|
+
provider?: 'openai';
|
|
2546
|
+
|
|
2547
|
+
/**
|
|
2548
|
+
* The reasoning configuration for the model.
|
|
2549
|
+
*/
|
|
2550
|
+
reasoning?: OpenAIModelSettings.Reasoning;
|
|
2551
|
+
|
|
2552
|
+
/**
|
|
2553
|
+
* The response format for the model.
|
|
2554
|
+
*/
|
|
2555
|
+
response_format?:
|
|
2556
|
+
| AgentsAPI.TextResponseFormat
|
|
2557
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2558
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2559
|
+
| null;
|
|
2560
|
+
|
|
2561
|
+
/**
|
|
2562
|
+
* The temperature of the model.
|
|
2563
|
+
*/
|
|
2564
|
+
temperature?: number;
|
|
2565
|
+
}
|
|
2566
|
+
|
|
2567
|
+
export namespace OpenAIModelSettings {
|
|
2568
|
+
/**
|
|
2569
|
+
* The reasoning configuration for the model.
|
|
2570
|
+
*/
|
|
2571
|
+
export interface Reasoning {
|
|
2572
|
+
/**
|
|
2573
|
+
* The reasoning effort to use when generating text reasoning models
|
|
2574
|
+
*/
|
|
2575
|
+
reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
|
|
2576
|
+
}
|
|
2577
|
+
}
|
|
2578
|
+
|
|
2579
|
+
export interface AnthropicModelSettings {
|
|
2580
|
+
/**
|
|
2581
|
+
* The maximum number of tokens the model can generate.
|
|
2582
|
+
*/
|
|
2583
|
+
max_output_tokens?: number;
|
|
2584
|
+
|
|
2585
|
+
/**
|
|
2586
|
+
* Whether to enable parallel tool calling.
|
|
2587
|
+
*/
|
|
2588
|
+
parallel_tool_calls?: boolean;
|
|
2589
|
+
|
|
2590
|
+
/**
|
|
2591
|
+
* The provider of the model.
|
|
2592
|
+
*/
|
|
2593
|
+
provider?: 'anthropic';
|
|
2594
|
+
|
|
2595
|
+
/**
|
|
2596
|
+
* The temperature of the model.
|
|
2597
|
+
*/
|
|
2598
|
+
temperature?: number;
|
|
2599
|
+
|
|
2600
|
+
/**
|
|
2601
|
+
* The thinking configuration for the model.
|
|
2602
|
+
*/
|
|
2603
|
+
thinking?: AnthropicModelSettings.Thinking;
|
|
2604
|
+
|
|
2605
|
+
/**
|
|
2606
|
+
* Soft control for how verbose model output should be, used for GPT-5 models.
|
|
2607
|
+
*/
|
|
2608
|
+
verbosity?: 'low' | 'medium' | 'high' | null;
|
|
2609
|
+
}
|
|
2610
|
+
|
|
2611
|
+
export namespace AnthropicModelSettings {
|
|
2612
|
+
/**
|
|
2613
|
+
* The thinking configuration for the model.
|
|
2614
|
+
*/
|
|
2615
|
+
export interface Thinking {
|
|
2616
|
+
/**
|
|
2617
|
+
* The maximum number of tokens the model can use for extended thinking.
|
|
2618
|
+
*/
|
|
2619
|
+
budget_tokens?: number;
|
|
2620
|
+
|
|
2621
|
+
/**
|
|
2622
|
+
* The type of thinking to use.
|
|
2623
|
+
*/
|
|
2624
|
+
type?: 'enabled' | 'disabled';
|
|
2625
|
+
}
|
|
2626
|
+
}
|
|
2627
|
+
|
|
2628
|
+
export interface GoogleAIModelSettings {
|
|
1746
2629
|
/**
|
|
1747
|
-
* The
|
|
2630
|
+
* The maximum number of tokens the model can generate.
|
|
1748
2631
|
*/
|
|
1749
|
-
|
|
2632
|
+
max_output_tokens?: number;
|
|
2633
|
+
|
|
2634
|
+
/**
|
|
2635
|
+
* Whether to enable parallel tool calling.
|
|
2636
|
+
*/
|
|
2637
|
+
parallel_tool_calls?: boolean;
|
|
1750
2638
|
|
|
1751
2639
|
/**
|
|
1752
2640
|
* The provider of the model.
|
|
1753
2641
|
*/
|
|
1754
|
-
provider
|
|
2642
|
+
provider?: 'google_ai';
|
|
2643
|
+
|
|
2644
|
+
/**
|
|
2645
|
+
* The response schema for the model.
|
|
2646
|
+
*/
|
|
2647
|
+
response_schema?:
|
|
2648
|
+
| AgentsAPI.TextResponseFormat
|
|
2649
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2650
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2651
|
+
| null;
|
|
2652
|
+
|
|
2653
|
+
/**
|
|
2654
|
+
* The temperature of the model.
|
|
2655
|
+
*/
|
|
2656
|
+
temperature?: number;
|
|
2657
|
+
|
|
2658
|
+
/**
|
|
2659
|
+
* The thinking configuration for the model.
|
|
2660
|
+
*/
|
|
2661
|
+
thinking_config?: GoogleAIModelSettings.ThinkingConfig;
|
|
2662
|
+
}
|
|
2663
|
+
|
|
2664
|
+
export namespace GoogleAIModelSettings {
|
|
2665
|
+
/**
|
|
2666
|
+
* The thinking configuration for the model.
|
|
2667
|
+
*/
|
|
2668
|
+
export interface ThinkingConfig {
|
|
2669
|
+
/**
|
|
2670
|
+
* Whether to include thoughts in the model's response.
|
|
2671
|
+
*/
|
|
2672
|
+
include_thoughts?: boolean;
|
|
2673
|
+
|
|
2674
|
+
/**
|
|
2675
|
+
* The thinking budget for the model.
|
|
2676
|
+
*/
|
|
2677
|
+
thinking_budget?: number;
|
|
2678
|
+
}
|
|
2679
|
+
}
|
|
2680
|
+
|
|
2681
|
+
export interface GoogleVertexModelSettings {
|
|
2682
|
+
/**
|
|
2683
|
+
* The maximum number of tokens the model can generate.
|
|
2684
|
+
*/
|
|
2685
|
+
max_output_tokens?: number;
|
|
2686
|
+
|
|
2687
|
+
/**
|
|
2688
|
+
* Whether to enable parallel tool calling.
|
|
2689
|
+
*/
|
|
2690
|
+
parallel_tool_calls?: boolean;
|
|
2691
|
+
|
|
2692
|
+
/**
|
|
2693
|
+
* The provider of the model.
|
|
2694
|
+
*/
|
|
2695
|
+
provider?: 'google_vertex';
|
|
2696
|
+
|
|
2697
|
+
/**
|
|
2698
|
+
* The response schema for the model.
|
|
2699
|
+
*/
|
|
2700
|
+
response_schema?:
|
|
2701
|
+
| AgentsAPI.TextResponseFormat
|
|
2702
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2703
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2704
|
+
| null;
|
|
2705
|
+
|
|
2706
|
+
/**
|
|
2707
|
+
* The temperature of the model.
|
|
2708
|
+
*/
|
|
2709
|
+
temperature?: number;
|
|
2710
|
+
|
|
2711
|
+
/**
|
|
2712
|
+
* The thinking configuration for the model.
|
|
2713
|
+
*/
|
|
2714
|
+
thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
|
|
2715
|
+
}
|
|
2716
|
+
|
|
2717
|
+
export namespace GoogleVertexModelSettings {
|
|
2718
|
+
/**
|
|
2719
|
+
* The thinking configuration for the model.
|
|
2720
|
+
*/
|
|
2721
|
+
export interface ThinkingConfig {
|
|
2722
|
+
/**
|
|
2723
|
+
* Whether to include thoughts in the model's response.
|
|
2724
|
+
*/
|
|
2725
|
+
include_thoughts?: boolean;
|
|
2726
|
+
|
|
2727
|
+
/**
|
|
2728
|
+
* The thinking budget for the model.
|
|
2729
|
+
*/
|
|
2730
|
+
thinking_budget?: number;
|
|
2731
|
+
}
|
|
1755
2732
|
}
|
|
1756
2733
|
|
|
1757
2734
|
/**
|
|
1758
|
-
*
|
|
2735
|
+
* Azure OpenAI model configuration (OpenAI-compatible).
|
|
1759
2736
|
*/
|
|
1760
|
-
export interface
|
|
2737
|
+
export interface AzureModelSettings {
|
|
2738
|
+
/**
|
|
2739
|
+
* The maximum number of tokens the model can generate.
|
|
2740
|
+
*/
|
|
2741
|
+
max_output_tokens?: number;
|
|
2742
|
+
|
|
2743
|
+
/**
|
|
2744
|
+
* Whether to enable parallel tool calling.
|
|
2745
|
+
*/
|
|
2746
|
+
parallel_tool_calls?: boolean;
|
|
2747
|
+
|
|
2748
|
+
/**
|
|
2749
|
+
* The provider of the model.
|
|
2750
|
+
*/
|
|
2751
|
+
provider?: 'azure';
|
|
2752
|
+
|
|
2753
|
+
/**
|
|
2754
|
+
* The response format for the model.
|
|
2755
|
+
*/
|
|
2756
|
+
response_format?:
|
|
2757
|
+
| AgentsAPI.TextResponseFormat
|
|
2758
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2759
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2760
|
+
| null;
|
|
2761
|
+
|
|
1761
2762
|
/**
|
|
1762
|
-
* The
|
|
2763
|
+
* The temperature of the model.
|
|
1763
2764
|
*/
|
|
1764
|
-
|
|
2765
|
+
temperature?: number;
|
|
2766
|
+
}
|
|
1765
2767
|
|
|
2768
|
+
/**
|
|
2769
|
+
* xAI model configuration (OpenAI-compatible).
|
|
2770
|
+
*/
|
|
2771
|
+
export interface XaiModelSettings {
|
|
1766
2772
|
/**
|
|
1767
2773
|
* The maximum number of tokens the model can generate.
|
|
1768
2774
|
*/
|
|
1769
2775
|
max_output_tokens?: number;
|
|
2776
|
+
|
|
2777
|
+
/**
|
|
2778
|
+
* Whether to enable parallel tool calling.
|
|
2779
|
+
*/
|
|
2780
|
+
parallel_tool_calls?: boolean;
|
|
2781
|
+
|
|
2782
|
+
/**
|
|
2783
|
+
* The provider of the model.
|
|
2784
|
+
*/
|
|
2785
|
+
provider?: 'xai';
|
|
2786
|
+
|
|
2787
|
+
/**
|
|
2788
|
+
* The response format for the model.
|
|
2789
|
+
*/
|
|
2790
|
+
response_format?:
|
|
2791
|
+
| AgentsAPI.TextResponseFormat
|
|
2792
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2793
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2794
|
+
| null;
|
|
2795
|
+
|
|
2796
|
+
/**
|
|
2797
|
+
* The temperature of the model.
|
|
2798
|
+
*/
|
|
2799
|
+
temperature?: number;
|
|
2800
|
+
}
|
|
2801
|
+
|
|
2802
|
+
/**
|
|
2803
|
+
* Groq model configuration (OpenAI-compatible).
|
|
2804
|
+
*/
|
|
2805
|
+
export interface GroqModelSettings {
|
|
2806
|
+
/**
|
|
2807
|
+
* The maximum number of tokens the model can generate.
|
|
2808
|
+
*/
|
|
2809
|
+
max_output_tokens?: number;
|
|
2810
|
+
|
|
2811
|
+
/**
|
|
2812
|
+
* Whether to enable parallel tool calling.
|
|
2813
|
+
*/
|
|
2814
|
+
parallel_tool_calls?: boolean;
|
|
2815
|
+
|
|
2816
|
+
/**
|
|
2817
|
+
* The provider of the model.
|
|
2818
|
+
*/
|
|
2819
|
+
provider?: 'groq';
|
|
2820
|
+
|
|
2821
|
+
/**
|
|
2822
|
+
* The response format for the model.
|
|
2823
|
+
*/
|
|
2824
|
+
response_format?:
|
|
2825
|
+
| AgentsAPI.TextResponseFormat
|
|
2826
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2827
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2828
|
+
| null;
|
|
2829
|
+
|
|
2830
|
+
/**
|
|
2831
|
+
* The temperature of the model.
|
|
2832
|
+
*/
|
|
2833
|
+
temperature?: number;
|
|
2834
|
+
}
|
|
2835
|
+
|
|
2836
|
+
/**
|
|
2837
|
+
* Deepseek model configuration (OpenAI-compatible).
|
|
2838
|
+
*/
|
|
2839
|
+
export interface DeepseekModelSettings {
|
|
2840
|
+
/**
|
|
2841
|
+
* The maximum number of tokens the model can generate.
|
|
2842
|
+
*/
|
|
2843
|
+
max_output_tokens?: number;
|
|
2844
|
+
|
|
2845
|
+
/**
|
|
2846
|
+
* Whether to enable parallel tool calling.
|
|
2847
|
+
*/
|
|
2848
|
+
parallel_tool_calls?: boolean;
|
|
2849
|
+
|
|
2850
|
+
/**
|
|
2851
|
+
* The provider of the model.
|
|
2852
|
+
*/
|
|
2853
|
+
provider?: 'deepseek';
|
|
2854
|
+
|
|
2855
|
+
/**
|
|
2856
|
+
* The response format for the model.
|
|
2857
|
+
*/
|
|
2858
|
+
response_format?:
|
|
2859
|
+
| AgentsAPI.TextResponseFormat
|
|
2860
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2861
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2862
|
+
| null;
|
|
2863
|
+
|
|
2864
|
+
/**
|
|
2865
|
+
* The temperature of the model.
|
|
2866
|
+
*/
|
|
2867
|
+
temperature?: number;
|
|
2868
|
+
}
|
|
2869
|
+
|
|
2870
|
+
/**
|
|
2871
|
+
* Together AI model configuration (OpenAI-compatible).
|
|
2872
|
+
*/
|
|
2873
|
+
export interface TogetherModelSettings {
|
|
2874
|
+
/**
|
|
2875
|
+
* The maximum number of tokens the model can generate.
|
|
2876
|
+
*/
|
|
2877
|
+
max_output_tokens?: number;
|
|
2878
|
+
|
|
2879
|
+
/**
|
|
2880
|
+
* Whether to enable parallel tool calling.
|
|
2881
|
+
*/
|
|
2882
|
+
parallel_tool_calls?: boolean;
|
|
2883
|
+
|
|
2884
|
+
/**
|
|
2885
|
+
* The provider of the model.
|
|
2886
|
+
*/
|
|
2887
|
+
provider?: 'together';
|
|
2888
|
+
|
|
2889
|
+
/**
|
|
2890
|
+
* The response format for the model.
|
|
2891
|
+
*/
|
|
2892
|
+
response_format?:
|
|
2893
|
+
| AgentsAPI.TextResponseFormat
|
|
2894
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2895
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2896
|
+
| null;
|
|
2897
|
+
|
|
2898
|
+
/**
|
|
2899
|
+
* The temperature of the model.
|
|
2900
|
+
*/
|
|
2901
|
+
temperature?: number;
|
|
2902
|
+
}
|
|
2903
|
+
|
|
2904
|
+
/**
|
|
2905
|
+
* AWS Bedrock model configuration.
|
|
2906
|
+
*/
|
|
2907
|
+
export interface BedrockModelSettings {
|
|
2908
|
+
/**
|
|
2909
|
+
* The maximum number of tokens the model can generate.
|
|
2910
|
+
*/
|
|
2911
|
+
max_output_tokens?: number;
|
|
2912
|
+
|
|
2913
|
+
/**
|
|
2914
|
+
* Whether to enable parallel tool calling.
|
|
2915
|
+
*/
|
|
2916
|
+
parallel_tool_calls?: boolean;
|
|
2917
|
+
|
|
2918
|
+
/**
|
|
2919
|
+
* The provider of the model.
|
|
2920
|
+
*/
|
|
2921
|
+
provider?: 'bedrock';
|
|
2922
|
+
|
|
2923
|
+
/**
|
|
2924
|
+
* The response format for the model.
|
|
2925
|
+
*/
|
|
2926
|
+
response_format?:
|
|
2927
|
+
| AgentsAPI.TextResponseFormat
|
|
2928
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2929
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2930
|
+
| null;
|
|
2931
|
+
|
|
2932
|
+
/**
|
|
2933
|
+
* The temperature of the model.
|
|
2934
|
+
*/
|
|
2935
|
+
temperature?: number;
|
|
1770
2936
|
}
|
|
1771
2937
|
}
|
|
1772
2938
|
|