@letta-ai/letta-client 1.0.0-alpha.16 → 1.0.0-alpha.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (93) hide show
  1. package/CHANGELOG.md +15 -0
  2. package/client.d.mts +2 -2
  3. package/client.d.mts.map +1 -1
  4. package/client.d.ts +2 -2
  5. package/client.d.ts.map +1 -1
  6. package/client.js.map +1 -1
  7. package/client.mjs.map +1 -1
  8. package/package.json +1 -1
  9. package/resources/agents/agents.d.mts +962 -111
  10. package/resources/agents/agents.d.mts.map +1 -1
  11. package/resources/agents/agents.d.ts +962 -111
  12. package/resources/agents/agents.d.ts.map +1 -1
  13. package/resources/agents/agents.js.map +1 -1
  14. package/resources/agents/agents.mjs.map +1 -1
  15. package/resources/agents/blocks.d.mts +1 -15
  16. package/resources/agents/blocks.d.mts.map +1 -1
  17. package/resources/agents/blocks.d.ts +1 -15
  18. package/resources/agents/blocks.d.ts.map +1 -1
  19. package/resources/agents/folders.d.mts +2 -7
  20. package/resources/agents/folders.d.mts.map +1 -1
  21. package/resources/agents/folders.d.ts +2 -7
  22. package/resources/agents/folders.d.ts.map +1 -1
  23. package/resources/agents/messages.d.mts +0 -10
  24. package/resources/agents/messages.d.mts.map +1 -1
  25. package/resources/agents/messages.d.ts +0 -10
  26. package/resources/agents/messages.d.ts.map +1 -1
  27. package/resources/archives.d.mts +0 -6
  28. package/resources/archives.d.mts.map +1 -1
  29. package/resources/archives.d.ts +0 -6
  30. package/resources/archives.d.ts.map +1 -1
  31. package/resources/folders/folders.d.mts +0 -6
  32. package/resources/folders/folders.d.mts.map +1 -1
  33. package/resources/folders/folders.d.ts +0 -6
  34. package/resources/folders/folders.d.ts.map +1 -1
  35. package/resources/folders/folders.js.map +1 -1
  36. package/resources/folders/folders.mjs.map +1 -1
  37. package/resources/index.d.mts +1 -1
  38. package/resources/index.d.mts.map +1 -1
  39. package/resources/index.d.ts +1 -1
  40. package/resources/index.d.ts.map +1 -1
  41. package/resources/index.js.map +1 -1
  42. package/resources/index.mjs.map +1 -1
  43. package/resources/mcp-servers/mcp-servers.d.mts +0 -10
  44. package/resources/mcp-servers/mcp-servers.d.mts.map +1 -1
  45. package/resources/mcp-servers/mcp-servers.d.ts +0 -10
  46. package/resources/mcp-servers/mcp-servers.d.ts.map +1 -1
  47. package/resources/mcp-servers/mcp-servers.js.map +1 -1
  48. package/resources/mcp-servers/mcp-servers.mjs.map +1 -1
  49. package/resources/models/embeddings.d.mts +1 -66
  50. package/resources/models/embeddings.d.mts.map +1 -1
  51. package/resources/models/embeddings.d.ts +1 -66
  52. package/resources/models/embeddings.d.ts.map +1 -1
  53. package/resources/models/index.d.mts +1 -1
  54. package/resources/models/index.d.mts.map +1 -1
  55. package/resources/models/index.d.ts +1 -1
  56. package/resources/models/index.d.ts.map +1 -1
  57. package/resources/models/index.js.map +1 -1
  58. package/resources/models/index.mjs.map +1 -1
  59. package/resources/models/models.d.mts +169 -109
  60. package/resources/models/models.d.mts.map +1 -1
  61. package/resources/models/models.d.ts +169 -109
  62. package/resources/models/models.d.ts.map +1 -1
  63. package/resources/models/models.js.map +1 -1
  64. package/resources/models/models.mjs.map +1 -1
  65. package/resources/runs/runs.d.mts +0 -5
  66. package/resources/runs/runs.d.mts.map +1 -1
  67. package/resources/runs/runs.d.ts +0 -5
  68. package/resources/runs/runs.d.ts.map +1 -1
  69. package/resources/runs/runs.js.map +1 -1
  70. package/resources/runs/runs.mjs.map +1 -1
  71. package/resources/tools.d.mts +0 -4
  72. package/resources/tools.d.mts.map +1 -1
  73. package/resources/tools.d.ts +0 -4
  74. package/resources/tools.d.ts.map +1 -1
  75. package/src/client.ts +4 -0
  76. package/src/resources/agents/agents.ts +1235 -69
  77. package/src/resources/agents/blocks.ts +1 -15
  78. package/src/resources/agents/folders.ts +2 -7
  79. package/src/resources/agents/messages.ts +0 -10
  80. package/src/resources/archives.ts +0 -6
  81. package/src/resources/folders/folders.ts +0 -6
  82. package/src/resources/index.ts +2 -0
  83. package/src/resources/mcp-servers/mcp-servers.ts +0 -10
  84. package/src/resources/models/embeddings.ts +1 -100
  85. package/src/resources/models/index.ts +2 -0
  86. package/src/resources/models/models.ts +249 -153
  87. package/src/resources/runs/runs.ts +0 -5
  88. package/src/resources/tools.ts +0 -4
  89. package/src/version.ts +1 -1
  90. package/version.d.mts +1 -1
  91. package/version.d.ts +1 -1
  92. package/version.js +1 -1
  93. package/version.mjs +1 -1
@@ -254,16 +254,6 @@ export interface AgentEnvironmentVariable {
254
254
  * Representation of an agent's state. This is the state of the agent at a given
255
255
  * time, and is persisted in the DB backend. The state has all the information
256
256
  * needed to recreate a persisted agent.
257
- *
258
- * Parameters: id (str): The unique identifier of the agent. name (str): The name
259
- * of the agent (must be unique to the user). created_at (datetime): The datetime
260
- * the agent was created. message_ids (List[str]): The ids of the messages in the
261
- * agent's in-context memory. memory (Memory): The in-context memory of the agent.
262
- * tools (List[str]): The tools used by the agent. This includes any memory editing
263
- * functions specified in `memory`. system (str): The system prompt used by the
264
- * agent. llm_config (LLMConfig): The LLM configuration used by the agent.
265
- * embedding_config (EmbeddingConfig): The embedding configuration used by the
266
- * agent.
267
257
  */
268
258
  export interface AgentState {
269
259
  /**
@@ -350,9 +340,9 @@ export interface AgentState {
350
340
  description?: string | null;
351
341
 
352
342
  /**
353
- * The embedding model used by the agent.
343
+ * The embedding model handle used by the agent (format: provider/model-name).
354
344
  */
355
- embedding?: AgentState.Embedding | null;
345
+ embedding?: string | null;
356
346
 
357
347
  /**
358
348
  * If set to True, memory management will move to a background agent thread.
@@ -429,9 +419,25 @@ export interface AgentState {
429
419
  metadata?: { [key: string]: unknown } | null;
430
420
 
431
421
  /**
432
- * Schema for defining settings for a model
422
+ * The model handle used by the agent (format: provider/model-name).
433
423
  */
434
- model?: AgentState.Model | null;
424
+ model?: string | null;
425
+
426
+ /**
427
+ * The model settings used by the agent.
428
+ */
429
+ model_settings?:
430
+ | AgentState.OpenAIModelSettings
431
+ | AgentState.AnthropicModelSettings
432
+ | AgentState.GoogleAIModelSettings
433
+ | AgentState.GoogleVertexModelSettings
434
+ | AgentState.AzureModelSettings
435
+ | AgentState.XaiModelSettings
436
+ | AgentState.GroqModelSettings
437
+ | AgentState.DeepseekModelSettings
438
+ | AgentState.TogetherModelSettings
439
+ | AgentState.BedrockModelSettings
440
+ | null;
435
441
 
436
442
  /**
437
443
  * @deprecated Deprecated: Use `managed_group` field instead. The multi-agent group
@@ -639,13 +645,8 @@ export namespace AgentState {
639
645
  }
640
646
 
641
647
  /**
642
- * Representation of a source, which is a collection of files and passages.
643
- *
644
- * Parameters: id (str): The ID of the source name (str): The name of the source.
645
- * embedding_config (EmbeddingConfig): The embedding configuration used by the
646
- * source. user_id (str): The ID of the user that created the source. metadata
647
- * (dict): Metadata associated with the source. description (str): The description
648
- * of the source.
648
+ * (Deprecated: Use Folder) Representation of a source, which is a collection of
649
+ * files and passages.
649
650
  */
650
651
  export interface Source {
651
652
  /**
@@ -704,34 +705,411 @@ export namespace AgentState {
704
705
  vector_db_provider?: ArchivesAPI.VectorDBProvider;
705
706
  }
706
707
 
708
+ export interface OpenAIModelSettings {
709
+ /**
710
+ * The maximum number of tokens the model can generate.
711
+ */
712
+ max_output_tokens?: number;
713
+
714
+ /**
715
+ * Whether to enable parallel tool calling.
716
+ */
717
+ parallel_tool_calls?: boolean;
718
+
719
+ /**
720
+ * The provider of the model.
721
+ */
722
+ provider?: 'openai';
723
+
724
+ /**
725
+ * The reasoning configuration for the model.
726
+ */
727
+ reasoning?: OpenAIModelSettings.Reasoning;
728
+
729
+ /**
730
+ * The response format for the model.
731
+ */
732
+ response_format?:
733
+ | AgentsAPI.TextResponseFormat
734
+ | AgentsAPI.JsonSchemaResponseFormat
735
+ | AgentsAPI.JsonObjectResponseFormat
736
+ | null;
737
+
738
+ /**
739
+ * The temperature of the model.
740
+ */
741
+ temperature?: number;
742
+ }
743
+
744
+ export namespace OpenAIModelSettings {
745
+ /**
746
+ * The reasoning configuration for the model.
747
+ */
748
+ export interface Reasoning {
749
+ /**
750
+ * The reasoning effort to use when generating text reasoning models
751
+ */
752
+ reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
753
+ }
754
+ }
755
+
756
+ export interface AnthropicModelSettings {
757
+ /**
758
+ * The maximum number of tokens the model can generate.
759
+ */
760
+ max_output_tokens?: number;
761
+
762
+ /**
763
+ * Whether to enable parallel tool calling.
764
+ */
765
+ parallel_tool_calls?: boolean;
766
+
767
+ /**
768
+ * The provider of the model.
769
+ */
770
+ provider?: 'anthropic';
771
+
772
+ /**
773
+ * The temperature of the model.
774
+ */
775
+ temperature?: number;
776
+
777
+ /**
778
+ * The thinking configuration for the model.
779
+ */
780
+ thinking?: AnthropicModelSettings.Thinking;
781
+
782
+ /**
783
+ * Soft control for how verbose model output should be, used for GPT-5 models.
784
+ */
785
+ verbosity?: 'low' | 'medium' | 'high' | null;
786
+ }
787
+
788
+ export namespace AnthropicModelSettings {
789
+ /**
790
+ * The thinking configuration for the model.
791
+ */
792
+ export interface Thinking {
793
+ /**
794
+ * The maximum number of tokens the model can use for extended thinking.
795
+ */
796
+ budget_tokens?: number;
797
+
798
+ /**
799
+ * The type of thinking to use.
800
+ */
801
+ type?: 'enabled' | 'disabled';
802
+ }
803
+ }
804
+
805
+ export interface GoogleAIModelSettings {
806
+ /**
807
+ * The maximum number of tokens the model can generate.
808
+ */
809
+ max_output_tokens?: number;
810
+
811
+ /**
812
+ * Whether to enable parallel tool calling.
813
+ */
814
+ parallel_tool_calls?: boolean;
815
+
816
+ /**
817
+ * The provider of the model.
818
+ */
819
+ provider?: 'google_ai';
820
+
821
+ /**
822
+ * The response schema for the model.
823
+ */
824
+ response_schema?:
825
+ | AgentsAPI.TextResponseFormat
826
+ | AgentsAPI.JsonSchemaResponseFormat
827
+ | AgentsAPI.JsonObjectResponseFormat
828
+ | null;
829
+
830
+ /**
831
+ * The temperature of the model.
832
+ */
833
+ temperature?: number;
834
+
835
+ /**
836
+ * The thinking configuration for the model.
837
+ */
838
+ thinking_config?: GoogleAIModelSettings.ThinkingConfig;
839
+ }
840
+
841
+ export namespace GoogleAIModelSettings {
842
+ /**
843
+ * The thinking configuration for the model.
844
+ */
845
+ export interface ThinkingConfig {
846
+ /**
847
+ * Whether to include thoughts in the model's response.
848
+ */
849
+ include_thoughts?: boolean;
850
+
851
+ /**
852
+ * The thinking budget for the model.
853
+ */
854
+ thinking_budget?: number;
855
+ }
856
+ }
857
+
858
+ export interface GoogleVertexModelSettings {
859
+ /**
860
+ * The maximum number of tokens the model can generate.
861
+ */
862
+ max_output_tokens?: number;
863
+
864
+ /**
865
+ * Whether to enable parallel tool calling.
866
+ */
867
+ parallel_tool_calls?: boolean;
868
+
869
+ /**
870
+ * The provider of the model.
871
+ */
872
+ provider?: 'google_vertex';
873
+
874
+ /**
875
+ * The response schema for the model.
876
+ */
877
+ response_schema?:
878
+ | AgentsAPI.TextResponseFormat
879
+ | AgentsAPI.JsonSchemaResponseFormat
880
+ | AgentsAPI.JsonObjectResponseFormat
881
+ | null;
882
+
883
+ /**
884
+ * The temperature of the model.
885
+ */
886
+ temperature?: number;
887
+
888
+ /**
889
+ * The thinking configuration for the model.
890
+ */
891
+ thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
892
+ }
893
+
894
+ export namespace GoogleVertexModelSettings {
895
+ /**
896
+ * The thinking configuration for the model.
897
+ */
898
+ export interface ThinkingConfig {
899
+ /**
900
+ * Whether to include thoughts in the model's response.
901
+ */
902
+ include_thoughts?: boolean;
903
+
904
+ /**
905
+ * The thinking budget for the model.
906
+ */
907
+ thinking_budget?: number;
908
+ }
909
+ }
910
+
911
+ /**
912
+ * Azure OpenAI model configuration (OpenAI-compatible).
913
+ */
914
+ export interface AzureModelSettings {
915
+ /**
916
+ * The maximum number of tokens the model can generate.
917
+ */
918
+ max_output_tokens?: number;
919
+
920
+ /**
921
+ * Whether to enable parallel tool calling.
922
+ */
923
+ parallel_tool_calls?: boolean;
924
+
925
+ /**
926
+ * The provider of the model.
927
+ */
928
+ provider?: 'azure';
929
+
930
+ /**
931
+ * The response format for the model.
932
+ */
933
+ response_format?:
934
+ | AgentsAPI.TextResponseFormat
935
+ | AgentsAPI.JsonSchemaResponseFormat
936
+ | AgentsAPI.JsonObjectResponseFormat
937
+ | null;
938
+
939
+ /**
940
+ * The temperature of the model.
941
+ */
942
+ temperature?: number;
943
+ }
944
+
945
+ /**
946
+ * xAI model configuration (OpenAI-compatible).
947
+ */
948
+ export interface XaiModelSettings {
949
+ /**
950
+ * The maximum number of tokens the model can generate.
951
+ */
952
+ max_output_tokens?: number;
953
+
954
+ /**
955
+ * Whether to enable parallel tool calling.
956
+ */
957
+ parallel_tool_calls?: boolean;
958
+
959
+ /**
960
+ * The provider of the model.
961
+ */
962
+ provider?: 'xai';
963
+
964
+ /**
965
+ * The response format for the model.
966
+ */
967
+ response_format?:
968
+ | AgentsAPI.TextResponseFormat
969
+ | AgentsAPI.JsonSchemaResponseFormat
970
+ | AgentsAPI.JsonObjectResponseFormat
971
+ | null;
972
+
973
+ /**
974
+ * The temperature of the model.
975
+ */
976
+ temperature?: number;
977
+ }
978
+
979
+ /**
980
+ * Groq model configuration (OpenAI-compatible).
981
+ */
982
+ export interface GroqModelSettings {
983
+ /**
984
+ * The maximum number of tokens the model can generate.
985
+ */
986
+ max_output_tokens?: number;
987
+
988
+ /**
989
+ * Whether to enable parallel tool calling.
990
+ */
991
+ parallel_tool_calls?: boolean;
992
+
993
+ /**
994
+ * The provider of the model.
995
+ */
996
+ provider?: 'groq';
997
+
998
+ /**
999
+ * The response format for the model.
1000
+ */
1001
+ response_format?:
1002
+ | AgentsAPI.TextResponseFormat
1003
+ | AgentsAPI.JsonSchemaResponseFormat
1004
+ | AgentsAPI.JsonObjectResponseFormat
1005
+ | null;
1006
+
1007
+ /**
1008
+ * The temperature of the model.
1009
+ */
1010
+ temperature?: number;
1011
+ }
1012
+
707
1013
  /**
708
- * The embedding model used by the agent.
1014
+ * Deepseek model configuration (OpenAI-compatible).
709
1015
  */
710
- export interface Embedding {
1016
+ export interface DeepseekModelSettings {
711
1017
  /**
712
- * The name of the model.
1018
+ * The maximum number of tokens the model can generate.
1019
+ */
1020
+ max_output_tokens?: number;
1021
+
1022
+ /**
1023
+ * Whether to enable parallel tool calling.
713
1024
  */
714
- model: string;
1025
+ parallel_tool_calls?: boolean;
715
1026
 
716
1027
  /**
717
1028
  * The provider of the model.
718
1029
  */
719
- provider: 'openai' | 'ollama';
1030
+ provider?: 'deepseek';
1031
+
1032
+ /**
1033
+ * The response format for the model.
1034
+ */
1035
+ response_format?:
1036
+ | AgentsAPI.TextResponseFormat
1037
+ | AgentsAPI.JsonSchemaResponseFormat
1038
+ | AgentsAPI.JsonObjectResponseFormat
1039
+ | null;
1040
+
1041
+ /**
1042
+ * The temperature of the model.
1043
+ */
1044
+ temperature?: number;
720
1045
  }
721
1046
 
722
1047
  /**
723
- * Schema for defining settings for a model
1048
+ * Together AI model configuration (OpenAI-compatible).
724
1049
  */
725
- export interface Model {
1050
+ export interface TogetherModelSettings {
1051
+ /**
1052
+ * The maximum number of tokens the model can generate.
1053
+ */
1054
+ max_output_tokens?: number;
1055
+
726
1056
  /**
727
- * The name of the model.
1057
+ * Whether to enable parallel tool calling.
728
1058
  */
729
- model: string;
1059
+ parallel_tool_calls?: boolean;
730
1060
 
1061
+ /**
1062
+ * The provider of the model.
1063
+ */
1064
+ provider?: 'together';
1065
+
1066
+ /**
1067
+ * The response format for the model.
1068
+ */
1069
+ response_format?:
1070
+ | AgentsAPI.TextResponseFormat
1071
+ | AgentsAPI.JsonSchemaResponseFormat
1072
+ | AgentsAPI.JsonObjectResponseFormat
1073
+ | null;
1074
+
1075
+ /**
1076
+ * The temperature of the model.
1077
+ */
1078
+ temperature?: number;
1079
+ }
1080
+
1081
+ /**
1082
+ * AWS Bedrock model configuration.
1083
+ */
1084
+ export interface BedrockModelSettings {
731
1085
  /**
732
1086
  * The maximum number of tokens the model can generate.
733
1087
  */
734
1088
  max_output_tokens?: number;
1089
+
1090
+ /**
1091
+ * Whether to enable parallel tool calling.
1092
+ */
1093
+ parallel_tool_calls?: boolean;
1094
+
1095
+ /**
1096
+ * The provider of the model.
1097
+ */
1098
+ provider?: 'bedrock';
1099
+
1100
+ /**
1101
+ * The response format for the model.
1102
+ */
1103
+ response_format?:
1104
+ | AgentsAPI.TextResponseFormat
1105
+ | AgentsAPI.JsonSchemaResponseFormat
1106
+ | AgentsAPI.JsonObjectResponseFormat
1107
+ | null;
1108
+
1109
+ /**
1110
+ * The temperature of the model.
1111
+ */
1112
+ temperature?: number;
735
1113
  }
736
1114
  }
737
1115
 
@@ -1111,10 +1489,9 @@ export interface AgentCreateParams {
1111
1489
  description?: string | null;
1112
1490
 
1113
1491
  /**
1114
- * The embedding configuration handle used by the agent, specified in the format
1115
- * provider/model-name.
1492
+ * The embedding model handle used by the agent (format: provider/model-name).
1116
1493
  */
1117
- embedding?: string | AgentCreateParams.EmbeddingModelSettings | null;
1494
+ embedding?: string | null;
1118
1495
 
1119
1496
  /**
1120
1497
  * @deprecated Deprecated: No longer used. The embedding chunk size used by the
@@ -1230,10 +1607,25 @@ export interface AgentCreateParams {
1230
1607
  metadata?: { [key: string]: unknown } | null;
1231
1608
 
1232
1609
  /**
1233
- * The model handle or model settings for the agent to use, specified either by a
1234
- * handle or an object. See the model schema for more information.
1610
+ * The model handle for the agent to use (format: provider/model-name).
1611
+ */
1612
+ model?: string | null;
1613
+
1614
+ /**
1615
+ * The model settings for the agent.
1235
1616
  */
1236
- model?: string | AgentCreateParams.ModelSettings | null;
1617
+ model_settings?:
1618
+ | AgentCreateParams.OpenAIModelSettings
1619
+ | AgentCreateParams.AnthropicModelSettings
1620
+ | AgentCreateParams.GoogleAIModelSettings
1621
+ | AgentCreateParams.GoogleVertexModelSettings
1622
+ | AgentCreateParams.AzureModelSettings
1623
+ | AgentCreateParams.XaiModelSettings
1624
+ | AgentCreateParams.GroqModelSettings
1625
+ | AgentCreateParams.DeepseekModelSettings
1626
+ | AgentCreateParams.TogetherModelSettings
1627
+ | AgentCreateParams.BedrockModelSettings
1628
+ | null;
1237
1629
 
1238
1630
  /**
1239
1631
  * The name of the agent.
@@ -1345,39 +1737,419 @@ export interface AgentCreateParams {
1345
1737
  }
1346
1738
 
1347
1739
  export namespace AgentCreateParams {
1348
- export interface EmbeddingModelSettings {
1740
+ export interface OpenAIModelSettings {
1741
+ /**
1742
+ * The maximum number of tokens the model can generate.
1743
+ */
1744
+ max_output_tokens?: number;
1745
+
1349
1746
  /**
1350
- * The name of the model.
1747
+ * Whether to enable parallel tool calling.
1351
1748
  */
1352
- model: string;
1749
+ parallel_tool_calls?: boolean;
1353
1750
 
1354
1751
  /**
1355
1752
  * The provider of the model.
1356
1753
  */
1357
- provider: 'openai' | 'ollama';
1358
- }
1754
+ provider?: 'openai';
1359
1755
 
1360
- /**
1361
- * Schema for defining settings for a model
1362
- */
1363
- export interface ModelSettings {
1364
1756
  /**
1365
- * The name of the model.
1757
+ * The reasoning configuration for the model.
1366
1758
  */
1367
- model: string;
1759
+ reasoning?: OpenAIModelSettings.Reasoning;
1368
1760
 
1369
1761
  /**
1370
- * The maximum number of tokens the model can generate.
1762
+ * The response format for the model.
1371
1763
  */
1372
- max_output_tokens?: number;
1764
+ response_format?:
1765
+ | AgentsAPI.TextResponseFormat
1766
+ | AgentsAPI.JsonSchemaResponseFormat
1767
+ | AgentsAPI.JsonObjectResponseFormat
1768
+ | null;
1769
+
1770
+ /**
1771
+ * The temperature of the model.
1772
+ */
1773
+ temperature?: number;
1373
1774
  }
1374
- }
1375
1775
 
1376
- export interface AgentRetrieveParams {
1377
- /**
1378
- * Specify which relational fields to include in the response. No relationships are
1379
- * included by default.
1380
- */
1776
+ export namespace OpenAIModelSettings {
1777
+ /**
1778
+ * The reasoning configuration for the model.
1779
+ */
1780
+ export interface Reasoning {
1781
+ /**
1782
+ * The reasoning effort to use when generating text reasoning models
1783
+ */
1784
+ reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
1785
+ }
1786
+ }
1787
+
1788
+ export interface AnthropicModelSettings {
1789
+ /**
1790
+ * The maximum number of tokens the model can generate.
1791
+ */
1792
+ max_output_tokens?: number;
1793
+
1794
+ /**
1795
+ * Whether to enable parallel tool calling.
1796
+ */
1797
+ parallel_tool_calls?: boolean;
1798
+
1799
+ /**
1800
+ * The provider of the model.
1801
+ */
1802
+ provider?: 'anthropic';
1803
+
1804
+ /**
1805
+ * The temperature of the model.
1806
+ */
1807
+ temperature?: number;
1808
+
1809
+ /**
1810
+ * The thinking configuration for the model.
1811
+ */
1812
+ thinking?: AnthropicModelSettings.Thinking;
1813
+
1814
+ /**
1815
+ * Soft control for how verbose model output should be, used for GPT-5 models.
1816
+ */
1817
+ verbosity?: 'low' | 'medium' | 'high' | null;
1818
+ }
1819
+
1820
+ export namespace AnthropicModelSettings {
1821
+ /**
1822
+ * The thinking configuration for the model.
1823
+ */
1824
+ export interface Thinking {
1825
+ /**
1826
+ * The maximum number of tokens the model can use for extended thinking.
1827
+ */
1828
+ budget_tokens?: number;
1829
+
1830
+ /**
1831
+ * The type of thinking to use.
1832
+ */
1833
+ type?: 'enabled' | 'disabled';
1834
+ }
1835
+ }
1836
+
1837
+ export interface GoogleAIModelSettings {
1838
+ /**
1839
+ * The maximum number of tokens the model can generate.
1840
+ */
1841
+ max_output_tokens?: number;
1842
+
1843
+ /**
1844
+ * Whether to enable parallel tool calling.
1845
+ */
1846
+ parallel_tool_calls?: boolean;
1847
+
1848
+ /**
1849
+ * The provider of the model.
1850
+ */
1851
+ provider?: 'google_ai';
1852
+
1853
+ /**
1854
+ * The response schema for the model.
1855
+ */
1856
+ response_schema?:
1857
+ | AgentsAPI.TextResponseFormat
1858
+ | AgentsAPI.JsonSchemaResponseFormat
1859
+ | AgentsAPI.JsonObjectResponseFormat
1860
+ | null;
1861
+
1862
+ /**
1863
+ * The temperature of the model.
1864
+ */
1865
+ temperature?: number;
1866
+
1867
+ /**
1868
+ * The thinking configuration for the model.
1869
+ */
1870
+ thinking_config?: GoogleAIModelSettings.ThinkingConfig;
1871
+ }
1872
+
1873
+ export namespace GoogleAIModelSettings {
1874
+ /**
1875
+ * The thinking configuration for the model.
1876
+ */
1877
+ export interface ThinkingConfig {
1878
+ /**
1879
+ * Whether to include thoughts in the model's response.
1880
+ */
1881
+ include_thoughts?: boolean;
1882
+
1883
+ /**
1884
+ * The thinking budget for the model.
1885
+ */
1886
+ thinking_budget?: number;
1887
+ }
1888
+ }
1889
+
1890
+ export interface GoogleVertexModelSettings {
1891
+ /**
1892
+ * The maximum number of tokens the model can generate.
1893
+ */
1894
+ max_output_tokens?: number;
1895
+
1896
+ /**
1897
+ * Whether to enable parallel tool calling.
1898
+ */
1899
+ parallel_tool_calls?: boolean;
1900
+
1901
+ /**
1902
+ * The provider of the model.
1903
+ */
1904
+ provider?: 'google_vertex';
1905
+
1906
+ /**
1907
+ * The response schema for the model.
1908
+ */
1909
+ response_schema?:
1910
+ | AgentsAPI.TextResponseFormat
1911
+ | AgentsAPI.JsonSchemaResponseFormat
1912
+ | AgentsAPI.JsonObjectResponseFormat
1913
+ | null;
1914
+
1915
+ /**
1916
+ * The temperature of the model.
1917
+ */
1918
+ temperature?: number;
1919
+
1920
+ /**
1921
+ * The thinking configuration for the model.
1922
+ */
1923
+ thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
1924
+ }
1925
+
1926
+ export namespace GoogleVertexModelSettings {
1927
+ /**
1928
+ * The thinking configuration for the model.
1929
+ */
1930
+ export interface ThinkingConfig {
1931
+ /**
1932
+ * Whether to include thoughts in the model's response.
1933
+ */
1934
+ include_thoughts?: boolean;
1935
+
1936
+ /**
1937
+ * The thinking budget for the model.
1938
+ */
1939
+ thinking_budget?: number;
1940
+ }
1941
+ }
1942
+
1943
+ /**
1944
+ * Azure OpenAI model configuration (OpenAI-compatible).
1945
+ */
1946
+ export interface AzureModelSettings {
1947
+ /**
1948
+ * The maximum number of tokens the model can generate.
1949
+ */
1950
+ max_output_tokens?: number;
1951
+
1952
+ /**
1953
+ * Whether to enable parallel tool calling.
1954
+ */
1955
+ parallel_tool_calls?: boolean;
1956
+
1957
+ /**
1958
+ * The provider of the model.
1959
+ */
1960
+ provider?: 'azure';
1961
+
1962
+ /**
1963
+ * The response format for the model.
1964
+ */
1965
+ response_format?:
1966
+ | AgentsAPI.TextResponseFormat
1967
+ | AgentsAPI.JsonSchemaResponseFormat
1968
+ | AgentsAPI.JsonObjectResponseFormat
1969
+ | null;
1970
+
1971
+ /**
1972
+ * The temperature of the model.
1973
+ */
1974
+ temperature?: number;
1975
+ }
1976
+
1977
+ /**
1978
+ * xAI model configuration (OpenAI-compatible).
1979
+ */
1980
+ export interface XaiModelSettings {
1981
+ /**
1982
+ * The maximum number of tokens the model can generate.
1983
+ */
1984
+ max_output_tokens?: number;
1985
+
1986
+ /**
1987
+ * Whether to enable parallel tool calling.
1988
+ */
1989
+ parallel_tool_calls?: boolean;
1990
+
1991
+ /**
1992
+ * The provider of the model.
1993
+ */
1994
+ provider?: 'xai';
1995
+
1996
+ /**
1997
+ * The response format for the model.
1998
+ */
1999
+ response_format?:
2000
+ | AgentsAPI.TextResponseFormat
2001
+ | AgentsAPI.JsonSchemaResponseFormat
2002
+ | AgentsAPI.JsonObjectResponseFormat
2003
+ | null;
2004
+
2005
+ /**
2006
+ * The temperature of the model.
2007
+ */
2008
+ temperature?: number;
2009
+ }
2010
+
2011
+ /**
2012
+ * Groq model configuration (OpenAI-compatible).
2013
+ */
2014
+ export interface GroqModelSettings {
2015
+ /**
2016
+ * The maximum number of tokens the model can generate.
2017
+ */
2018
+ max_output_tokens?: number;
2019
+
2020
+ /**
2021
+ * Whether to enable parallel tool calling.
2022
+ */
2023
+ parallel_tool_calls?: boolean;
2024
+
2025
+ /**
2026
+ * The provider of the model.
2027
+ */
2028
+ provider?: 'groq';
2029
+
2030
+ /**
2031
+ * The response format for the model.
2032
+ */
2033
+ response_format?:
2034
+ | AgentsAPI.TextResponseFormat
2035
+ | AgentsAPI.JsonSchemaResponseFormat
2036
+ | AgentsAPI.JsonObjectResponseFormat
2037
+ | null;
2038
+
2039
+ /**
2040
+ * The temperature of the model.
2041
+ */
2042
+ temperature?: number;
2043
+ }
2044
+
2045
+ /**
2046
+ * Deepseek model configuration (OpenAI-compatible).
2047
+ */
2048
+ export interface DeepseekModelSettings {
2049
+ /**
2050
+ * The maximum number of tokens the model can generate.
2051
+ */
2052
+ max_output_tokens?: number;
2053
+
2054
+ /**
2055
+ * Whether to enable parallel tool calling.
2056
+ */
2057
+ parallel_tool_calls?: boolean;
2058
+
2059
+ /**
2060
+ * The provider of the model.
2061
+ */
2062
+ provider?: 'deepseek';
2063
+
2064
+ /**
2065
+ * The response format for the model.
2066
+ */
2067
+ response_format?:
2068
+ | AgentsAPI.TextResponseFormat
2069
+ | AgentsAPI.JsonSchemaResponseFormat
2070
+ | AgentsAPI.JsonObjectResponseFormat
2071
+ | null;
2072
+
2073
+ /**
2074
+ * The temperature of the model.
2075
+ */
2076
+ temperature?: number;
2077
+ }
2078
+
2079
+ /**
2080
+ * Together AI model configuration (OpenAI-compatible).
2081
+ */
2082
+ export interface TogetherModelSettings {
2083
+ /**
2084
+ * The maximum number of tokens the model can generate.
2085
+ */
2086
+ max_output_tokens?: number;
2087
+
2088
+ /**
2089
+ * Whether to enable parallel tool calling.
2090
+ */
2091
+ parallel_tool_calls?: boolean;
2092
+
2093
+ /**
2094
+ * The provider of the model.
2095
+ */
2096
+ provider?: 'together';
2097
+
2098
+ /**
2099
+ * The response format for the model.
2100
+ */
2101
+ response_format?:
2102
+ | AgentsAPI.TextResponseFormat
2103
+ | AgentsAPI.JsonSchemaResponseFormat
2104
+ | AgentsAPI.JsonObjectResponseFormat
2105
+ | null;
2106
+
2107
+ /**
2108
+ * The temperature of the model.
2109
+ */
2110
+ temperature?: number;
2111
+ }
2112
+
2113
+ /**
2114
+ * AWS Bedrock model configuration.
2115
+ */
2116
+ export interface BedrockModelSettings {
2117
+ /**
2118
+ * The maximum number of tokens the model can generate.
2119
+ */
2120
+ max_output_tokens?: number;
2121
+
2122
+ /**
2123
+ * Whether to enable parallel tool calling.
2124
+ */
2125
+ parallel_tool_calls?: boolean;
2126
+
2127
+ /**
2128
+ * The provider of the model.
2129
+ */
2130
+ provider?: 'bedrock';
2131
+
2132
+ /**
2133
+ * The response format for the model.
2134
+ */
2135
+ response_format?:
2136
+ | AgentsAPI.TextResponseFormat
2137
+ | AgentsAPI.JsonSchemaResponseFormat
2138
+ | AgentsAPI.JsonObjectResponseFormat
2139
+ | null;
2140
+
2141
+ /**
2142
+ * The temperature of the model.
2143
+ */
2144
+ temperature?: number;
2145
+ }
2146
+ }
2147
+
2148
+ export interface AgentRetrieveParams {
2149
+ /**
2150
+ * Specify which relational fields to include in the response. No relationships are
2151
+ * included by default.
2152
+ */
1381
2153
  include?: Array<
1382
2154
  | 'agent.blocks'
1383
2155
  | 'agent.identities'
@@ -1571,10 +2343,9 @@ export interface AgentModifyParams {
1571
2343
  description?: string | null;
1572
2344
 
1573
2345
  /**
1574
- * The embedding configuration handle used by the agent, specified in the format
1575
- * provider/model-name.
2346
+ * The embedding model handle used by the agent (format: provider/model-name).
1576
2347
  */
1577
- embedding?: string | AgentModifyParams.EmbeddingModelSettings | null;
2348
+ embedding?: string | null;
1578
2349
 
1579
2350
  /**
1580
2351
  * Configuration for embedding model connection and processing parameters.
@@ -1646,10 +2417,25 @@ export interface AgentModifyParams {
1646
2417
  metadata?: { [key: string]: unknown } | null;
1647
2418
 
1648
2419
  /**
1649
- * The model used by the agent, specified either by a handle or an object. See the
1650
- * model schema for more information.
2420
+ * The model handle used by the agent (format: provider/model-name).
1651
2421
  */
1652
- model?: string | AgentModifyParams.ModelSettings | null;
2422
+ model?: string | null;
2423
+
2424
+ /**
2425
+ * The model settings for the agent.
2426
+ */
2427
+ model_settings?:
2428
+ | AgentModifyParams.OpenAIModelSettings
2429
+ | AgentModifyParams.AnthropicModelSettings
2430
+ | AgentModifyParams.GoogleAIModelSettings
2431
+ | AgentModifyParams.GoogleVertexModelSettings
2432
+ | AgentModifyParams.AzureModelSettings
2433
+ | AgentModifyParams.XaiModelSettings
2434
+ | AgentModifyParams.GroqModelSettings
2435
+ | AgentModifyParams.DeepseekModelSettings
2436
+ | AgentModifyParams.TogetherModelSettings
2437
+ | AgentModifyParams.BedrockModelSettings
2438
+ | null;
1653
2439
 
1654
2440
  /**
1655
2441
  * The name of the agent.
@@ -1742,31 +2528,411 @@ export interface AgentModifyParams {
1742
2528
  }
1743
2529
 
1744
2530
  export namespace AgentModifyParams {
1745
- export interface EmbeddingModelSettings {
2531
+ export interface OpenAIModelSettings {
2532
+ /**
2533
+ * The maximum number of tokens the model can generate.
2534
+ */
2535
+ max_output_tokens?: number;
2536
+
2537
+ /**
2538
+ * Whether to enable parallel tool calling.
2539
+ */
2540
+ parallel_tool_calls?: boolean;
2541
+
2542
+ /**
2543
+ * The provider of the model.
2544
+ */
2545
+ provider?: 'openai';
2546
+
2547
+ /**
2548
+ * The reasoning configuration for the model.
2549
+ */
2550
+ reasoning?: OpenAIModelSettings.Reasoning;
2551
+
2552
+ /**
2553
+ * The response format for the model.
2554
+ */
2555
+ response_format?:
2556
+ | AgentsAPI.TextResponseFormat
2557
+ | AgentsAPI.JsonSchemaResponseFormat
2558
+ | AgentsAPI.JsonObjectResponseFormat
2559
+ | null;
2560
+
2561
+ /**
2562
+ * The temperature of the model.
2563
+ */
2564
+ temperature?: number;
2565
+ }
2566
+
2567
+ export namespace OpenAIModelSettings {
2568
+ /**
2569
+ * The reasoning configuration for the model.
2570
+ */
2571
+ export interface Reasoning {
2572
+ /**
2573
+ * The reasoning effort to use when generating text reasoning models
2574
+ */
2575
+ reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
2576
+ }
2577
+ }
2578
+
2579
+ export interface AnthropicModelSettings {
2580
+ /**
2581
+ * The maximum number of tokens the model can generate.
2582
+ */
2583
+ max_output_tokens?: number;
2584
+
2585
+ /**
2586
+ * Whether to enable parallel tool calling.
2587
+ */
2588
+ parallel_tool_calls?: boolean;
2589
+
2590
+ /**
2591
+ * The provider of the model.
2592
+ */
2593
+ provider?: 'anthropic';
2594
+
2595
+ /**
2596
+ * The temperature of the model.
2597
+ */
2598
+ temperature?: number;
2599
+
2600
+ /**
2601
+ * The thinking configuration for the model.
2602
+ */
2603
+ thinking?: AnthropicModelSettings.Thinking;
2604
+
2605
+ /**
2606
+ * Soft control for how verbose model output should be, used for GPT-5 models.
2607
+ */
2608
+ verbosity?: 'low' | 'medium' | 'high' | null;
2609
+ }
2610
+
2611
+ export namespace AnthropicModelSettings {
2612
+ /**
2613
+ * The thinking configuration for the model.
2614
+ */
2615
+ export interface Thinking {
2616
+ /**
2617
+ * The maximum number of tokens the model can use for extended thinking.
2618
+ */
2619
+ budget_tokens?: number;
2620
+
2621
+ /**
2622
+ * The type of thinking to use.
2623
+ */
2624
+ type?: 'enabled' | 'disabled';
2625
+ }
2626
+ }
2627
+
2628
+ export interface GoogleAIModelSettings {
1746
2629
  /**
1747
- * The name of the model.
2630
+ * The maximum number of tokens the model can generate.
1748
2631
  */
1749
- model: string;
2632
+ max_output_tokens?: number;
2633
+
2634
+ /**
2635
+ * Whether to enable parallel tool calling.
2636
+ */
2637
+ parallel_tool_calls?: boolean;
1750
2638
 
1751
2639
  /**
1752
2640
  * The provider of the model.
1753
2641
  */
1754
- provider: 'openai' | 'ollama';
2642
+ provider?: 'google_ai';
2643
+
2644
+ /**
2645
+ * The response schema for the model.
2646
+ */
2647
+ response_schema?:
2648
+ | AgentsAPI.TextResponseFormat
2649
+ | AgentsAPI.JsonSchemaResponseFormat
2650
+ | AgentsAPI.JsonObjectResponseFormat
2651
+ | null;
2652
+
2653
+ /**
2654
+ * The temperature of the model.
2655
+ */
2656
+ temperature?: number;
2657
+
2658
+ /**
2659
+ * The thinking configuration for the model.
2660
+ */
2661
+ thinking_config?: GoogleAIModelSettings.ThinkingConfig;
2662
+ }
2663
+
2664
+ export namespace GoogleAIModelSettings {
2665
+ /**
2666
+ * The thinking configuration for the model.
2667
+ */
2668
+ export interface ThinkingConfig {
2669
+ /**
2670
+ * Whether to include thoughts in the model's response.
2671
+ */
2672
+ include_thoughts?: boolean;
2673
+
2674
+ /**
2675
+ * The thinking budget for the model.
2676
+ */
2677
+ thinking_budget?: number;
2678
+ }
2679
+ }
2680
+
2681
+ export interface GoogleVertexModelSettings {
2682
+ /**
2683
+ * The maximum number of tokens the model can generate.
2684
+ */
2685
+ max_output_tokens?: number;
2686
+
2687
+ /**
2688
+ * Whether to enable parallel tool calling.
2689
+ */
2690
+ parallel_tool_calls?: boolean;
2691
+
2692
+ /**
2693
+ * The provider of the model.
2694
+ */
2695
+ provider?: 'google_vertex';
2696
+
2697
+ /**
2698
+ * The response schema for the model.
2699
+ */
2700
+ response_schema?:
2701
+ | AgentsAPI.TextResponseFormat
2702
+ | AgentsAPI.JsonSchemaResponseFormat
2703
+ | AgentsAPI.JsonObjectResponseFormat
2704
+ | null;
2705
+
2706
+ /**
2707
+ * The temperature of the model.
2708
+ */
2709
+ temperature?: number;
2710
+
2711
+ /**
2712
+ * The thinking configuration for the model.
2713
+ */
2714
+ thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
2715
+ }
2716
+
2717
+ export namespace GoogleVertexModelSettings {
2718
+ /**
2719
+ * The thinking configuration for the model.
2720
+ */
2721
+ export interface ThinkingConfig {
2722
+ /**
2723
+ * Whether to include thoughts in the model's response.
2724
+ */
2725
+ include_thoughts?: boolean;
2726
+
2727
+ /**
2728
+ * The thinking budget for the model.
2729
+ */
2730
+ thinking_budget?: number;
2731
+ }
1755
2732
  }
1756
2733
 
1757
2734
  /**
1758
- * Schema for defining settings for a model
2735
+ * Azure OpenAI model configuration (OpenAI-compatible).
1759
2736
  */
1760
- export interface ModelSettings {
2737
+ export interface AzureModelSettings {
2738
+ /**
2739
+ * The maximum number of tokens the model can generate.
2740
+ */
2741
+ max_output_tokens?: number;
2742
+
2743
+ /**
2744
+ * Whether to enable parallel tool calling.
2745
+ */
2746
+ parallel_tool_calls?: boolean;
2747
+
2748
+ /**
2749
+ * The provider of the model.
2750
+ */
2751
+ provider?: 'azure';
2752
+
2753
+ /**
2754
+ * The response format for the model.
2755
+ */
2756
+ response_format?:
2757
+ | AgentsAPI.TextResponseFormat
2758
+ | AgentsAPI.JsonSchemaResponseFormat
2759
+ | AgentsAPI.JsonObjectResponseFormat
2760
+ | null;
2761
+
1761
2762
  /**
1762
- * The name of the model.
2763
+ * The temperature of the model.
1763
2764
  */
1764
- model: string;
2765
+ temperature?: number;
2766
+ }
1765
2767
 
2768
+ /**
2769
+ * xAI model configuration (OpenAI-compatible).
2770
+ */
2771
+ export interface XaiModelSettings {
1766
2772
  /**
1767
2773
  * The maximum number of tokens the model can generate.
1768
2774
  */
1769
2775
  max_output_tokens?: number;
2776
+
2777
+ /**
2778
+ * Whether to enable parallel tool calling.
2779
+ */
2780
+ parallel_tool_calls?: boolean;
2781
+
2782
+ /**
2783
+ * The provider of the model.
2784
+ */
2785
+ provider?: 'xai';
2786
+
2787
+ /**
2788
+ * The response format for the model.
2789
+ */
2790
+ response_format?:
2791
+ | AgentsAPI.TextResponseFormat
2792
+ | AgentsAPI.JsonSchemaResponseFormat
2793
+ | AgentsAPI.JsonObjectResponseFormat
2794
+ | null;
2795
+
2796
+ /**
2797
+ * The temperature of the model.
2798
+ */
2799
+ temperature?: number;
2800
+ }
2801
+
2802
+ /**
2803
+ * Groq model configuration (OpenAI-compatible).
2804
+ */
2805
+ export interface GroqModelSettings {
2806
+ /**
2807
+ * The maximum number of tokens the model can generate.
2808
+ */
2809
+ max_output_tokens?: number;
2810
+
2811
+ /**
2812
+ * Whether to enable parallel tool calling.
2813
+ */
2814
+ parallel_tool_calls?: boolean;
2815
+
2816
+ /**
2817
+ * The provider of the model.
2818
+ */
2819
+ provider?: 'groq';
2820
+
2821
+ /**
2822
+ * The response format for the model.
2823
+ */
2824
+ response_format?:
2825
+ | AgentsAPI.TextResponseFormat
2826
+ | AgentsAPI.JsonSchemaResponseFormat
2827
+ | AgentsAPI.JsonObjectResponseFormat
2828
+ | null;
2829
+
2830
+ /**
2831
+ * The temperature of the model.
2832
+ */
2833
+ temperature?: number;
2834
+ }
2835
+
2836
+ /**
2837
+ * Deepseek model configuration (OpenAI-compatible).
2838
+ */
2839
+ export interface DeepseekModelSettings {
2840
+ /**
2841
+ * The maximum number of tokens the model can generate.
2842
+ */
2843
+ max_output_tokens?: number;
2844
+
2845
+ /**
2846
+ * Whether to enable parallel tool calling.
2847
+ */
2848
+ parallel_tool_calls?: boolean;
2849
+
2850
+ /**
2851
+ * The provider of the model.
2852
+ */
2853
+ provider?: 'deepseek';
2854
+
2855
+ /**
2856
+ * The response format for the model.
2857
+ */
2858
+ response_format?:
2859
+ | AgentsAPI.TextResponseFormat
2860
+ | AgentsAPI.JsonSchemaResponseFormat
2861
+ | AgentsAPI.JsonObjectResponseFormat
2862
+ | null;
2863
+
2864
+ /**
2865
+ * The temperature of the model.
2866
+ */
2867
+ temperature?: number;
2868
+ }
2869
+
2870
+ /**
2871
+ * Together AI model configuration (OpenAI-compatible).
2872
+ */
2873
+ export interface TogetherModelSettings {
2874
+ /**
2875
+ * The maximum number of tokens the model can generate.
2876
+ */
2877
+ max_output_tokens?: number;
2878
+
2879
+ /**
2880
+ * Whether to enable parallel tool calling.
2881
+ */
2882
+ parallel_tool_calls?: boolean;
2883
+
2884
+ /**
2885
+ * The provider of the model.
2886
+ */
2887
+ provider?: 'together';
2888
+
2889
+ /**
2890
+ * The response format for the model.
2891
+ */
2892
+ response_format?:
2893
+ | AgentsAPI.TextResponseFormat
2894
+ | AgentsAPI.JsonSchemaResponseFormat
2895
+ | AgentsAPI.JsonObjectResponseFormat
2896
+ | null;
2897
+
2898
+ /**
2899
+ * The temperature of the model.
2900
+ */
2901
+ temperature?: number;
2902
+ }
2903
+
2904
+ /**
2905
+ * AWS Bedrock model configuration.
2906
+ */
2907
+ export interface BedrockModelSettings {
2908
+ /**
2909
+ * The maximum number of tokens the model can generate.
2910
+ */
2911
+ max_output_tokens?: number;
2912
+
2913
+ /**
2914
+ * Whether to enable parallel tool calling.
2915
+ */
2916
+ parallel_tool_calls?: boolean;
2917
+
2918
+ /**
2919
+ * The provider of the model.
2920
+ */
2921
+ provider?: 'bedrock';
2922
+
2923
+ /**
2924
+ * The response format for the model.
2925
+ */
2926
+ response_format?:
2927
+ | AgentsAPI.TextResponseFormat
2928
+ | AgentsAPI.JsonSchemaResponseFormat
2929
+ | AgentsAPI.JsonObjectResponseFormat
2930
+ | null;
2931
+
2932
+ /**
2933
+ * The temperature of the model.
2934
+ */
2935
+ temperature?: number;
1770
2936
  }
1771
2937
  }
1772
2938