@letta-ai/letta-client 1.0.0-alpha.15 → 1.0.0-alpha.17
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +41 -0
- package/client.d.mts +22 -22
- package/client.d.mts.map +1 -1
- package/client.d.ts +22 -22
- package/client.d.ts.map +1 -1
- package/client.js.map +1 -1
- package/client.mjs.map +1 -1
- package/package.json +1 -1
- package/resources/agents/agents.d.mts +1098 -136
- package/resources/agents/agents.d.mts.map +1 -1
- package/resources/agents/agents.d.ts +1098 -136
- package/resources/agents/agents.d.ts.map +1 -1
- package/resources/agents/agents.js +0 -6
- package/resources/agents/agents.js.map +1 -1
- package/resources/agents/agents.mjs +1 -7
- package/resources/agents/agents.mjs.map +1 -1
- package/resources/agents/blocks.d.mts +1 -15
- package/resources/agents/blocks.d.mts.map +1 -1
- package/resources/agents/blocks.d.ts +1 -15
- package/resources/agents/blocks.d.ts.map +1 -1
- package/resources/agents/folders.d.mts +47 -76
- package/resources/agents/folders.d.mts.map +1 -1
- package/resources/agents/folders.d.ts +47 -76
- package/resources/agents/folders.d.ts.map +1 -1
- package/resources/agents/folders.js +5 -1
- package/resources/agents/folders.js.map +1 -1
- package/resources/agents/folders.mjs +5 -1
- package/resources/agents/folders.mjs.map +1 -1
- package/resources/agents/index.d.mts +4 -4
- package/resources/agents/index.d.mts.map +1 -1
- package/resources/agents/index.d.ts +4 -4
- package/resources/agents/index.d.ts.map +1 -1
- package/resources/agents/index.js.map +1 -1
- package/resources/agents/index.mjs.map +1 -1
- package/resources/agents/messages.d.mts +87 -28
- package/resources/agents/messages.d.mts.map +1 -1
- package/resources/agents/messages.d.ts +87 -28
- package/resources/agents/messages.d.ts.map +1 -1
- package/resources/agents/messages.js +5 -5
- package/resources/agents/messages.js.map +1 -1
- package/resources/agents/messages.mjs +5 -5
- package/resources/agents/messages.mjs.map +1 -1
- package/resources/agents/tools.d.mts +6 -27
- package/resources/agents/tools.d.mts.map +1 -1
- package/resources/agents/tools.d.ts +6 -27
- package/resources/agents/tools.d.ts.map +1 -1
- package/resources/agents/tools.js +5 -1
- package/resources/agents/tools.js.map +1 -1
- package/resources/agents/tools.mjs +5 -1
- package/resources/agents/tools.mjs.map +1 -1
- package/resources/archives.d.mts +9 -37
- package/resources/archives.d.mts.map +1 -1
- package/resources/archives.d.ts +9 -37
- package/resources/archives.d.ts.map +1 -1
- package/resources/archives.js +2 -1
- package/resources/archives.js.map +1 -1
- package/resources/archives.mjs +2 -1
- package/resources/archives.mjs.map +1 -1
- package/resources/batches/batches.d.mts +8 -30
- package/resources/batches/batches.d.mts.map +1 -1
- package/resources/batches/batches.d.ts +8 -30
- package/resources/batches/batches.d.ts.map +1 -1
- package/resources/batches/batches.js +2 -1
- package/resources/batches/batches.js.map +1 -1
- package/resources/batches/batches.mjs +2 -1
- package/resources/batches/batches.mjs.map +1 -1
- package/resources/batches/index.d.mts +2 -2
- package/resources/batches/index.d.mts.map +1 -1
- package/resources/batches/index.d.ts +2 -2
- package/resources/batches/index.d.ts.map +1 -1
- package/resources/batches/index.js.map +1 -1
- package/resources/batches/index.mjs.map +1 -1
- package/resources/batches/messages.d.mts +6 -30
- package/resources/batches/messages.d.mts.map +1 -1
- package/resources/batches/messages.d.ts +6 -30
- package/resources/batches/messages.d.ts.map +1 -1
- package/resources/batches/messages.js +2 -1
- package/resources/batches/messages.js.map +1 -1
- package/resources/batches/messages.mjs +2 -1
- package/resources/batches/messages.mjs.map +1 -1
- package/resources/blocks/blocks.d.mts +1 -6
- package/resources/blocks/blocks.d.mts.map +1 -1
- package/resources/blocks/blocks.d.ts +1 -6
- package/resources/blocks/blocks.d.ts.map +1 -1
- package/resources/blocks/blocks.js +0 -6
- package/resources/blocks/blocks.js.map +1 -1
- package/resources/blocks/blocks.mjs +0 -6
- package/resources/blocks/blocks.mjs.map +1 -1
- package/resources/blocks/index.d.mts +1 -1
- package/resources/blocks/index.d.mts.map +1 -1
- package/resources/blocks/index.d.ts +1 -1
- package/resources/blocks/index.d.ts.map +1 -1
- package/resources/blocks/index.js.map +1 -1
- package/resources/blocks/index.mjs.map +1 -1
- package/resources/folders/files.d.mts +73 -97
- package/resources/folders/files.d.mts.map +1 -1
- package/resources/folders/files.d.ts +73 -97
- package/resources/folders/files.d.ts.map +1 -1
- package/resources/folders/files.js +5 -1
- package/resources/folders/files.js.map +1 -1
- package/resources/folders/files.mjs +5 -1
- package/resources/folders/files.mjs.map +1 -1
- package/resources/folders/folders.d.mts +11 -44
- package/resources/folders/folders.d.mts.map +1 -1
- package/resources/folders/folders.d.ts +11 -44
- package/resources/folders/folders.d.ts.map +1 -1
- package/resources/folders/folders.js +2 -7
- package/resources/folders/folders.js.map +1 -1
- package/resources/folders/folders.mjs +2 -7
- package/resources/folders/folders.mjs.map +1 -1
- package/resources/folders/index.d.mts +2 -2
- package/resources/folders/index.d.mts.map +1 -1
- package/resources/folders/index.d.ts +2 -2
- package/resources/folders/index.d.ts.map +1 -1
- package/resources/folders/index.js.map +1 -1
- package/resources/folders/index.mjs.map +1 -1
- package/resources/groups/groups.d.mts +1 -6
- package/resources/groups/groups.d.mts.map +1 -1
- package/resources/groups/groups.d.ts +1 -6
- package/resources/groups/groups.d.ts.map +1 -1
- package/resources/groups/groups.js +0 -6
- package/resources/groups/groups.js.map +1 -1
- package/resources/groups/groups.mjs +0 -6
- package/resources/groups/groups.mjs.map +1 -1
- package/resources/groups/index.d.mts +1 -1
- package/resources/groups/index.d.mts.map +1 -1
- package/resources/groups/index.d.ts +1 -1
- package/resources/groups/index.d.ts.map +1 -1
- package/resources/groups/index.js.map +1 -1
- package/resources/groups/index.mjs.map +1 -1
- package/resources/groups/messages.d.mts +9 -3
- package/resources/groups/messages.d.mts.map +1 -1
- package/resources/groups/messages.d.ts +9 -3
- package/resources/groups/messages.d.ts.map +1 -1
- package/resources/identities/identities.d.mts +9 -36
- package/resources/identities/identities.d.mts.map +1 -1
- package/resources/identities/identities.d.ts +9 -36
- package/resources/identities/identities.d.ts.map +1 -1
- package/resources/identities/identities.js +2 -7
- package/resources/identities/identities.js.map +1 -1
- package/resources/identities/identities.mjs +2 -7
- package/resources/identities/identities.mjs.map +1 -1
- package/resources/identities/index.d.mts +1 -1
- package/resources/identities/index.d.mts.map +1 -1
- package/resources/identities/index.d.ts +1 -1
- package/resources/identities/index.d.ts.map +1 -1
- package/resources/identities/index.js.map +1 -1
- package/resources/identities/index.mjs.map +1 -1
- package/resources/index.d.mts +11 -11
- package/resources/index.d.mts.map +1 -1
- package/resources/index.d.ts +11 -11
- package/resources/index.d.ts.map +1 -1
- package/resources/index.js.map +1 -1
- package/resources/index.mjs.map +1 -1
- package/resources/mcp-servers/index.d.mts +2 -3
- package/resources/mcp-servers/index.d.mts.map +1 -1
- package/resources/mcp-servers/index.d.ts +2 -3
- package/resources/mcp-servers/index.d.ts.map +1 -1
- package/resources/mcp-servers/index.js +1 -3
- package/resources/mcp-servers/index.js.map +1 -1
- package/resources/mcp-servers/index.mjs +1 -2
- package/resources/mcp-servers/index.mjs.map +1 -1
- package/resources/mcp-servers/mcp-servers.d.mts +292 -378
- package/resources/mcp-servers/mcp-servers.d.mts.map +1 -1
- package/resources/mcp-servers/mcp-servers.d.ts +292 -378
- package/resources/mcp-servers/mcp-servers.d.ts.map +1 -1
- package/resources/mcp-servers/mcp-servers.js +17 -4
- package/resources/mcp-servers/mcp-servers.js.map +1 -1
- package/resources/mcp-servers/mcp-servers.mjs +17 -4
- package/resources/mcp-servers/mcp-servers.mjs.map +1 -1
- package/resources/mcp-servers/tools.d.mts +3 -41
- package/resources/mcp-servers/tools.d.mts.map +1 -1
- package/resources/mcp-servers/tools.d.ts +3 -41
- package/resources/mcp-servers/tools.d.ts.map +1 -1
- package/resources/mcp-servers/tools.js.map +1 -1
- package/resources/mcp-servers/tools.mjs.map +1 -1
- package/resources/models/embeddings.d.mts +6 -2
- package/resources/models/embeddings.d.mts.map +1 -1
- package/resources/models/embeddings.d.ts +6 -2
- package/resources/models/embeddings.d.ts.map +1 -1
- package/resources/models/embeddings.js +5 -1
- package/resources/models/embeddings.js.map +1 -1
- package/resources/models/embeddings.mjs +5 -1
- package/resources/models/embeddings.mjs.map +1 -1
- package/resources/models/index.d.mts +1 -1
- package/resources/models/index.d.mts.map +1 -1
- package/resources/models/index.d.ts +1 -1
- package/resources/models/index.d.ts.map +1 -1
- package/resources/models/index.js.map +1 -1
- package/resources/models/index.mjs.map +1 -1
- package/resources/models/models.d.mts +174 -3
- package/resources/models/models.d.mts.map +1 -1
- package/resources/models/models.d.ts +174 -3
- package/resources/models/models.d.ts.map +1 -1
- package/resources/models/models.js +5 -1
- package/resources/models/models.js.map +1 -1
- package/resources/models/models.mjs +5 -1
- package/resources/models/models.mjs.map +1 -1
- package/resources/runs/index.d.mts +1 -1
- package/resources/runs/index.d.mts.map +1 -1
- package/resources/runs/index.d.ts +1 -1
- package/resources/runs/index.d.ts.map +1 -1
- package/resources/runs/index.js.map +1 -1
- package/resources/runs/index.mjs.map +1 -1
- package/resources/runs/runs.d.mts +6 -32
- package/resources/runs/runs.d.mts.map +1 -1
- package/resources/runs/runs.d.ts +6 -32
- package/resources/runs/runs.d.ts.map +1 -1
- package/resources/runs/runs.js +2 -1
- package/resources/runs/runs.js.map +1 -1
- package/resources/runs/runs.mjs +2 -1
- package/resources/runs/runs.mjs.map +1 -1
- package/resources/tools.d.mts +6 -68
- package/resources/tools.d.mts.map +1 -1
- package/resources/tools.d.ts +6 -68
- package/resources/tools.d.ts.map +1 -1
- package/resources/tools.js +2 -7
- package/resources/tools.js.map +1 -1
- package/resources/tools.mjs +2 -7
- package/resources/tools.mjs.map +1 -1
- package/src/client.ts +41 -28
- package/src/resources/agents/agents.ts +1364 -77
- package/src/resources/agents/blocks.ts +1 -15
- package/src/resources/agents/folders.ts +55 -85
- package/src/resources/agents/index.ts +5 -2
- package/src/resources/agents/messages.ts +123 -28
- package/src/resources/agents/tools.ts +10 -34
- package/src/resources/archives.ts +12 -45
- package/src/resources/batches/batches.ts +10 -41
- package/src/resources/batches/index.ts +2 -2
- package/src/resources/batches/messages.ts +12 -37
- package/src/resources/blocks/blocks.ts +0 -10
- package/src/resources/blocks/index.ts +0 -1
- package/src/resources/folders/files.ts +88 -113
- package/src/resources/folders/folders.ts +14 -55
- package/src/resources/folders/index.ts +2 -2
- package/src/resources/groups/groups.ts +0 -10
- package/src/resources/groups/index.ts +0 -1
- package/src/resources/groups/messages.ts +10 -3
- package/src/resources/identities/identities.ts +12 -49
- package/src/resources/identities/index.ts +1 -2
- package/src/resources/index.ts +21 -14
- package/src/resources/mcp-servers/index.ts +14 -9
- package/src/resources/mcp-servers/mcp-servers.ts +303 -402
- package/src/resources/mcp-servers/tools.ts +6 -47
- package/src/resources/models/embeddings.ts +6 -2
- package/src/resources/models/index.ts +2 -0
- package/src/resources/models/models.ts +254 -2
- package/src/resources/runs/index.ts +1 -1
- package/src/resources/runs/runs.ts +11 -44
- package/src/resources/tools.ts +8 -92
- package/src/version.ts +1 -1
- package/version.d.mts +1 -1
- package/version.d.ts +1 -1
- package/version.js +1 -1
- package/version.mjs +1 -1
- package/resources/mcp-servers/refresh.d.mts +0 -24
- package/resources/mcp-servers/refresh.d.mts.map +0 -1
- package/resources/mcp-servers/refresh.d.ts +0 -24
- package/resources/mcp-servers/refresh.d.ts.map +0 -1
- package/resources/mcp-servers/refresh.js +0 -27
- package/resources/mcp-servers/refresh.js.map +0 -1
- package/resources/mcp-servers/refresh.mjs +0 -23
- package/resources/mcp-servers/refresh.mjs.map +0 -1
- package/src/resources/mcp-servers/refresh.ts +0 -43
|
@@ -33,6 +33,7 @@ import {
|
|
|
33
33
|
FolderDetachParams,
|
|
34
34
|
FolderListParams,
|
|
35
35
|
FolderListResponse,
|
|
36
|
+
FolderListResponsesArrayPage,
|
|
36
37
|
Folders,
|
|
37
38
|
} from './folders';
|
|
38
39
|
import * as GroupsAPI from './groups';
|
|
@@ -66,6 +67,8 @@ import {
|
|
|
66
67
|
MessageRole,
|
|
67
68
|
MessageSendAsyncParams,
|
|
68
69
|
MessageSendParams,
|
|
70
|
+
MessageSendParamsNonStreaming,
|
|
71
|
+
MessageSendParamsStreaming,
|
|
69
72
|
MessageStreamParams,
|
|
70
73
|
MessageType,
|
|
71
74
|
Messages,
|
|
@@ -90,18 +93,12 @@ import {
|
|
|
90
93
|
UserMessage,
|
|
91
94
|
} from './messages';
|
|
92
95
|
import * as AgentsToolsAPI from './tools';
|
|
93
|
-
import {
|
|
94
|
-
ToolAttachParams,
|
|
95
|
-
ToolDetachParams,
|
|
96
|
-
ToolListParams,
|
|
97
|
-
ToolListResponse,
|
|
98
|
-
ToolUpdateApprovalParams,
|
|
99
|
-
Tools,
|
|
100
|
-
} from './tools';
|
|
96
|
+
import { ToolAttachParams, ToolDetachParams, ToolListParams, ToolUpdateApprovalParams, Tools } from './tools';
|
|
101
97
|
import * as BlocksBlocksAPI from '../blocks/blocks';
|
|
102
98
|
import * as GroupsGroupsAPI from '../groups/groups';
|
|
103
99
|
import * as IdentitiesAPI from '../identities/identities';
|
|
104
100
|
import * as ModelsAPI from '../models/models';
|
|
101
|
+
import * as RunsAPI from '../runs/runs';
|
|
105
102
|
import { APIPromise } from '../../core/api-promise';
|
|
106
103
|
import { ArrayPage, type ArrayPageParams, PagePromise } from '../../core/pagination';
|
|
107
104
|
import { type Uploadable } from '../../core/uploads';
|
|
@@ -153,13 +150,6 @@ export class Agents extends APIResource {
|
|
|
153
150
|
return this._client.delete(path`/v1/agents/${agentID}`, options);
|
|
154
151
|
}
|
|
155
152
|
|
|
156
|
-
/**
|
|
157
|
-
* Get the total number of agents.
|
|
158
|
-
*/
|
|
159
|
-
count(options?: RequestOptions): APIPromise<AgentCountResponse> {
|
|
160
|
-
return this._client.get('/v1/agents/count', options);
|
|
161
|
-
}
|
|
162
|
-
|
|
163
153
|
/**
|
|
164
154
|
* Export the serialized JSON representation of an agent, formatted with
|
|
165
155
|
* indentation.
|
|
@@ -264,16 +254,6 @@ export interface AgentEnvironmentVariable {
|
|
|
264
254
|
* Representation of an agent's state. This is the state of the agent at a given
|
|
265
255
|
* time, and is persisted in the DB backend. The state has all the information
|
|
266
256
|
* needed to recreate a persisted agent.
|
|
267
|
-
*
|
|
268
|
-
* Parameters: id (str): The unique identifier of the agent. name (str): The name
|
|
269
|
-
* of the agent (must be unique to the user). created_at (datetime): The datetime
|
|
270
|
-
* the agent was created. message_ids (List[str]): The ids of the messages in the
|
|
271
|
-
* agent's in-context memory. memory (Memory): The in-context memory of the agent.
|
|
272
|
-
* tools (List[str]): The tools used by the agent. This includes any memory editing
|
|
273
|
-
* functions specified in `memory`. system (str): The system prompt used by the
|
|
274
|
-
* agent. llm_config (LLMConfig): The LLM configuration used by the agent.
|
|
275
|
-
* embedding_config (EmbeddingConfig): The embedding configuration used by the
|
|
276
|
-
* agent.
|
|
277
257
|
*/
|
|
278
258
|
export interface AgentState {
|
|
279
259
|
/**
|
|
@@ -292,17 +272,20 @@ export interface AgentState {
|
|
|
292
272
|
blocks: Array<BlocksAPI.Block>;
|
|
293
273
|
|
|
294
274
|
/**
|
|
295
|
-
*
|
|
275
|
+
* @deprecated Deprecated: Use `embedding` field instead. The embedding
|
|
276
|
+
* configuration used by the agent.
|
|
296
277
|
*/
|
|
297
278
|
embedding_config: ModelsAPI.EmbeddingConfig;
|
|
298
279
|
|
|
299
280
|
/**
|
|
300
|
-
* The LLM configuration used by
|
|
281
|
+
* @deprecated Deprecated: Use `model` field instead. The LLM configuration used by
|
|
282
|
+
* the agent.
|
|
301
283
|
*/
|
|
302
284
|
llm_config: ModelsAPI.LlmConfig;
|
|
303
285
|
|
|
304
286
|
/**
|
|
305
|
-
* @deprecated The in-context memory of the
|
|
287
|
+
* @deprecated Deprecated: Use `blocks` field instead. The in-context memory of the
|
|
288
|
+
* agent.
|
|
306
289
|
*/
|
|
307
290
|
memory: AgentState.Memory;
|
|
308
291
|
|
|
@@ -356,6 +339,11 @@ export interface AgentState {
|
|
|
356
339
|
*/
|
|
357
340
|
description?: string | null;
|
|
358
341
|
|
|
342
|
+
/**
|
|
343
|
+
* The embedding model handle used by the agent (format: provider/model-name).
|
|
344
|
+
*/
|
|
345
|
+
embedding?: string | null;
|
|
346
|
+
|
|
359
347
|
/**
|
|
360
348
|
* If set to True, memory management will move to a background agent thread.
|
|
361
349
|
*/
|
|
@@ -377,7 +365,8 @@ export interface AgentState {
|
|
|
377
365
|
identities?: Array<IdentitiesAPI.Identity>;
|
|
378
366
|
|
|
379
367
|
/**
|
|
380
|
-
* @deprecated
|
|
368
|
+
* @deprecated Deprecated: Use `identities` field instead. The ids of the
|
|
369
|
+
* identities associated with this agent.
|
|
381
370
|
*/
|
|
382
371
|
identity_ids?: Array<string>;
|
|
383
372
|
|
|
@@ -391,6 +380,11 @@ export interface AgentState {
|
|
|
391
380
|
*/
|
|
392
381
|
last_run_duration_ms?: number | null;
|
|
393
382
|
|
|
383
|
+
/**
|
|
384
|
+
* The stop reason from the agent's last run.
|
|
385
|
+
*/
|
|
386
|
+
last_stop_reason?: RunsAPI.StopReasonType | null;
|
|
387
|
+
|
|
394
388
|
/**
|
|
395
389
|
* The id of the user that made this object.
|
|
396
390
|
*/
|
|
@@ -425,7 +419,29 @@ export interface AgentState {
|
|
|
425
419
|
metadata?: { [key: string]: unknown } | null;
|
|
426
420
|
|
|
427
421
|
/**
|
|
428
|
-
*
|
|
422
|
+
* The model handle used by the agent (format: provider/model-name).
|
|
423
|
+
*/
|
|
424
|
+
model?: string | null;
|
|
425
|
+
|
|
426
|
+
/**
|
|
427
|
+
* The model settings used by the agent.
|
|
428
|
+
*/
|
|
429
|
+
model_settings?:
|
|
430
|
+
| AgentState.OpenAIModelSettings
|
|
431
|
+
| AgentState.AnthropicModelSettings
|
|
432
|
+
| AgentState.GoogleAIModelSettings
|
|
433
|
+
| AgentState.GoogleVertexModelSettings
|
|
434
|
+
| AgentState.AzureModelSettings
|
|
435
|
+
| AgentState.XaiModelSettings
|
|
436
|
+
| AgentState.GroqModelSettings
|
|
437
|
+
| AgentState.DeepseekModelSettings
|
|
438
|
+
| AgentState.TogetherModelSettings
|
|
439
|
+
| AgentState.BedrockModelSettings
|
|
440
|
+
| null;
|
|
441
|
+
|
|
442
|
+
/**
|
|
443
|
+
* @deprecated Deprecated: Use `managed_group` field instead. The multi-agent group
|
|
444
|
+
* that this agent manages.
|
|
429
445
|
*/
|
|
430
446
|
multi_agent_group?: GroupsGroupsAPI.Group | null;
|
|
431
447
|
|
|
@@ -441,7 +457,7 @@ export interface AgentState {
|
|
|
441
457
|
project_id?: string | null;
|
|
442
458
|
|
|
443
459
|
/**
|
|
444
|
-
* The response format used by the agent
|
|
460
|
+
* The response format used by the agent
|
|
445
461
|
*/
|
|
446
462
|
response_format?: TextResponseFormat | JsonSchemaResponseFormat | JsonObjectResponseFormat | null;
|
|
447
463
|
|
|
@@ -488,7 +504,8 @@ export interface AgentState {
|
|
|
488
504
|
|
|
489
505
|
export namespace AgentState {
|
|
490
506
|
/**
|
|
491
|
-
* @deprecated The in-context memory of the
|
|
507
|
+
* @deprecated Deprecated: Use `blocks` field instead. The in-context memory of the
|
|
508
|
+
* agent.
|
|
492
509
|
*/
|
|
493
510
|
export interface Memory {
|
|
494
511
|
/**
|
|
@@ -628,15 +645,15 @@ export namespace AgentState {
|
|
|
628
645
|
}
|
|
629
646
|
|
|
630
647
|
/**
|
|
631
|
-
* Representation of a source, which is a collection of
|
|
632
|
-
*
|
|
633
|
-
* Parameters: id (str): The ID of the source name (str): The name of the source.
|
|
634
|
-
* embedding_config (EmbeddingConfig): The embedding configuration used by the
|
|
635
|
-
* source. user_id (str): The ID of the user that created the source. metadata
|
|
636
|
-
* (dict): Metadata associated with the source. description (str): The description
|
|
637
|
-
* of the source.
|
|
648
|
+
* (Deprecated: Use Folder) Representation of a source, which is a collection of
|
|
649
|
+
* files and passages.
|
|
638
650
|
*/
|
|
639
651
|
export interface Source {
|
|
652
|
+
/**
|
|
653
|
+
* The human-friendly ID of the Source
|
|
654
|
+
*/
|
|
655
|
+
id: string;
|
|
656
|
+
|
|
640
657
|
/**
|
|
641
658
|
* The embedding configuration used by the source.
|
|
642
659
|
*/
|
|
@@ -647,11 +664,6 @@ export namespace AgentState {
|
|
|
647
664
|
*/
|
|
648
665
|
name: string;
|
|
649
666
|
|
|
650
|
-
/**
|
|
651
|
-
* The human-friendly ID of the Source
|
|
652
|
-
*/
|
|
653
|
-
id?: string;
|
|
654
|
-
|
|
655
667
|
/**
|
|
656
668
|
* The timestamp when the source was created.
|
|
657
669
|
*/
|
|
@@ -692,6 +704,413 @@ export namespace AgentState {
|
|
|
692
704
|
*/
|
|
693
705
|
vector_db_provider?: ArchivesAPI.VectorDBProvider;
|
|
694
706
|
}
|
|
707
|
+
|
|
708
|
+
export interface OpenAIModelSettings {
|
|
709
|
+
/**
|
|
710
|
+
* The maximum number of tokens the model can generate.
|
|
711
|
+
*/
|
|
712
|
+
max_output_tokens?: number;
|
|
713
|
+
|
|
714
|
+
/**
|
|
715
|
+
* Whether to enable parallel tool calling.
|
|
716
|
+
*/
|
|
717
|
+
parallel_tool_calls?: boolean;
|
|
718
|
+
|
|
719
|
+
/**
|
|
720
|
+
* The provider of the model.
|
|
721
|
+
*/
|
|
722
|
+
provider?: 'openai';
|
|
723
|
+
|
|
724
|
+
/**
|
|
725
|
+
* The reasoning configuration for the model.
|
|
726
|
+
*/
|
|
727
|
+
reasoning?: OpenAIModelSettings.Reasoning;
|
|
728
|
+
|
|
729
|
+
/**
|
|
730
|
+
* The response format for the model.
|
|
731
|
+
*/
|
|
732
|
+
response_format?:
|
|
733
|
+
| AgentsAPI.TextResponseFormat
|
|
734
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
735
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
736
|
+
| null;
|
|
737
|
+
|
|
738
|
+
/**
|
|
739
|
+
* The temperature of the model.
|
|
740
|
+
*/
|
|
741
|
+
temperature?: number;
|
|
742
|
+
}
|
|
743
|
+
|
|
744
|
+
export namespace OpenAIModelSettings {
|
|
745
|
+
/**
|
|
746
|
+
* The reasoning configuration for the model.
|
|
747
|
+
*/
|
|
748
|
+
export interface Reasoning {
|
|
749
|
+
/**
|
|
750
|
+
* The reasoning effort to use when generating text reasoning models
|
|
751
|
+
*/
|
|
752
|
+
reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
|
|
753
|
+
}
|
|
754
|
+
}
|
|
755
|
+
|
|
756
|
+
export interface AnthropicModelSettings {
|
|
757
|
+
/**
|
|
758
|
+
* The maximum number of tokens the model can generate.
|
|
759
|
+
*/
|
|
760
|
+
max_output_tokens?: number;
|
|
761
|
+
|
|
762
|
+
/**
|
|
763
|
+
* Whether to enable parallel tool calling.
|
|
764
|
+
*/
|
|
765
|
+
parallel_tool_calls?: boolean;
|
|
766
|
+
|
|
767
|
+
/**
|
|
768
|
+
* The provider of the model.
|
|
769
|
+
*/
|
|
770
|
+
provider?: 'anthropic';
|
|
771
|
+
|
|
772
|
+
/**
|
|
773
|
+
* The temperature of the model.
|
|
774
|
+
*/
|
|
775
|
+
temperature?: number;
|
|
776
|
+
|
|
777
|
+
/**
|
|
778
|
+
* The thinking configuration for the model.
|
|
779
|
+
*/
|
|
780
|
+
thinking?: AnthropicModelSettings.Thinking;
|
|
781
|
+
|
|
782
|
+
/**
|
|
783
|
+
* Soft control for how verbose model output should be, used for GPT-5 models.
|
|
784
|
+
*/
|
|
785
|
+
verbosity?: 'low' | 'medium' | 'high' | null;
|
|
786
|
+
}
|
|
787
|
+
|
|
788
|
+
export namespace AnthropicModelSettings {
|
|
789
|
+
/**
|
|
790
|
+
* The thinking configuration for the model.
|
|
791
|
+
*/
|
|
792
|
+
export interface Thinking {
|
|
793
|
+
/**
|
|
794
|
+
* The maximum number of tokens the model can use for extended thinking.
|
|
795
|
+
*/
|
|
796
|
+
budget_tokens?: number;
|
|
797
|
+
|
|
798
|
+
/**
|
|
799
|
+
* The type of thinking to use.
|
|
800
|
+
*/
|
|
801
|
+
type?: 'enabled' | 'disabled';
|
|
802
|
+
}
|
|
803
|
+
}
|
|
804
|
+
|
|
805
|
+
export interface GoogleAIModelSettings {
|
|
806
|
+
/**
|
|
807
|
+
* The maximum number of tokens the model can generate.
|
|
808
|
+
*/
|
|
809
|
+
max_output_tokens?: number;
|
|
810
|
+
|
|
811
|
+
/**
|
|
812
|
+
* Whether to enable parallel tool calling.
|
|
813
|
+
*/
|
|
814
|
+
parallel_tool_calls?: boolean;
|
|
815
|
+
|
|
816
|
+
/**
|
|
817
|
+
* The provider of the model.
|
|
818
|
+
*/
|
|
819
|
+
provider?: 'google_ai';
|
|
820
|
+
|
|
821
|
+
/**
|
|
822
|
+
* The response schema for the model.
|
|
823
|
+
*/
|
|
824
|
+
response_schema?:
|
|
825
|
+
| AgentsAPI.TextResponseFormat
|
|
826
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
827
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
828
|
+
| null;
|
|
829
|
+
|
|
830
|
+
/**
|
|
831
|
+
* The temperature of the model.
|
|
832
|
+
*/
|
|
833
|
+
temperature?: number;
|
|
834
|
+
|
|
835
|
+
/**
|
|
836
|
+
* The thinking configuration for the model.
|
|
837
|
+
*/
|
|
838
|
+
thinking_config?: GoogleAIModelSettings.ThinkingConfig;
|
|
839
|
+
}
|
|
840
|
+
|
|
841
|
+
export namespace GoogleAIModelSettings {
|
|
842
|
+
/**
|
|
843
|
+
* The thinking configuration for the model.
|
|
844
|
+
*/
|
|
845
|
+
export interface ThinkingConfig {
|
|
846
|
+
/**
|
|
847
|
+
* Whether to include thoughts in the model's response.
|
|
848
|
+
*/
|
|
849
|
+
include_thoughts?: boolean;
|
|
850
|
+
|
|
851
|
+
/**
|
|
852
|
+
* The thinking budget for the model.
|
|
853
|
+
*/
|
|
854
|
+
thinking_budget?: number;
|
|
855
|
+
}
|
|
856
|
+
}
|
|
857
|
+
|
|
858
|
+
export interface GoogleVertexModelSettings {
|
|
859
|
+
/**
|
|
860
|
+
* The maximum number of tokens the model can generate.
|
|
861
|
+
*/
|
|
862
|
+
max_output_tokens?: number;
|
|
863
|
+
|
|
864
|
+
/**
|
|
865
|
+
* Whether to enable parallel tool calling.
|
|
866
|
+
*/
|
|
867
|
+
parallel_tool_calls?: boolean;
|
|
868
|
+
|
|
869
|
+
/**
|
|
870
|
+
* The provider of the model.
|
|
871
|
+
*/
|
|
872
|
+
provider?: 'google_vertex';
|
|
873
|
+
|
|
874
|
+
/**
|
|
875
|
+
* The response schema for the model.
|
|
876
|
+
*/
|
|
877
|
+
response_schema?:
|
|
878
|
+
| AgentsAPI.TextResponseFormat
|
|
879
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
880
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
881
|
+
| null;
|
|
882
|
+
|
|
883
|
+
/**
|
|
884
|
+
* The temperature of the model.
|
|
885
|
+
*/
|
|
886
|
+
temperature?: number;
|
|
887
|
+
|
|
888
|
+
/**
|
|
889
|
+
* The thinking configuration for the model.
|
|
890
|
+
*/
|
|
891
|
+
thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
|
|
892
|
+
}
|
|
893
|
+
|
|
894
|
+
export namespace GoogleVertexModelSettings {
|
|
895
|
+
/**
|
|
896
|
+
* The thinking configuration for the model.
|
|
897
|
+
*/
|
|
898
|
+
export interface ThinkingConfig {
|
|
899
|
+
/**
|
|
900
|
+
* Whether to include thoughts in the model's response.
|
|
901
|
+
*/
|
|
902
|
+
include_thoughts?: boolean;
|
|
903
|
+
|
|
904
|
+
/**
|
|
905
|
+
* The thinking budget for the model.
|
|
906
|
+
*/
|
|
907
|
+
thinking_budget?: number;
|
|
908
|
+
}
|
|
909
|
+
}
|
|
910
|
+
|
|
911
|
+
/**
|
|
912
|
+
* Azure OpenAI model configuration (OpenAI-compatible).
|
|
913
|
+
*/
|
|
914
|
+
export interface AzureModelSettings {
|
|
915
|
+
/**
|
|
916
|
+
* The maximum number of tokens the model can generate.
|
|
917
|
+
*/
|
|
918
|
+
max_output_tokens?: number;
|
|
919
|
+
|
|
920
|
+
/**
|
|
921
|
+
* Whether to enable parallel tool calling.
|
|
922
|
+
*/
|
|
923
|
+
parallel_tool_calls?: boolean;
|
|
924
|
+
|
|
925
|
+
/**
|
|
926
|
+
* The provider of the model.
|
|
927
|
+
*/
|
|
928
|
+
provider?: 'azure';
|
|
929
|
+
|
|
930
|
+
/**
|
|
931
|
+
* The response format for the model.
|
|
932
|
+
*/
|
|
933
|
+
response_format?:
|
|
934
|
+
| AgentsAPI.TextResponseFormat
|
|
935
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
936
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
937
|
+
| null;
|
|
938
|
+
|
|
939
|
+
/**
|
|
940
|
+
* The temperature of the model.
|
|
941
|
+
*/
|
|
942
|
+
temperature?: number;
|
|
943
|
+
}
|
|
944
|
+
|
|
945
|
+
/**
|
|
946
|
+
* xAI model configuration (OpenAI-compatible).
|
|
947
|
+
*/
|
|
948
|
+
export interface XaiModelSettings {
|
|
949
|
+
/**
|
|
950
|
+
* The maximum number of tokens the model can generate.
|
|
951
|
+
*/
|
|
952
|
+
max_output_tokens?: number;
|
|
953
|
+
|
|
954
|
+
/**
|
|
955
|
+
* Whether to enable parallel tool calling.
|
|
956
|
+
*/
|
|
957
|
+
parallel_tool_calls?: boolean;
|
|
958
|
+
|
|
959
|
+
/**
|
|
960
|
+
* The provider of the model.
|
|
961
|
+
*/
|
|
962
|
+
provider?: 'xai';
|
|
963
|
+
|
|
964
|
+
/**
|
|
965
|
+
* The response format for the model.
|
|
966
|
+
*/
|
|
967
|
+
response_format?:
|
|
968
|
+
| AgentsAPI.TextResponseFormat
|
|
969
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
970
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
971
|
+
| null;
|
|
972
|
+
|
|
973
|
+
/**
|
|
974
|
+
* The temperature of the model.
|
|
975
|
+
*/
|
|
976
|
+
temperature?: number;
|
|
977
|
+
}
|
|
978
|
+
|
|
979
|
+
/**
|
|
980
|
+
* Groq model configuration (OpenAI-compatible).
|
|
981
|
+
*/
|
|
982
|
+
export interface GroqModelSettings {
|
|
983
|
+
/**
|
|
984
|
+
* The maximum number of tokens the model can generate.
|
|
985
|
+
*/
|
|
986
|
+
max_output_tokens?: number;
|
|
987
|
+
|
|
988
|
+
/**
|
|
989
|
+
* Whether to enable parallel tool calling.
|
|
990
|
+
*/
|
|
991
|
+
parallel_tool_calls?: boolean;
|
|
992
|
+
|
|
993
|
+
/**
|
|
994
|
+
* The provider of the model.
|
|
995
|
+
*/
|
|
996
|
+
provider?: 'groq';
|
|
997
|
+
|
|
998
|
+
/**
|
|
999
|
+
* The response format for the model.
|
|
1000
|
+
*/
|
|
1001
|
+
response_format?:
|
|
1002
|
+
| AgentsAPI.TextResponseFormat
|
|
1003
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1004
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1005
|
+
| null;
|
|
1006
|
+
|
|
1007
|
+
/**
|
|
1008
|
+
* The temperature of the model.
|
|
1009
|
+
*/
|
|
1010
|
+
temperature?: number;
|
|
1011
|
+
}
|
|
1012
|
+
|
|
1013
|
+
/**
|
|
1014
|
+
* Deepseek model configuration (OpenAI-compatible).
|
|
1015
|
+
*/
|
|
1016
|
+
export interface DeepseekModelSettings {
|
|
1017
|
+
/**
|
|
1018
|
+
* The maximum number of tokens the model can generate.
|
|
1019
|
+
*/
|
|
1020
|
+
max_output_tokens?: number;
|
|
1021
|
+
|
|
1022
|
+
/**
|
|
1023
|
+
* Whether to enable parallel tool calling.
|
|
1024
|
+
*/
|
|
1025
|
+
parallel_tool_calls?: boolean;
|
|
1026
|
+
|
|
1027
|
+
/**
|
|
1028
|
+
* The provider of the model.
|
|
1029
|
+
*/
|
|
1030
|
+
provider?: 'deepseek';
|
|
1031
|
+
|
|
1032
|
+
/**
|
|
1033
|
+
* The response format for the model.
|
|
1034
|
+
*/
|
|
1035
|
+
response_format?:
|
|
1036
|
+
| AgentsAPI.TextResponseFormat
|
|
1037
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1038
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1039
|
+
| null;
|
|
1040
|
+
|
|
1041
|
+
/**
|
|
1042
|
+
* The temperature of the model.
|
|
1043
|
+
*/
|
|
1044
|
+
temperature?: number;
|
|
1045
|
+
}
|
|
1046
|
+
|
|
1047
|
+
/**
|
|
1048
|
+
* Together AI model configuration (OpenAI-compatible).
|
|
1049
|
+
*/
|
|
1050
|
+
export interface TogetherModelSettings {
|
|
1051
|
+
/**
|
|
1052
|
+
* The maximum number of tokens the model can generate.
|
|
1053
|
+
*/
|
|
1054
|
+
max_output_tokens?: number;
|
|
1055
|
+
|
|
1056
|
+
/**
|
|
1057
|
+
* Whether to enable parallel tool calling.
|
|
1058
|
+
*/
|
|
1059
|
+
parallel_tool_calls?: boolean;
|
|
1060
|
+
|
|
1061
|
+
/**
|
|
1062
|
+
* The provider of the model.
|
|
1063
|
+
*/
|
|
1064
|
+
provider?: 'together';
|
|
1065
|
+
|
|
1066
|
+
/**
|
|
1067
|
+
* The response format for the model.
|
|
1068
|
+
*/
|
|
1069
|
+
response_format?:
|
|
1070
|
+
| AgentsAPI.TextResponseFormat
|
|
1071
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1072
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1073
|
+
| null;
|
|
1074
|
+
|
|
1075
|
+
/**
|
|
1076
|
+
* The temperature of the model.
|
|
1077
|
+
*/
|
|
1078
|
+
temperature?: number;
|
|
1079
|
+
}
|
|
1080
|
+
|
|
1081
|
+
/**
|
|
1082
|
+
* AWS Bedrock model configuration.
|
|
1083
|
+
*/
|
|
1084
|
+
export interface BedrockModelSettings {
|
|
1085
|
+
/**
|
|
1086
|
+
* The maximum number of tokens the model can generate.
|
|
1087
|
+
*/
|
|
1088
|
+
max_output_tokens?: number;
|
|
1089
|
+
|
|
1090
|
+
/**
|
|
1091
|
+
* Whether to enable parallel tool calling.
|
|
1092
|
+
*/
|
|
1093
|
+
parallel_tool_calls?: boolean;
|
|
1094
|
+
|
|
1095
|
+
/**
|
|
1096
|
+
* The provider of the model.
|
|
1097
|
+
*/
|
|
1098
|
+
provider?: 'bedrock';
|
|
1099
|
+
|
|
1100
|
+
/**
|
|
1101
|
+
* The response format for the model.
|
|
1102
|
+
*/
|
|
1103
|
+
response_format?:
|
|
1104
|
+
| AgentsAPI.TextResponseFormat
|
|
1105
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1106
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1107
|
+
| null;
|
|
1108
|
+
|
|
1109
|
+
/**
|
|
1110
|
+
* The temperature of the model.
|
|
1111
|
+
*/
|
|
1112
|
+
temperature?: number;
|
|
1113
|
+
}
|
|
695
1114
|
}
|
|
696
1115
|
|
|
697
1116
|
/**
|
|
@@ -1031,8 +1450,6 @@ export interface TextResponseFormat {
|
|
|
1031
1450
|
|
|
1032
1451
|
export type AgentDeleteResponse = unknown;
|
|
1033
1452
|
|
|
1034
|
-
export type AgentCountResponse = number;
|
|
1035
|
-
|
|
1036
1453
|
export type AgentExportFileResponse = string;
|
|
1037
1454
|
|
|
1038
1455
|
/**
|
|
@@ -1052,7 +1469,7 @@ export interface AgentCreateParams {
|
|
|
1052
1469
|
agent_type?: AgentType;
|
|
1053
1470
|
|
|
1054
1471
|
/**
|
|
1055
|
-
* The base template id of the agent.
|
|
1472
|
+
* @deprecated Deprecated: No longer used. The base template id of the agent.
|
|
1056
1473
|
*/
|
|
1057
1474
|
base_template_id?: string | null;
|
|
1058
1475
|
|
|
@@ -1072,13 +1489,13 @@ export interface AgentCreateParams {
|
|
|
1072
1489
|
description?: string | null;
|
|
1073
1490
|
|
|
1074
1491
|
/**
|
|
1075
|
-
* The embedding
|
|
1076
|
-
* provider/model-name.
|
|
1492
|
+
* The embedding model handle used by the agent (format: provider/model-name).
|
|
1077
1493
|
*/
|
|
1078
1494
|
embedding?: string | null;
|
|
1079
1495
|
|
|
1080
1496
|
/**
|
|
1081
|
-
* The embedding chunk size used by the
|
|
1497
|
+
* @deprecated Deprecated: No longer used. The embedding chunk size used by the
|
|
1498
|
+
* agent.
|
|
1082
1499
|
*/
|
|
1083
1500
|
embedding_chunk_size?: number | null;
|
|
1084
1501
|
|
|
@@ -1088,6 +1505,7 @@ export interface AgentCreateParams {
|
|
|
1088
1505
|
embedding_config?: ModelsAPI.EmbeddingConfig | null;
|
|
1089
1506
|
|
|
1090
1507
|
/**
|
|
1508
|
+
* @deprecated Deprecated: Use `model` field to configure reasoning instead.
|
|
1091
1509
|
* Whether to enable internal extended thinking step for a reasoner model.
|
|
1092
1510
|
*/
|
|
1093
1511
|
enable_reasoner?: boolean | null;
|
|
@@ -1098,12 +1516,14 @@ export interface AgentCreateParams {
|
|
|
1098
1516
|
enable_sleeptime?: boolean | null;
|
|
1099
1517
|
|
|
1100
1518
|
/**
|
|
1101
|
-
* Deprecated: please use the 'create agents from a template' endpoint
|
|
1519
|
+
* @deprecated Deprecated: please use the 'create agents from a template' endpoint
|
|
1520
|
+
* instead.
|
|
1102
1521
|
*/
|
|
1103
1522
|
from_template?: string | null;
|
|
1104
1523
|
|
|
1105
1524
|
/**
|
|
1106
|
-
* If set to True, the agent will be
|
|
1525
|
+
* @deprecated Deprecated: No longer used. If set to True, the agent will be
|
|
1526
|
+
* hidden.
|
|
1107
1527
|
*/
|
|
1108
1528
|
hidden?: boolean | null;
|
|
1109
1529
|
|
|
@@ -1124,8 +1544,8 @@ export interface AgentCreateParams {
|
|
|
1124
1544
|
include_base_tools?: boolean;
|
|
1125
1545
|
|
|
1126
1546
|
/**
|
|
1127
|
-
* If true, automatically creates and attaches a default data source
|
|
1128
|
-
* agent.
|
|
1547
|
+
* @deprecated If true, automatically creates and attaches a default data source
|
|
1548
|
+
* for this agent.
|
|
1129
1549
|
*/
|
|
1130
1550
|
include_default_source?: boolean;
|
|
1131
1551
|
|
|
@@ -1152,14 +1572,14 @@ export interface AgentCreateParams {
|
|
|
1152
1572
|
max_files_open?: number | null;
|
|
1153
1573
|
|
|
1154
1574
|
/**
|
|
1155
|
-
*
|
|
1156
|
-
*
|
|
1575
|
+
* @deprecated Deprecated: Use `model` field to configure reasoning tokens instead.
|
|
1576
|
+
* The maximum number of tokens to generate for reasoning step.
|
|
1157
1577
|
*/
|
|
1158
1578
|
max_reasoning_tokens?: number | null;
|
|
1159
1579
|
|
|
1160
1580
|
/**
|
|
1161
|
-
*
|
|
1162
|
-
*
|
|
1581
|
+
* @deprecated Deprecated: Use `model` field to configure max output tokens
|
|
1582
|
+
* instead. The maximum number of tokens to generate, including reasoning step.
|
|
1163
1583
|
*/
|
|
1164
1584
|
max_tokens?: number | null;
|
|
1165
1585
|
|
|
@@ -1169,7 +1589,8 @@ export interface AgentCreateParams {
|
|
|
1169
1589
|
memory_blocks?: Array<BlocksBlocksAPI.CreateBlock> | null;
|
|
1170
1590
|
|
|
1171
1591
|
/**
|
|
1172
|
-
*
|
|
1592
|
+
* @deprecated Deprecated: Only relevant for creating agents from a template. Use
|
|
1593
|
+
* the 'create agents from a template' endpoint instead.
|
|
1173
1594
|
*/
|
|
1174
1595
|
memory_variables?: { [key: string]: string } | null;
|
|
1175
1596
|
|
|
@@ -1186,18 +1607,34 @@ export interface AgentCreateParams {
|
|
|
1186
1607
|
metadata?: { [key: string]: unknown } | null;
|
|
1187
1608
|
|
|
1188
1609
|
/**
|
|
1189
|
-
* The
|
|
1190
|
-
* provider/model-name, as an alternative to specifying llm_config.
|
|
1610
|
+
* The model handle for the agent to use (format: provider/model-name).
|
|
1191
1611
|
*/
|
|
1192
1612
|
model?: string | null;
|
|
1193
1613
|
|
|
1614
|
+
/**
|
|
1615
|
+
* The model settings for the agent.
|
|
1616
|
+
*/
|
|
1617
|
+
model_settings?:
|
|
1618
|
+
| AgentCreateParams.OpenAIModelSettings
|
|
1619
|
+
| AgentCreateParams.AnthropicModelSettings
|
|
1620
|
+
| AgentCreateParams.GoogleAIModelSettings
|
|
1621
|
+
| AgentCreateParams.GoogleVertexModelSettings
|
|
1622
|
+
| AgentCreateParams.AzureModelSettings
|
|
1623
|
+
| AgentCreateParams.XaiModelSettings
|
|
1624
|
+
| AgentCreateParams.GroqModelSettings
|
|
1625
|
+
| AgentCreateParams.DeepseekModelSettings
|
|
1626
|
+
| AgentCreateParams.TogetherModelSettings
|
|
1627
|
+
| AgentCreateParams.BedrockModelSettings
|
|
1628
|
+
| null;
|
|
1629
|
+
|
|
1194
1630
|
/**
|
|
1195
1631
|
* The name of the agent.
|
|
1196
1632
|
*/
|
|
1197
1633
|
name?: string;
|
|
1198
1634
|
|
|
1199
1635
|
/**
|
|
1200
|
-
*
|
|
1636
|
+
* @deprecated Deprecated: Use `model` field to configure parallel tool calls
|
|
1637
|
+
* instead. If set to True, enables parallel tool calling.
|
|
1201
1638
|
*/
|
|
1202
1639
|
parallel_tool_calls?: boolean | null;
|
|
1203
1640
|
|
|
@@ -1209,17 +1646,19 @@ export interface AgentCreateParams {
|
|
|
1209
1646
|
|
|
1210
1647
|
/**
|
|
1211
1648
|
* @deprecated Deprecated: Project should now be passed via the X-Project header
|
|
1212
|
-
* instead of in the request body. If using the
|
|
1213
|
-
* x_project
|
|
1649
|
+
* instead of in the request body. If using the SDK, this can be done via the
|
|
1650
|
+
* x_project parameter.
|
|
1214
1651
|
*/
|
|
1215
1652
|
project?: string | null;
|
|
1216
1653
|
|
|
1217
1654
|
/**
|
|
1218
|
-
* The id of the project the agent belongs
|
|
1655
|
+
* @deprecated Deprecated: No longer used. The id of the project the agent belongs
|
|
1656
|
+
* to.
|
|
1219
1657
|
*/
|
|
1220
1658
|
project_id?: string | null;
|
|
1221
1659
|
|
|
1222
1660
|
/**
|
|
1661
|
+
* @deprecated Deprecated: Use `model` field to configure reasoning instead.
|
|
1223
1662
|
* Whether to enable reasoning for this agent.
|
|
1224
1663
|
*/
|
|
1225
1664
|
reasoning?: boolean | null;
|
|
@@ -1250,12 +1689,13 @@ export interface AgentCreateParams {
|
|
|
1250
1689
|
tags?: Array<string> | null;
|
|
1251
1690
|
|
|
1252
1691
|
/**
|
|
1253
|
-
* Deprecated: No longer used
|
|
1692
|
+
* @deprecated Deprecated: No longer used.
|
|
1254
1693
|
*/
|
|
1255
1694
|
template?: boolean;
|
|
1256
1695
|
|
|
1257
1696
|
/**
|
|
1258
|
-
* The id of the template the agent belongs
|
|
1697
|
+
* @deprecated Deprecated: No longer used. The id of the template the agent belongs
|
|
1698
|
+
* to.
|
|
1259
1699
|
*/
|
|
1260
1700
|
template_id?: string | null;
|
|
1261
1701
|
|
|
@@ -1265,7 +1705,8 @@ export interface AgentCreateParams {
|
|
|
1265
1705
|
timezone?: string | null;
|
|
1266
1706
|
|
|
1267
1707
|
/**
|
|
1268
|
-
* Deprecated:
|
|
1708
|
+
* @deprecated Deprecated: Use `secrets` field instead. Environment variables for
|
|
1709
|
+
* tool execution.
|
|
1269
1710
|
*/
|
|
1270
1711
|
tool_exec_environment_variables?: { [key: string]: string } | null;
|
|
1271
1712
|
|
|
@@ -1295,6 +1736,415 @@ export interface AgentCreateParams {
|
|
|
1295
1736
|
tools?: Array<string> | null;
|
|
1296
1737
|
}
|
|
1297
1738
|
|
|
1739
|
+
export namespace AgentCreateParams {
|
|
1740
|
+
export interface OpenAIModelSettings {
|
|
1741
|
+
/**
|
|
1742
|
+
* The maximum number of tokens the model can generate.
|
|
1743
|
+
*/
|
|
1744
|
+
max_output_tokens?: number;
|
|
1745
|
+
|
|
1746
|
+
/**
|
|
1747
|
+
* Whether to enable parallel tool calling.
|
|
1748
|
+
*/
|
|
1749
|
+
parallel_tool_calls?: boolean;
|
|
1750
|
+
|
|
1751
|
+
/**
|
|
1752
|
+
* The provider of the model.
|
|
1753
|
+
*/
|
|
1754
|
+
provider?: 'openai';
|
|
1755
|
+
|
|
1756
|
+
/**
|
|
1757
|
+
* The reasoning configuration for the model.
|
|
1758
|
+
*/
|
|
1759
|
+
reasoning?: OpenAIModelSettings.Reasoning;
|
|
1760
|
+
|
|
1761
|
+
/**
|
|
1762
|
+
* The response format for the model.
|
|
1763
|
+
*/
|
|
1764
|
+
response_format?:
|
|
1765
|
+
| AgentsAPI.TextResponseFormat
|
|
1766
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1767
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1768
|
+
| null;
|
|
1769
|
+
|
|
1770
|
+
/**
|
|
1771
|
+
* The temperature of the model.
|
|
1772
|
+
*/
|
|
1773
|
+
temperature?: number;
|
|
1774
|
+
}
|
|
1775
|
+
|
|
1776
|
+
export namespace OpenAIModelSettings {
|
|
1777
|
+
/**
|
|
1778
|
+
* The reasoning configuration for the model.
|
|
1779
|
+
*/
|
|
1780
|
+
export interface Reasoning {
|
|
1781
|
+
/**
|
|
1782
|
+
* The reasoning effort to use when generating text reasoning models
|
|
1783
|
+
*/
|
|
1784
|
+
reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
|
|
1785
|
+
}
|
|
1786
|
+
}
|
|
1787
|
+
|
|
1788
|
+
export interface AnthropicModelSettings {
|
|
1789
|
+
/**
|
|
1790
|
+
* The maximum number of tokens the model can generate.
|
|
1791
|
+
*/
|
|
1792
|
+
max_output_tokens?: number;
|
|
1793
|
+
|
|
1794
|
+
/**
|
|
1795
|
+
* Whether to enable parallel tool calling.
|
|
1796
|
+
*/
|
|
1797
|
+
parallel_tool_calls?: boolean;
|
|
1798
|
+
|
|
1799
|
+
/**
|
|
1800
|
+
* The provider of the model.
|
|
1801
|
+
*/
|
|
1802
|
+
provider?: 'anthropic';
|
|
1803
|
+
|
|
1804
|
+
/**
|
|
1805
|
+
* The temperature of the model.
|
|
1806
|
+
*/
|
|
1807
|
+
temperature?: number;
|
|
1808
|
+
|
|
1809
|
+
/**
|
|
1810
|
+
* The thinking configuration for the model.
|
|
1811
|
+
*/
|
|
1812
|
+
thinking?: AnthropicModelSettings.Thinking;
|
|
1813
|
+
|
|
1814
|
+
/**
|
|
1815
|
+
* Soft control for how verbose model output should be, used for GPT-5 models.
|
|
1816
|
+
*/
|
|
1817
|
+
verbosity?: 'low' | 'medium' | 'high' | null;
|
|
1818
|
+
}
|
|
1819
|
+
|
|
1820
|
+
export namespace AnthropicModelSettings {
|
|
1821
|
+
/**
|
|
1822
|
+
* The thinking configuration for the model.
|
|
1823
|
+
*/
|
|
1824
|
+
export interface Thinking {
|
|
1825
|
+
/**
|
|
1826
|
+
* The maximum number of tokens the model can use for extended thinking.
|
|
1827
|
+
*/
|
|
1828
|
+
budget_tokens?: number;
|
|
1829
|
+
|
|
1830
|
+
/**
|
|
1831
|
+
* The type of thinking to use.
|
|
1832
|
+
*/
|
|
1833
|
+
type?: 'enabled' | 'disabled';
|
|
1834
|
+
}
|
|
1835
|
+
}
|
|
1836
|
+
|
|
1837
|
+
export interface GoogleAIModelSettings {
|
|
1838
|
+
/**
|
|
1839
|
+
* The maximum number of tokens the model can generate.
|
|
1840
|
+
*/
|
|
1841
|
+
max_output_tokens?: number;
|
|
1842
|
+
|
|
1843
|
+
/**
|
|
1844
|
+
* Whether to enable parallel tool calling.
|
|
1845
|
+
*/
|
|
1846
|
+
parallel_tool_calls?: boolean;
|
|
1847
|
+
|
|
1848
|
+
/**
|
|
1849
|
+
* The provider of the model.
|
|
1850
|
+
*/
|
|
1851
|
+
provider?: 'google_ai';
|
|
1852
|
+
|
|
1853
|
+
/**
|
|
1854
|
+
* The response schema for the model.
|
|
1855
|
+
*/
|
|
1856
|
+
response_schema?:
|
|
1857
|
+
| AgentsAPI.TextResponseFormat
|
|
1858
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1859
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1860
|
+
| null;
|
|
1861
|
+
|
|
1862
|
+
/**
|
|
1863
|
+
* The temperature of the model.
|
|
1864
|
+
*/
|
|
1865
|
+
temperature?: number;
|
|
1866
|
+
|
|
1867
|
+
/**
|
|
1868
|
+
* The thinking configuration for the model.
|
|
1869
|
+
*/
|
|
1870
|
+
thinking_config?: GoogleAIModelSettings.ThinkingConfig;
|
|
1871
|
+
}
|
|
1872
|
+
|
|
1873
|
+
export namespace GoogleAIModelSettings {
|
|
1874
|
+
/**
|
|
1875
|
+
* The thinking configuration for the model.
|
|
1876
|
+
*/
|
|
1877
|
+
export interface ThinkingConfig {
|
|
1878
|
+
/**
|
|
1879
|
+
* Whether to include thoughts in the model's response.
|
|
1880
|
+
*/
|
|
1881
|
+
include_thoughts?: boolean;
|
|
1882
|
+
|
|
1883
|
+
/**
|
|
1884
|
+
* The thinking budget for the model.
|
|
1885
|
+
*/
|
|
1886
|
+
thinking_budget?: number;
|
|
1887
|
+
}
|
|
1888
|
+
}
|
|
1889
|
+
|
|
1890
|
+
export interface GoogleVertexModelSettings {
|
|
1891
|
+
/**
|
|
1892
|
+
* The maximum number of tokens the model can generate.
|
|
1893
|
+
*/
|
|
1894
|
+
max_output_tokens?: number;
|
|
1895
|
+
|
|
1896
|
+
/**
|
|
1897
|
+
* Whether to enable parallel tool calling.
|
|
1898
|
+
*/
|
|
1899
|
+
parallel_tool_calls?: boolean;
|
|
1900
|
+
|
|
1901
|
+
/**
|
|
1902
|
+
* The provider of the model.
|
|
1903
|
+
*/
|
|
1904
|
+
provider?: 'google_vertex';
|
|
1905
|
+
|
|
1906
|
+
/**
|
|
1907
|
+
* The response schema for the model.
|
|
1908
|
+
*/
|
|
1909
|
+
response_schema?:
|
|
1910
|
+
| AgentsAPI.TextResponseFormat
|
|
1911
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1912
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1913
|
+
| null;
|
|
1914
|
+
|
|
1915
|
+
/**
|
|
1916
|
+
* The temperature of the model.
|
|
1917
|
+
*/
|
|
1918
|
+
temperature?: number;
|
|
1919
|
+
|
|
1920
|
+
/**
|
|
1921
|
+
* The thinking configuration for the model.
|
|
1922
|
+
*/
|
|
1923
|
+
thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
|
|
1924
|
+
}
|
|
1925
|
+
|
|
1926
|
+
export namespace GoogleVertexModelSettings {
|
|
1927
|
+
/**
|
|
1928
|
+
* The thinking configuration for the model.
|
|
1929
|
+
*/
|
|
1930
|
+
export interface ThinkingConfig {
|
|
1931
|
+
/**
|
|
1932
|
+
* Whether to include thoughts in the model's response.
|
|
1933
|
+
*/
|
|
1934
|
+
include_thoughts?: boolean;
|
|
1935
|
+
|
|
1936
|
+
/**
|
|
1937
|
+
* The thinking budget for the model.
|
|
1938
|
+
*/
|
|
1939
|
+
thinking_budget?: number;
|
|
1940
|
+
}
|
|
1941
|
+
}
|
|
1942
|
+
|
|
1943
|
+
/**
|
|
1944
|
+
* Azure OpenAI model configuration (OpenAI-compatible).
|
|
1945
|
+
*/
|
|
1946
|
+
export interface AzureModelSettings {
|
|
1947
|
+
/**
|
|
1948
|
+
* The maximum number of tokens the model can generate.
|
|
1949
|
+
*/
|
|
1950
|
+
max_output_tokens?: number;
|
|
1951
|
+
|
|
1952
|
+
/**
|
|
1953
|
+
* Whether to enable parallel tool calling.
|
|
1954
|
+
*/
|
|
1955
|
+
parallel_tool_calls?: boolean;
|
|
1956
|
+
|
|
1957
|
+
/**
|
|
1958
|
+
* The provider of the model.
|
|
1959
|
+
*/
|
|
1960
|
+
provider?: 'azure';
|
|
1961
|
+
|
|
1962
|
+
/**
|
|
1963
|
+
* The response format for the model.
|
|
1964
|
+
*/
|
|
1965
|
+
response_format?:
|
|
1966
|
+
| AgentsAPI.TextResponseFormat
|
|
1967
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
1968
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
1969
|
+
| null;
|
|
1970
|
+
|
|
1971
|
+
/**
|
|
1972
|
+
* The temperature of the model.
|
|
1973
|
+
*/
|
|
1974
|
+
temperature?: number;
|
|
1975
|
+
}
|
|
1976
|
+
|
|
1977
|
+
/**
|
|
1978
|
+
* xAI model configuration (OpenAI-compatible).
|
|
1979
|
+
*/
|
|
1980
|
+
export interface XaiModelSettings {
|
|
1981
|
+
/**
|
|
1982
|
+
* The maximum number of tokens the model can generate.
|
|
1983
|
+
*/
|
|
1984
|
+
max_output_tokens?: number;
|
|
1985
|
+
|
|
1986
|
+
/**
|
|
1987
|
+
* Whether to enable parallel tool calling.
|
|
1988
|
+
*/
|
|
1989
|
+
parallel_tool_calls?: boolean;
|
|
1990
|
+
|
|
1991
|
+
/**
|
|
1992
|
+
* The provider of the model.
|
|
1993
|
+
*/
|
|
1994
|
+
provider?: 'xai';
|
|
1995
|
+
|
|
1996
|
+
/**
|
|
1997
|
+
* The response format for the model.
|
|
1998
|
+
*/
|
|
1999
|
+
response_format?:
|
|
2000
|
+
| AgentsAPI.TextResponseFormat
|
|
2001
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2002
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2003
|
+
| null;
|
|
2004
|
+
|
|
2005
|
+
/**
|
|
2006
|
+
* The temperature of the model.
|
|
2007
|
+
*/
|
|
2008
|
+
temperature?: number;
|
|
2009
|
+
}
|
|
2010
|
+
|
|
2011
|
+
/**
|
|
2012
|
+
* Groq model configuration (OpenAI-compatible).
|
|
2013
|
+
*/
|
|
2014
|
+
export interface GroqModelSettings {
|
|
2015
|
+
/**
|
|
2016
|
+
* The maximum number of tokens the model can generate.
|
|
2017
|
+
*/
|
|
2018
|
+
max_output_tokens?: number;
|
|
2019
|
+
|
|
2020
|
+
/**
|
|
2021
|
+
* Whether to enable parallel tool calling.
|
|
2022
|
+
*/
|
|
2023
|
+
parallel_tool_calls?: boolean;
|
|
2024
|
+
|
|
2025
|
+
/**
|
|
2026
|
+
* The provider of the model.
|
|
2027
|
+
*/
|
|
2028
|
+
provider?: 'groq';
|
|
2029
|
+
|
|
2030
|
+
/**
|
|
2031
|
+
* The response format for the model.
|
|
2032
|
+
*/
|
|
2033
|
+
response_format?:
|
|
2034
|
+
| AgentsAPI.TextResponseFormat
|
|
2035
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2036
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2037
|
+
| null;
|
|
2038
|
+
|
|
2039
|
+
/**
|
|
2040
|
+
* The temperature of the model.
|
|
2041
|
+
*/
|
|
2042
|
+
temperature?: number;
|
|
2043
|
+
}
|
|
2044
|
+
|
|
2045
|
+
/**
|
|
2046
|
+
* Deepseek model configuration (OpenAI-compatible).
|
|
2047
|
+
*/
|
|
2048
|
+
export interface DeepseekModelSettings {
|
|
2049
|
+
/**
|
|
2050
|
+
* The maximum number of tokens the model can generate.
|
|
2051
|
+
*/
|
|
2052
|
+
max_output_tokens?: number;
|
|
2053
|
+
|
|
2054
|
+
/**
|
|
2055
|
+
* Whether to enable parallel tool calling.
|
|
2056
|
+
*/
|
|
2057
|
+
parallel_tool_calls?: boolean;
|
|
2058
|
+
|
|
2059
|
+
/**
|
|
2060
|
+
* The provider of the model.
|
|
2061
|
+
*/
|
|
2062
|
+
provider?: 'deepseek';
|
|
2063
|
+
|
|
2064
|
+
/**
|
|
2065
|
+
* The response format for the model.
|
|
2066
|
+
*/
|
|
2067
|
+
response_format?:
|
|
2068
|
+
| AgentsAPI.TextResponseFormat
|
|
2069
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2070
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2071
|
+
| null;
|
|
2072
|
+
|
|
2073
|
+
/**
|
|
2074
|
+
* The temperature of the model.
|
|
2075
|
+
*/
|
|
2076
|
+
temperature?: number;
|
|
2077
|
+
}
|
|
2078
|
+
|
|
2079
|
+
/**
|
|
2080
|
+
* Together AI model configuration (OpenAI-compatible).
|
|
2081
|
+
*/
|
|
2082
|
+
export interface TogetherModelSettings {
|
|
2083
|
+
/**
|
|
2084
|
+
* The maximum number of tokens the model can generate.
|
|
2085
|
+
*/
|
|
2086
|
+
max_output_tokens?: number;
|
|
2087
|
+
|
|
2088
|
+
/**
|
|
2089
|
+
* Whether to enable parallel tool calling.
|
|
2090
|
+
*/
|
|
2091
|
+
parallel_tool_calls?: boolean;
|
|
2092
|
+
|
|
2093
|
+
/**
|
|
2094
|
+
* The provider of the model.
|
|
2095
|
+
*/
|
|
2096
|
+
provider?: 'together';
|
|
2097
|
+
|
|
2098
|
+
/**
|
|
2099
|
+
* The response format for the model.
|
|
2100
|
+
*/
|
|
2101
|
+
response_format?:
|
|
2102
|
+
| AgentsAPI.TextResponseFormat
|
|
2103
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2104
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2105
|
+
| null;
|
|
2106
|
+
|
|
2107
|
+
/**
|
|
2108
|
+
* The temperature of the model.
|
|
2109
|
+
*/
|
|
2110
|
+
temperature?: number;
|
|
2111
|
+
}
|
|
2112
|
+
|
|
2113
|
+
/**
|
|
2114
|
+
* AWS Bedrock model configuration.
|
|
2115
|
+
*/
|
|
2116
|
+
export interface BedrockModelSettings {
|
|
2117
|
+
/**
|
|
2118
|
+
* The maximum number of tokens the model can generate.
|
|
2119
|
+
*/
|
|
2120
|
+
max_output_tokens?: number;
|
|
2121
|
+
|
|
2122
|
+
/**
|
|
2123
|
+
* Whether to enable parallel tool calling.
|
|
2124
|
+
*/
|
|
2125
|
+
parallel_tool_calls?: boolean;
|
|
2126
|
+
|
|
2127
|
+
/**
|
|
2128
|
+
* The provider of the model.
|
|
2129
|
+
*/
|
|
2130
|
+
provider?: 'bedrock';
|
|
2131
|
+
|
|
2132
|
+
/**
|
|
2133
|
+
* The response format for the model.
|
|
2134
|
+
*/
|
|
2135
|
+
response_format?:
|
|
2136
|
+
| AgentsAPI.TextResponseFormat
|
|
2137
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2138
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2139
|
+
| null;
|
|
2140
|
+
|
|
2141
|
+
/**
|
|
2142
|
+
* The temperature of the model.
|
|
2143
|
+
*/
|
|
2144
|
+
temperature?: number;
|
|
2145
|
+
}
|
|
2146
|
+
}
|
|
2147
|
+
|
|
1298
2148
|
export interface AgentRetrieveParams {
|
|
1299
2149
|
/**
|
|
1300
2150
|
* Specify which relational fields to include in the response. No relationships are
|
|
@@ -1363,6 +2213,11 @@ export interface AgentListParams extends ArrayPageParams {
|
|
|
1363
2213
|
*/
|
|
1364
2214
|
include_relationships?: Array<string> | null;
|
|
1365
2215
|
|
|
2216
|
+
/**
|
|
2217
|
+
* Filter agents by their last stop reason.
|
|
2218
|
+
*/
|
|
2219
|
+
last_stop_reason?: RunsAPI.StopReasonType | null;
|
|
2220
|
+
|
|
1366
2221
|
/**
|
|
1367
2222
|
* If True, only returns agents that match ALL given tags. Otherwise, return agents
|
|
1368
2223
|
* that have ANY of the passed-in tags.
|
|
@@ -1488,8 +2343,7 @@ export interface AgentModifyParams {
|
|
|
1488
2343
|
description?: string | null;
|
|
1489
2344
|
|
|
1490
2345
|
/**
|
|
1491
|
-
* The embedding
|
|
1492
|
-
* provider/model-name.
|
|
2346
|
+
* The embedding model handle used by the agent (format: provider/model-name).
|
|
1493
2347
|
*/
|
|
1494
2348
|
embedding?: string | null;
|
|
1495
2349
|
|
|
@@ -1523,6 +2377,11 @@ export interface AgentModifyParams {
|
|
|
1523
2377
|
*/
|
|
1524
2378
|
last_run_duration_ms?: number | null;
|
|
1525
2379
|
|
|
2380
|
+
/**
|
|
2381
|
+
* The stop reason from the agent's last run.
|
|
2382
|
+
*/
|
|
2383
|
+
last_stop_reason?: RunsAPI.StopReasonType | null;
|
|
2384
|
+
|
|
1526
2385
|
/**
|
|
1527
2386
|
* Configuration for Language Model (LLM) connection and generation parameters.
|
|
1528
2387
|
*/
|
|
@@ -1535,8 +2394,8 @@ export interface AgentModifyParams {
|
|
|
1535
2394
|
max_files_open?: number | null;
|
|
1536
2395
|
|
|
1537
2396
|
/**
|
|
1538
|
-
*
|
|
1539
|
-
*
|
|
2397
|
+
* @deprecated Deprecated: Use `model` field to configure max output tokens
|
|
2398
|
+
* instead. The maximum number of tokens to generate, including reasoning step.
|
|
1540
2399
|
*/
|
|
1541
2400
|
max_tokens?: number | null;
|
|
1542
2401
|
|
|
@@ -1558,18 +2417,34 @@ export interface AgentModifyParams {
|
|
|
1558
2417
|
metadata?: { [key: string]: unknown } | null;
|
|
1559
2418
|
|
|
1560
2419
|
/**
|
|
1561
|
-
* The
|
|
1562
|
-
* provider/model-name, as an alternative to specifying llm_config.
|
|
2420
|
+
* The model handle used by the agent (format: provider/model-name).
|
|
1563
2421
|
*/
|
|
1564
2422
|
model?: string | null;
|
|
1565
2423
|
|
|
2424
|
+
/**
|
|
2425
|
+
* The model settings for the agent.
|
|
2426
|
+
*/
|
|
2427
|
+
model_settings?:
|
|
2428
|
+
| AgentModifyParams.OpenAIModelSettings
|
|
2429
|
+
| AgentModifyParams.AnthropicModelSettings
|
|
2430
|
+
| AgentModifyParams.GoogleAIModelSettings
|
|
2431
|
+
| AgentModifyParams.GoogleVertexModelSettings
|
|
2432
|
+
| AgentModifyParams.AzureModelSettings
|
|
2433
|
+
| AgentModifyParams.XaiModelSettings
|
|
2434
|
+
| AgentModifyParams.GroqModelSettings
|
|
2435
|
+
| AgentModifyParams.DeepseekModelSettings
|
|
2436
|
+
| AgentModifyParams.TogetherModelSettings
|
|
2437
|
+
| AgentModifyParams.BedrockModelSettings
|
|
2438
|
+
| null;
|
|
2439
|
+
|
|
1566
2440
|
/**
|
|
1567
2441
|
* The name of the agent.
|
|
1568
2442
|
*/
|
|
1569
2443
|
name?: string | null;
|
|
1570
2444
|
|
|
1571
2445
|
/**
|
|
1572
|
-
*
|
|
2446
|
+
* @deprecated Deprecated: Use `model` field to configure parallel tool calls
|
|
2447
|
+
* instead. If set to True, enables parallel tool calling.
|
|
1573
2448
|
*/
|
|
1574
2449
|
parallel_tool_calls?: boolean | null;
|
|
1575
2450
|
|
|
@@ -1585,11 +2460,13 @@ export interface AgentModifyParams {
|
|
|
1585
2460
|
project_id?: string | null;
|
|
1586
2461
|
|
|
1587
2462
|
/**
|
|
2463
|
+
* @deprecated Deprecated: Use `model` field to configure reasoning instead.
|
|
1588
2464
|
* Whether to enable reasoning for this agent.
|
|
1589
2465
|
*/
|
|
1590
2466
|
reasoning?: boolean | null;
|
|
1591
2467
|
|
|
1592
2468
|
/**
|
|
2469
|
+
* @deprecated Deprecated: Use `model` field to configure response format instead.
|
|
1593
2470
|
* The response format for the agent.
|
|
1594
2471
|
*/
|
|
1595
2472
|
response_format?: TextResponseFormat | JsonSchemaResponseFormat | JsonObjectResponseFormat | null;
|
|
@@ -1650,6 +2527,415 @@ export interface AgentModifyParams {
|
|
|
1650
2527
|
> | null;
|
|
1651
2528
|
}
|
|
1652
2529
|
|
|
2530
|
+
export namespace AgentModifyParams {
|
|
2531
|
+
export interface OpenAIModelSettings {
|
|
2532
|
+
/**
|
|
2533
|
+
* The maximum number of tokens the model can generate.
|
|
2534
|
+
*/
|
|
2535
|
+
max_output_tokens?: number;
|
|
2536
|
+
|
|
2537
|
+
/**
|
|
2538
|
+
* Whether to enable parallel tool calling.
|
|
2539
|
+
*/
|
|
2540
|
+
parallel_tool_calls?: boolean;
|
|
2541
|
+
|
|
2542
|
+
/**
|
|
2543
|
+
* The provider of the model.
|
|
2544
|
+
*/
|
|
2545
|
+
provider?: 'openai';
|
|
2546
|
+
|
|
2547
|
+
/**
|
|
2548
|
+
* The reasoning configuration for the model.
|
|
2549
|
+
*/
|
|
2550
|
+
reasoning?: OpenAIModelSettings.Reasoning;
|
|
2551
|
+
|
|
2552
|
+
/**
|
|
2553
|
+
* The response format for the model.
|
|
2554
|
+
*/
|
|
2555
|
+
response_format?:
|
|
2556
|
+
| AgentsAPI.TextResponseFormat
|
|
2557
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2558
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2559
|
+
| null;
|
|
2560
|
+
|
|
2561
|
+
/**
|
|
2562
|
+
* The temperature of the model.
|
|
2563
|
+
*/
|
|
2564
|
+
temperature?: number;
|
|
2565
|
+
}
|
|
2566
|
+
|
|
2567
|
+
export namespace OpenAIModelSettings {
|
|
2568
|
+
/**
|
|
2569
|
+
* The reasoning configuration for the model.
|
|
2570
|
+
*/
|
|
2571
|
+
export interface Reasoning {
|
|
2572
|
+
/**
|
|
2573
|
+
* The reasoning effort to use when generating text reasoning models
|
|
2574
|
+
*/
|
|
2575
|
+
reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
|
|
2576
|
+
}
|
|
2577
|
+
}
|
|
2578
|
+
|
|
2579
|
+
export interface AnthropicModelSettings {
|
|
2580
|
+
/**
|
|
2581
|
+
* The maximum number of tokens the model can generate.
|
|
2582
|
+
*/
|
|
2583
|
+
max_output_tokens?: number;
|
|
2584
|
+
|
|
2585
|
+
/**
|
|
2586
|
+
* Whether to enable parallel tool calling.
|
|
2587
|
+
*/
|
|
2588
|
+
parallel_tool_calls?: boolean;
|
|
2589
|
+
|
|
2590
|
+
/**
|
|
2591
|
+
* The provider of the model.
|
|
2592
|
+
*/
|
|
2593
|
+
provider?: 'anthropic';
|
|
2594
|
+
|
|
2595
|
+
/**
|
|
2596
|
+
* The temperature of the model.
|
|
2597
|
+
*/
|
|
2598
|
+
temperature?: number;
|
|
2599
|
+
|
|
2600
|
+
/**
|
|
2601
|
+
* The thinking configuration for the model.
|
|
2602
|
+
*/
|
|
2603
|
+
thinking?: AnthropicModelSettings.Thinking;
|
|
2604
|
+
|
|
2605
|
+
/**
|
|
2606
|
+
* Soft control for how verbose model output should be, used for GPT-5 models.
|
|
2607
|
+
*/
|
|
2608
|
+
verbosity?: 'low' | 'medium' | 'high' | null;
|
|
2609
|
+
}
|
|
2610
|
+
|
|
2611
|
+
export namespace AnthropicModelSettings {
|
|
2612
|
+
/**
|
|
2613
|
+
* The thinking configuration for the model.
|
|
2614
|
+
*/
|
|
2615
|
+
export interface Thinking {
|
|
2616
|
+
/**
|
|
2617
|
+
* The maximum number of tokens the model can use for extended thinking.
|
|
2618
|
+
*/
|
|
2619
|
+
budget_tokens?: number;
|
|
2620
|
+
|
|
2621
|
+
/**
|
|
2622
|
+
* The type of thinking to use.
|
|
2623
|
+
*/
|
|
2624
|
+
type?: 'enabled' | 'disabled';
|
|
2625
|
+
}
|
|
2626
|
+
}
|
|
2627
|
+
|
|
2628
|
+
export interface GoogleAIModelSettings {
|
|
2629
|
+
/**
|
|
2630
|
+
* The maximum number of tokens the model can generate.
|
|
2631
|
+
*/
|
|
2632
|
+
max_output_tokens?: number;
|
|
2633
|
+
|
|
2634
|
+
/**
|
|
2635
|
+
* Whether to enable parallel tool calling.
|
|
2636
|
+
*/
|
|
2637
|
+
parallel_tool_calls?: boolean;
|
|
2638
|
+
|
|
2639
|
+
/**
|
|
2640
|
+
* The provider of the model.
|
|
2641
|
+
*/
|
|
2642
|
+
provider?: 'google_ai';
|
|
2643
|
+
|
|
2644
|
+
/**
|
|
2645
|
+
* The response schema for the model.
|
|
2646
|
+
*/
|
|
2647
|
+
response_schema?:
|
|
2648
|
+
| AgentsAPI.TextResponseFormat
|
|
2649
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2650
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2651
|
+
| null;
|
|
2652
|
+
|
|
2653
|
+
/**
|
|
2654
|
+
* The temperature of the model.
|
|
2655
|
+
*/
|
|
2656
|
+
temperature?: number;
|
|
2657
|
+
|
|
2658
|
+
/**
|
|
2659
|
+
* The thinking configuration for the model.
|
|
2660
|
+
*/
|
|
2661
|
+
thinking_config?: GoogleAIModelSettings.ThinkingConfig;
|
|
2662
|
+
}
|
|
2663
|
+
|
|
2664
|
+
export namespace GoogleAIModelSettings {
|
|
2665
|
+
/**
|
|
2666
|
+
* The thinking configuration for the model.
|
|
2667
|
+
*/
|
|
2668
|
+
export interface ThinkingConfig {
|
|
2669
|
+
/**
|
|
2670
|
+
* Whether to include thoughts in the model's response.
|
|
2671
|
+
*/
|
|
2672
|
+
include_thoughts?: boolean;
|
|
2673
|
+
|
|
2674
|
+
/**
|
|
2675
|
+
* The thinking budget for the model.
|
|
2676
|
+
*/
|
|
2677
|
+
thinking_budget?: number;
|
|
2678
|
+
}
|
|
2679
|
+
}
|
|
2680
|
+
|
|
2681
|
+
export interface GoogleVertexModelSettings {
|
|
2682
|
+
/**
|
|
2683
|
+
* The maximum number of tokens the model can generate.
|
|
2684
|
+
*/
|
|
2685
|
+
max_output_tokens?: number;
|
|
2686
|
+
|
|
2687
|
+
/**
|
|
2688
|
+
* Whether to enable parallel tool calling.
|
|
2689
|
+
*/
|
|
2690
|
+
parallel_tool_calls?: boolean;
|
|
2691
|
+
|
|
2692
|
+
/**
|
|
2693
|
+
* The provider of the model.
|
|
2694
|
+
*/
|
|
2695
|
+
provider?: 'google_vertex';
|
|
2696
|
+
|
|
2697
|
+
/**
|
|
2698
|
+
* The response schema for the model.
|
|
2699
|
+
*/
|
|
2700
|
+
response_schema?:
|
|
2701
|
+
| AgentsAPI.TextResponseFormat
|
|
2702
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2703
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2704
|
+
| null;
|
|
2705
|
+
|
|
2706
|
+
/**
|
|
2707
|
+
* The temperature of the model.
|
|
2708
|
+
*/
|
|
2709
|
+
temperature?: number;
|
|
2710
|
+
|
|
2711
|
+
/**
|
|
2712
|
+
* The thinking configuration for the model.
|
|
2713
|
+
*/
|
|
2714
|
+
thinking_config?: GoogleVertexModelSettings.ThinkingConfig;
|
|
2715
|
+
}
|
|
2716
|
+
|
|
2717
|
+
export namespace GoogleVertexModelSettings {
|
|
2718
|
+
/**
|
|
2719
|
+
* The thinking configuration for the model.
|
|
2720
|
+
*/
|
|
2721
|
+
export interface ThinkingConfig {
|
|
2722
|
+
/**
|
|
2723
|
+
* Whether to include thoughts in the model's response.
|
|
2724
|
+
*/
|
|
2725
|
+
include_thoughts?: boolean;
|
|
2726
|
+
|
|
2727
|
+
/**
|
|
2728
|
+
* The thinking budget for the model.
|
|
2729
|
+
*/
|
|
2730
|
+
thinking_budget?: number;
|
|
2731
|
+
}
|
|
2732
|
+
}
|
|
2733
|
+
|
|
2734
|
+
/**
|
|
2735
|
+
* Azure OpenAI model configuration (OpenAI-compatible).
|
|
2736
|
+
*/
|
|
2737
|
+
export interface AzureModelSettings {
|
|
2738
|
+
/**
|
|
2739
|
+
* The maximum number of tokens the model can generate.
|
|
2740
|
+
*/
|
|
2741
|
+
max_output_tokens?: number;
|
|
2742
|
+
|
|
2743
|
+
/**
|
|
2744
|
+
* Whether to enable parallel tool calling.
|
|
2745
|
+
*/
|
|
2746
|
+
parallel_tool_calls?: boolean;
|
|
2747
|
+
|
|
2748
|
+
/**
|
|
2749
|
+
* The provider of the model.
|
|
2750
|
+
*/
|
|
2751
|
+
provider?: 'azure';
|
|
2752
|
+
|
|
2753
|
+
/**
|
|
2754
|
+
* The response format for the model.
|
|
2755
|
+
*/
|
|
2756
|
+
response_format?:
|
|
2757
|
+
| AgentsAPI.TextResponseFormat
|
|
2758
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2759
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2760
|
+
| null;
|
|
2761
|
+
|
|
2762
|
+
/**
|
|
2763
|
+
* The temperature of the model.
|
|
2764
|
+
*/
|
|
2765
|
+
temperature?: number;
|
|
2766
|
+
}
|
|
2767
|
+
|
|
2768
|
+
/**
|
|
2769
|
+
* xAI model configuration (OpenAI-compatible).
|
|
2770
|
+
*/
|
|
2771
|
+
export interface XaiModelSettings {
|
|
2772
|
+
/**
|
|
2773
|
+
* The maximum number of tokens the model can generate.
|
|
2774
|
+
*/
|
|
2775
|
+
max_output_tokens?: number;
|
|
2776
|
+
|
|
2777
|
+
/**
|
|
2778
|
+
* Whether to enable parallel tool calling.
|
|
2779
|
+
*/
|
|
2780
|
+
parallel_tool_calls?: boolean;
|
|
2781
|
+
|
|
2782
|
+
/**
|
|
2783
|
+
* The provider of the model.
|
|
2784
|
+
*/
|
|
2785
|
+
provider?: 'xai';
|
|
2786
|
+
|
|
2787
|
+
/**
|
|
2788
|
+
* The response format for the model.
|
|
2789
|
+
*/
|
|
2790
|
+
response_format?:
|
|
2791
|
+
| AgentsAPI.TextResponseFormat
|
|
2792
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2793
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2794
|
+
| null;
|
|
2795
|
+
|
|
2796
|
+
/**
|
|
2797
|
+
* The temperature of the model.
|
|
2798
|
+
*/
|
|
2799
|
+
temperature?: number;
|
|
2800
|
+
}
|
|
2801
|
+
|
|
2802
|
+
/**
|
|
2803
|
+
* Groq model configuration (OpenAI-compatible).
|
|
2804
|
+
*/
|
|
2805
|
+
export interface GroqModelSettings {
|
|
2806
|
+
/**
|
|
2807
|
+
* The maximum number of tokens the model can generate.
|
|
2808
|
+
*/
|
|
2809
|
+
max_output_tokens?: number;
|
|
2810
|
+
|
|
2811
|
+
/**
|
|
2812
|
+
* Whether to enable parallel tool calling.
|
|
2813
|
+
*/
|
|
2814
|
+
parallel_tool_calls?: boolean;
|
|
2815
|
+
|
|
2816
|
+
/**
|
|
2817
|
+
* The provider of the model.
|
|
2818
|
+
*/
|
|
2819
|
+
provider?: 'groq';
|
|
2820
|
+
|
|
2821
|
+
/**
|
|
2822
|
+
* The response format for the model.
|
|
2823
|
+
*/
|
|
2824
|
+
response_format?:
|
|
2825
|
+
| AgentsAPI.TextResponseFormat
|
|
2826
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2827
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2828
|
+
| null;
|
|
2829
|
+
|
|
2830
|
+
/**
|
|
2831
|
+
* The temperature of the model.
|
|
2832
|
+
*/
|
|
2833
|
+
temperature?: number;
|
|
2834
|
+
}
|
|
2835
|
+
|
|
2836
|
+
/**
|
|
2837
|
+
* Deepseek model configuration (OpenAI-compatible).
|
|
2838
|
+
*/
|
|
2839
|
+
export interface DeepseekModelSettings {
|
|
2840
|
+
/**
|
|
2841
|
+
* The maximum number of tokens the model can generate.
|
|
2842
|
+
*/
|
|
2843
|
+
max_output_tokens?: number;
|
|
2844
|
+
|
|
2845
|
+
/**
|
|
2846
|
+
* Whether to enable parallel tool calling.
|
|
2847
|
+
*/
|
|
2848
|
+
parallel_tool_calls?: boolean;
|
|
2849
|
+
|
|
2850
|
+
/**
|
|
2851
|
+
* The provider of the model.
|
|
2852
|
+
*/
|
|
2853
|
+
provider?: 'deepseek';
|
|
2854
|
+
|
|
2855
|
+
/**
|
|
2856
|
+
* The response format for the model.
|
|
2857
|
+
*/
|
|
2858
|
+
response_format?:
|
|
2859
|
+
| AgentsAPI.TextResponseFormat
|
|
2860
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2861
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2862
|
+
| null;
|
|
2863
|
+
|
|
2864
|
+
/**
|
|
2865
|
+
* The temperature of the model.
|
|
2866
|
+
*/
|
|
2867
|
+
temperature?: number;
|
|
2868
|
+
}
|
|
2869
|
+
|
|
2870
|
+
/**
|
|
2871
|
+
* Together AI model configuration (OpenAI-compatible).
|
|
2872
|
+
*/
|
|
2873
|
+
export interface TogetherModelSettings {
|
|
2874
|
+
/**
|
|
2875
|
+
* The maximum number of tokens the model can generate.
|
|
2876
|
+
*/
|
|
2877
|
+
max_output_tokens?: number;
|
|
2878
|
+
|
|
2879
|
+
/**
|
|
2880
|
+
* Whether to enable parallel tool calling.
|
|
2881
|
+
*/
|
|
2882
|
+
parallel_tool_calls?: boolean;
|
|
2883
|
+
|
|
2884
|
+
/**
|
|
2885
|
+
* The provider of the model.
|
|
2886
|
+
*/
|
|
2887
|
+
provider?: 'together';
|
|
2888
|
+
|
|
2889
|
+
/**
|
|
2890
|
+
* The response format for the model.
|
|
2891
|
+
*/
|
|
2892
|
+
response_format?:
|
|
2893
|
+
| AgentsAPI.TextResponseFormat
|
|
2894
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2895
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2896
|
+
| null;
|
|
2897
|
+
|
|
2898
|
+
/**
|
|
2899
|
+
* The temperature of the model.
|
|
2900
|
+
*/
|
|
2901
|
+
temperature?: number;
|
|
2902
|
+
}
|
|
2903
|
+
|
|
2904
|
+
/**
|
|
2905
|
+
* AWS Bedrock model configuration.
|
|
2906
|
+
*/
|
|
2907
|
+
export interface BedrockModelSettings {
|
|
2908
|
+
/**
|
|
2909
|
+
* The maximum number of tokens the model can generate.
|
|
2910
|
+
*/
|
|
2911
|
+
max_output_tokens?: number;
|
|
2912
|
+
|
|
2913
|
+
/**
|
|
2914
|
+
* Whether to enable parallel tool calling.
|
|
2915
|
+
*/
|
|
2916
|
+
parallel_tool_calls?: boolean;
|
|
2917
|
+
|
|
2918
|
+
/**
|
|
2919
|
+
* The provider of the model.
|
|
2920
|
+
*/
|
|
2921
|
+
provider?: 'bedrock';
|
|
2922
|
+
|
|
2923
|
+
/**
|
|
2924
|
+
* The response format for the model.
|
|
2925
|
+
*/
|
|
2926
|
+
response_format?:
|
|
2927
|
+
| AgentsAPI.TextResponseFormat
|
|
2928
|
+
| AgentsAPI.JsonSchemaResponseFormat
|
|
2929
|
+
| AgentsAPI.JsonObjectResponseFormat
|
|
2930
|
+
| null;
|
|
2931
|
+
|
|
2932
|
+
/**
|
|
2933
|
+
* The temperature of the model.
|
|
2934
|
+
*/
|
|
2935
|
+
temperature?: number;
|
|
2936
|
+
}
|
|
2937
|
+
}
|
|
2938
|
+
|
|
1653
2939
|
Agents.Tools = Tools;
|
|
1654
2940
|
Agents.Folders = Folders;
|
|
1655
2941
|
Agents.Files = Files;
|
|
@@ -1677,7 +2963,6 @@ export declare namespace Agents {
|
|
|
1677
2963
|
type TerminalToolRule as TerminalToolRule,
|
|
1678
2964
|
type TextResponseFormat as TextResponseFormat,
|
|
1679
2965
|
type AgentDeleteResponse as AgentDeleteResponse,
|
|
1680
|
-
type AgentCountResponse as AgentCountResponse,
|
|
1681
2966
|
type AgentExportFileResponse as AgentExportFileResponse,
|
|
1682
2967
|
type AgentImportFileResponse as AgentImportFileResponse,
|
|
1683
2968
|
type AgentStatesArrayPage as AgentStatesArrayPage,
|
|
@@ -1691,7 +2976,6 @@ export declare namespace Agents {
|
|
|
1691
2976
|
|
|
1692
2977
|
export {
|
|
1693
2978
|
Tools as Tools,
|
|
1694
|
-
type ToolListResponse as ToolListResponse,
|
|
1695
2979
|
type ToolListParams as ToolListParams,
|
|
1696
2980
|
type ToolAttachParams as ToolAttachParams,
|
|
1697
2981
|
type ToolDetachParams as ToolDetachParams,
|
|
@@ -1701,6 +2985,7 @@ export declare namespace Agents {
|
|
|
1701
2985
|
export {
|
|
1702
2986
|
Folders as Folders,
|
|
1703
2987
|
type FolderListResponse as FolderListResponse,
|
|
2988
|
+
type FolderListResponsesArrayPage as FolderListResponsesArrayPage,
|
|
1704
2989
|
type FolderListParams as FolderListParams,
|
|
1705
2990
|
type FolderAttachParams as FolderAttachParams,
|
|
1706
2991
|
type FolderDetachParams as FolderDetachParams,
|
|
@@ -1779,6 +3064,8 @@ export declare namespace Agents {
|
|
|
1779
3064
|
type MessageModifyParams as MessageModifyParams,
|
|
1780
3065
|
type MessageResetParams as MessageResetParams,
|
|
1781
3066
|
type MessageSendParams as MessageSendParams,
|
|
3067
|
+
type MessageSendParamsNonStreaming as MessageSendParamsNonStreaming,
|
|
3068
|
+
type MessageSendParamsStreaming as MessageSendParamsStreaming,
|
|
1782
3069
|
type MessageSendAsyncParams as MessageSendAsyncParams,
|
|
1783
3070
|
type MessageStreamParams as MessageStreamParams,
|
|
1784
3071
|
};
|