ai 4.0.0-canary.5 → 4.0.0-canary.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +24 -0
- package/dist/index.d.mts +50 -168
- package/dist/index.d.ts +50 -168
- package/dist/index.js +55 -84
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +55 -81
- package/dist/index.mjs.map +1 -1
- package/package.json +1 -7
- package/rsc/dist/index.d.ts +18 -18
- package/rsc/dist/rsc-server.d.mts +18 -18
- package/rsc/dist/rsc-server.mjs +7 -4
- package/rsc/dist/rsc-server.mjs.map +1 -1
package/dist/index.d.ts
CHANGED
@@ -46,33 +46,6 @@ type TelemetrySettings = {
|
|
46
46
|
tracer?: Tracer;
|
47
47
|
};
|
48
48
|
|
49
|
-
/**
|
50
|
-
Represents the number of tokens used in a prompt and completion.
|
51
|
-
*/
|
52
|
-
type LanguageModelUsage$1 = {
|
53
|
-
/**
|
54
|
-
The number of tokens used in the prompt.
|
55
|
-
*/
|
56
|
-
promptTokens: number;
|
57
|
-
/**
|
58
|
-
The number of tokens used in the completion.
|
59
|
-
*/
|
60
|
-
completionTokens: number;
|
61
|
-
/**
|
62
|
-
The total number of tokens used (promptTokens + completionTokens).
|
63
|
-
*/
|
64
|
-
totalTokens: number;
|
65
|
-
};
|
66
|
-
/**
|
67
|
-
Represents the number of tokens used in an embedding.
|
68
|
-
*/
|
69
|
-
type EmbeddingModelUsage$1 = {
|
70
|
-
/**
|
71
|
-
The number of tokens used in the embedding.
|
72
|
-
*/
|
73
|
-
tokens: number;
|
74
|
-
};
|
75
|
-
|
76
49
|
/**
|
77
50
|
Embedding model that is used by the AI SDK Core functions.
|
78
51
|
*/
|
@@ -147,10 +120,6 @@ type LanguageModelResponseMetadata = {
|
|
147
120
|
*/
|
148
121
|
headers?: Record<string, string>;
|
149
122
|
};
|
150
|
-
/**
|
151
|
-
@deprecated Use `LanguageModelResponseMetadata` instead.
|
152
|
-
*/
|
153
|
-
type LanguageModelResponseMetadataWithHeaders = LanguageModelResponseMetadata;
|
154
123
|
|
155
124
|
/**
|
156
125
|
* Provider for language and text embedding models.
|
@@ -188,19 +157,31 @@ functionality that can be fully encapsulated in the provider.
|
|
188
157
|
type ProviderMetadata = LanguageModelV1ProviderMetadata;
|
189
158
|
|
190
159
|
/**
|
191
|
-
|
192
|
-
*/
|
193
|
-
type TokenUsage = LanguageModelUsage$1;
|
194
|
-
/**
|
195
|
-
* @deprecated Use LanguageModelUsage instead.
|
160
|
+
Represents the number of tokens used in a prompt and completion.
|
196
161
|
*/
|
197
|
-
type
|
198
|
-
|
162
|
+
type LanguageModelUsage = {
|
163
|
+
/**
|
164
|
+
The number of tokens used in the prompt.
|
165
|
+
*/
|
166
|
+
promptTokens: number;
|
167
|
+
/**
|
168
|
+
The number of tokens used in the completion.
|
169
|
+
*/
|
170
|
+
completionTokens: number;
|
171
|
+
/**
|
172
|
+
The total number of tokens used (promptTokens + completionTokens).
|
173
|
+
*/
|
174
|
+
totalTokens: number;
|
175
|
+
};
|
199
176
|
/**
|
200
|
-
|
177
|
+
Represents the number of tokens used in an embedding.
|
201
178
|
*/
|
202
|
-
type
|
203
|
-
|
179
|
+
type EmbeddingModelUsage = {
|
180
|
+
/**
|
181
|
+
The number of tokens used in the embedding.
|
182
|
+
*/
|
183
|
+
tokens: number;
|
184
|
+
};
|
204
185
|
|
205
186
|
/**
|
206
187
|
The result of an `embed` call.
|
@@ -218,7 +199,7 @@ interface EmbedResult<VALUE> {
|
|
218
199
|
/**
|
219
200
|
The embedding token usage.
|
220
201
|
*/
|
221
|
-
readonly usage: EmbeddingModelUsage
|
202
|
+
readonly usage: EmbeddingModelUsage;
|
222
203
|
/**
|
223
204
|
Optional raw response data.
|
224
205
|
*/
|
@@ -288,7 +269,7 @@ interface EmbedManyResult<VALUE> {
|
|
288
269
|
/**
|
289
270
|
The embedding token usage.
|
290
271
|
*/
|
291
|
-
readonly usage: EmbeddingModelUsage
|
272
|
+
readonly usage: EmbeddingModelUsage;
|
292
273
|
}
|
293
274
|
|
294
275
|
/**
|
@@ -659,7 +640,7 @@ interface GenerateObjectResult<T> {
|
|
659
640
|
/**
|
660
641
|
The token usage of the generated text.
|
661
642
|
*/
|
662
|
-
readonly usage: LanguageModelUsage
|
643
|
+
readonly usage: LanguageModelUsage;
|
663
644
|
/**
|
664
645
|
Warnings from the model provider (e.g. unsupported settings).
|
665
646
|
*/
|
@@ -916,7 +897,7 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
|
|
916
897
|
/**
|
917
898
|
The token usage of the generated response. Resolved when the response is finished.
|
918
899
|
*/
|
919
|
-
readonly usage: Promise<LanguageModelUsage
|
900
|
+
readonly usage: Promise<LanguageModelUsage>;
|
920
901
|
/**
|
921
902
|
Additional provider-specific metadata. They are passed through
|
922
903
|
from the provider to the AI SDK and enable provider-specific
|
@@ -988,7 +969,7 @@ type ObjectStreamPart<PARTIAL> = {
|
|
988
969
|
type: 'finish';
|
989
970
|
finishReason: FinishReason;
|
990
971
|
logprobs?: LogProbs;
|
991
|
-
usage: LanguageModelUsage
|
972
|
+
usage: LanguageModelUsage;
|
992
973
|
response: LanguageModelResponseMetadata;
|
993
974
|
providerMetadata?: ProviderMetadata;
|
994
975
|
};
|
@@ -997,7 +978,7 @@ type OnFinishCallback<RESULT> = (event: {
|
|
997
978
|
/**
|
998
979
|
The token usage of the generated response.
|
999
980
|
*/
|
1000
|
-
usage: LanguageModelUsage
|
981
|
+
usage: LanguageModelUsage;
|
1001
982
|
/**
|
1002
983
|
The generated object. Can be undefined if the final object does not match the schema.
|
1003
984
|
*/
|
@@ -1266,10 +1247,6 @@ declare function tool<PARAMETERS extends Parameters, RESULT>(tool: CoreTool<PARA
|
|
1266
1247
|
}): CoreTool<PARAMETERS, RESULT> & {
|
1267
1248
|
execute: undefined;
|
1268
1249
|
};
|
1269
|
-
/**
|
1270
|
-
* @deprecated Use `CoreTool` instead.
|
1271
|
-
*/
|
1272
|
-
type ExperimentalTool = CoreTool;
|
1273
1250
|
|
1274
1251
|
/**
|
1275
1252
|
Converts an array of messages from useChat into an array of CoreMessages that can be used
|
@@ -1374,7 +1351,7 @@ type StepResult<TOOLS extends Record<string, CoreTool>> = {
|
|
1374
1351
|
/**
|
1375
1352
|
The token usage of the generated text.
|
1376
1353
|
*/
|
1377
|
-
readonly usage: LanguageModelUsage
|
1354
|
+
readonly usage: LanguageModelUsage;
|
1378
1355
|
/**
|
1379
1356
|
Warnings from the model provider (e.g. unsupported settings).
|
1380
1357
|
*/
|
@@ -1440,16 +1417,12 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1440
1417
|
/**
|
1441
1418
|
The token usage of the generated text.
|
1442
1419
|
*/
|
1443
|
-
readonly usage: LanguageModelUsage
|
1420
|
+
readonly usage: LanguageModelUsage;
|
1444
1421
|
/**
|
1445
1422
|
Warnings from the model provider (e.g. unsupported settings)
|
1446
1423
|
*/
|
1447
1424
|
readonly warnings: CallWarning[] | undefined;
|
1448
1425
|
/**
|
1449
|
-
@deprecated use `response.messages` instead.
|
1450
|
-
*/
|
1451
|
-
readonly responseMessages: Array<CoreAssistantMessage | CoreToolMessage>;
|
1452
|
-
/**
|
1453
1426
|
Details for all steps.
|
1454
1427
|
You can use this to get information about intermediate steps,
|
1455
1428
|
such as the tool calls or the response headers.
|
@@ -1534,7 +1507,7 @@ If set and supported by the model, calls will generate deterministic results.
|
|
1534
1507
|
@returns
|
1535
1508
|
A result object that contains the generated text, the results of the tool calls, and additional information.
|
1536
1509
|
*/
|
1537
|
-
declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxSteps,
|
1510
|
+
declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxSteps, experimental_continueSteps: continueSteps, experimental_telemetry: telemetry, experimental_providerMetadata: providerMetadata, experimental_activeTools: activeTools, _internal: { generateId, currentDate, }, onStepFinish, ...settings }: CallSettings & Prompt & {
|
1538
1511
|
/**
|
1539
1512
|
The language model to use.
|
1540
1513
|
*/
|
@@ -1556,10 +1529,6 @@ By default, it's set to 1, which means that only a single LLM call is made.
|
|
1556
1529
|
*/
|
1557
1530
|
maxSteps?: number;
|
1558
1531
|
/**
|
1559
|
-
@deprecated Use `experimental_continueSteps` instead.
|
1560
|
-
*/
|
1561
|
-
experimental_continuationSteps?: boolean;
|
1562
|
-
/**
|
1563
1532
|
When enabled, the model will perform additional steps if the finish reason is "length" (experimental).
|
1564
1533
|
|
1565
1534
|
By default, it's set to false.
|
@@ -1612,11 +1581,6 @@ declare class StreamData {
|
|
1612
1581
|
* This assumes every chunk is a 'text' chunk.
|
1613
1582
|
*/
|
1614
1583
|
declare function createStreamDataTransformer(): TransformStream<any, any>;
|
1615
|
-
/**
|
1616
|
-
@deprecated Use `StreamData` instead.
|
1617
|
-
*/
|
1618
|
-
declare class experimental_StreamData extends StreamData {
|
1619
|
-
}
|
1620
1584
|
|
1621
1585
|
/**
|
1622
1586
|
A result object for accessing different stream types and additional information.
|
@@ -1632,7 +1596,7 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1632
1596
|
|
1633
1597
|
Resolved when the response is finished.
|
1634
1598
|
*/
|
1635
|
-
readonly usage: Promise<LanguageModelUsage
|
1599
|
+
readonly usage: Promise<LanguageModelUsage>;
|
1636
1600
|
/**
|
1637
1601
|
The reason why the generation finished. Taken from the last step.
|
1638
1602
|
|
@@ -1664,10 +1628,6 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1664
1628
|
*/
|
1665
1629
|
readonly toolResults: Promise<ToolResultUnion<TOOLS>[]>;
|
1666
1630
|
/**
|
1667
|
-
@deprecated use `response.messages` instead.
|
1668
|
-
*/
|
1669
|
-
readonly responseMessages: Promise<Array<CoreAssistantMessage | CoreToolMessage>>;
|
1670
|
-
/**
|
1671
1631
|
Details for all steps.
|
1672
1632
|
You can use this to get information about intermediate steps,
|
1673
1633
|
such as the tool calls or the response headers.
|
@@ -1722,11 +1682,14 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1722
1682
|
Writes data stream output to a Node.js response-like object.
|
1723
1683
|
|
1724
1684
|
@param response A Node.js response-like object (ServerResponse).
|
1725
|
-
@param options
|
1726
|
-
|
1685
|
+
@param options.status The status code.
|
1686
|
+
@param options.statusText The status text.
|
1687
|
+
@param options.headers The headers.
|
1688
|
+
@param options.data The stream data.
|
1689
|
+
@param options.getErrorMessage An optional function that converts an error to an error message.
|
1690
|
+
@param options.sendUsage Whether to send the usage information to the client. Defaults to true.
|
1727
1691
|
*/
|
1728
|
-
pipeDataStreamToResponse(response: ServerResponse$1, options?: ResponseInit
|
1729
|
-
init?: ResponseInit;
|
1692
|
+
pipeDataStreamToResponse(response: ServerResponse$1, options?: ResponseInit & {
|
1730
1693
|
data?: StreamData;
|
1731
1694
|
getErrorMessage?: (error: unknown) => string;
|
1732
1695
|
sendUsage?: boolean;
|
@@ -1744,13 +1707,16 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
|
|
1744
1707
|
Converts the result to a streamed response object with a stream data part stream.
|
1745
1708
|
It can be used with the `useChat` and `useCompletion` hooks.
|
1746
1709
|
|
1747
|
-
@param options
|
1748
|
-
|
1710
|
+
@param options.status The status code.
|
1711
|
+
@param options.statusText The status text.
|
1712
|
+
@param options.headers The headers.
|
1713
|
+
@param options.data The stream data.
|
1714
|
+
@param options.getErrorMessage An optional function that converts an error to an error message.
|
1715
|
+
@param options.sendUsage Whether to send the usage information to the client. Defaults to true.
|
1749
1716
|
|
1750
1717
|
@return A response object.
|
1751
1718
|
*/
|
1752
|
-
toDataStreamResponse(options?: ResponseInit
|
1753
|
-
init?: ResponseInit;
|
1719
|
+
toDataStreamResponse(options?: ResponseInit & {
|
1754
1720
|
data?: StreamData;
|
1755
1721
|
getErrorMessage?: (error: unknown) => string;
|
1756
1722
|
sendUsage?: boolean;
|
@@ -1784,7 +1750,7 @@ type TextStreamPart<TOOLS extends Record<string, CoreTool>> = {
|
|
1784
1750
|
type: 'step-finish';
|
1785
1751
|
finishReason: FinishReason;
|
1786
1752
|
logprobs?: LogProbs;
|
1787
|
-
usage: LanguageModelUsage
|
1753
|
+
usage: LanguageModelUsage;
|
1788
1754
|
response: LanguageModelResponseMetadata;
|
1789
1755
|
experimental_providerMetadata?: ProviderMetadata;
|
1790
1756
|
isContinued: boolean;
|
@@ -1792,7 +1758,7 @@ type TextStreamPart<TOOLS extends Record<string, CoreTool>> = {
|
|
1792
1758
|
type: 'finish';
|
1793
1759
|
finishReason: FinishReason;
|
1794
1760
|
logprobs?: LogProbs;
|
1795
|
-
usage: LanguageModelUsage
|
1761
|
+
usage: LanguageModelUsage;
|
1796
1762
|
response: LanguageModelResponseMetadata;
|
1797
1763
|
experimental_providerMetadata?: ProviderMetadata;
|
1798
1764
|
} | {
|
@@ -2034,70 +2000,10 @@ declare class NoSuchProviderError extends NoSuchModelError {
|
|
2034
2000
|
static isInstance(error: unknown): error is NoSuchProviderError;
|
2035
2001
|
}
|
2036
2002
|
|
2037
|
-
/**
|
2038
|
-
* Provides for language and text embedding models.
|
2039
|
-
*
|
2040
|
-
* @deprecated Use `ProviderV1` instead.
|
2041
|
-
*/
|
2042
|
-
interface experimental_Provider {
|
2043
|
-
/**
|
2044
|
-
Returns the language model with the given id in the format `providerId:modelId`.
|
2045
|
-
The model id is then passed to the provider function to get the model.
|
2046
|
-
|
2047
|
-
@param {string} id - The id of the model to return.
|
2048
|
-
|
2049
|
-
@throws {NoSuchModelError} If no model with the given id exists.
|
2050
|
-
@throws {NoSuchProviderError} If no provider with the given id exists.
|
2051
|
-
|
2052
|
-
@returns {LanguageModel} The language model associated with the id.
|
2053
|
-
*/
|
2054
|
-
languageModel?: (modelId: string) => LanguageModel;
|
2055
|
-
/**
|
2056
|
-
Returns the text embedding model with the given id in the format `providerId:modelId`.
|
2057
|
-
The model id is then passed to the provider function to get the model.
|
2058
|
-
|
2059
|
-
@param {string} id - The id of the model to return.
|
2060
|
-
|
2061
|
-
@throws {NoSuchModelError} If no model with the given id exists.
|
2062
|
-
@throws {NoSuchProviderError} If no provider with the given id exists.
|
2063
|
-
|
2064
|
-
@returns {LanguageModel} The language model associated with the id.
|
2065
|
-
*/
|
2066
|
-
textEmbeddingModel?: (modelId: string) => EmbeddingModel<string>;
|
2067
|
-
/**
|
2068
|
-
Returns the text embedding model with the given id in the format `providerId:modelId`.
|
2069
|
-
The model id is then passed to the provider function to get the model.
|
2070
|
-
|
2071
|
-
@param {string} id - The id of the model to return.
|
2072
|
-
|
2073
|
-
@throws {NoSuchModelError} If no model with the given id exists.
|
2074
|
-
@throws {NoSuchProviderError} If no provider with the given id exists.
|
2075
|
-
|
2076
|
-
@returns {LanguageModel} The language model associated with the id.
|
2077
|
-
|
2078
|
-
@deprecated use `textEmbeddingModel` instead.
|
2079
|
-
*/
|
2080
|
-
textEmbedding?: (modelId: string) => EmbeddingModel<string>;
|
2081
|
-
}
|
2082
|
-
|
2083
|
-
/**
|
2084
|
-
Registry for managing models. It enables getting a model with a string id.
|
2085
|
-
|
2086
|
-
@deprecated Use `experimental_Provider` instead.
|
2087
|
-
*/
|
2088
|
-
type experimental_ProviderRegistry = Provider;
|
2089
|
-
/**
|
2090
|
-
* @deprecated Use `experimental_ProviderRegistry` instead.
|
2091
|
-
*/
|
2092
|
-
type experimental_ModelRegistry = experimental_ProviderRegistry;
|
2093
2003
|
/**
|
2094
2004
|
* Creates a registry for the given providers.
|
2095
2005
|
*/
|
2096
|
-
declare function experimental_createProviderRegistry(providers: Record<string,
|
2097
|
-
/**
|
2098
|
-
* @deprecated Use `experimental_createProviderRegistry` instead.
|
2099
|
-
*/
|
2100
|
-
declare const experimental_createModelRegistry: typeof experimental_createProviderRegistry;
|
2006
|
+
declare function experimental_createProviderRegistry(providers: Record<string, Provider>): Provider;
|
2101
2007
|
|
2102
2008
|
/**
|
2103
2009
|
* Calculates the cosine similarity between two vectors. This is a useful metric for
|
@@ -2246,14 +2152,6 @@ The process parameter is a callback in which you can run the assistant on thread
|
|
2246
2152
|
*/
|
2247
2153
|
type AssistantResponseCallback = (options: {
|
2248
2154
|
/**
|
2249
|
-
@deprecated use variable from outer scope instead.
|
2250
|
-
*/
|
2251
|
-
threadId: string;
|
2252
|
-
/**
|
2253
|
-
@deprecated use variable from outer scope instead.
|
2254
|
-
*/
|
2255
|
-
messageId: string;
|
2256
|
-
/**
|
2257
2155
|
Forwards an assistant message (non-streaming) to the client.
|
2258
2156
|
*/
|
2259
2157
|
sendMessage: (message: AssistantMessage) => void;
|
@@ -2272,10 +2170,6 @@ It is designed to facilitate streaming assistant responses to the `useAssistant`
|
|
2272
2170
|
It receives an assistant thread and a current message, and can send messages and data messages to the client.
|
2273
2171
|
*/
|
2274
2172
|
declare function AssistantResponse({ threadId, messageId }: AssistantResponseSettings, process: AssistantResponseCallback): Response;
|
2275
|
-
/**
|
2276
|
-
@deprecated Use `AssistantResponse` instead.
|
2277
|
-
*/
|
2278
|
-
declare const experimental_AssistantResponse: typeof AssistantResponse;
|
2279
2173
|
|
2280
2174
|
/**
|
2281
2175
|
* Configuration options and helper callback methods for stream lifecycle events.
|
@@ -2321,16 +2215,6 @@ type LangChainStreamEvent = {
|
|
2321
2215
|
/**
|
2322
2216
|
Converts LangChain output streams to AIStream.
|
2323
2217
|
|
2324
|
-
The following streams are supported:
|
2325
|
-
- `LangChainAIMessageChunk` streams (LangChain `model.stream` output)
|
2326
|
-
- `string` streams (LangChain `StringOutputParser` output)
|
2327
|
-
|
2328
|
-
@deprecated Use `toDataStream` instead.
|
2329
|
-
*/
|
2330
|
-
declare function toAIStream(stream: ReadableStream<LangChainStreamEvent> | ReadableStream<LangChainAIMessageChunk> | ReadableStream<string>, callbacks?: StreamCallbacks): ReadableStream<any>;
|
2331
|
-
/**
|
2332
|
-
Converts LangChain output streams to AIStream.
|
2333
|
-
|
2334
2218
|
The following streams are supported:
|
2335
2219
|
- `LangChainAIMessageChunk` streams (LangChain `model.stream` output)
|
2336
2220
|
- `string` streams (LangChain `StringOutputParser` output)
|
@@ -2342,10 +2226,8 @@ declare function toDataStreamResponse$1(stream: ReadableStream<LangChainStreamEv
|
|
2342
2226
|
callbacks?: StreamCallbacks;
|
2343
2227
|
}): Response;
|
2344
2228
|
|
2345
|
-
declare const langchainAdapter_toAIStream: typeof toAIStream;
|
2346
2229
|
declare namespace langchainAdapter {
|
2347
2230
|
export {
|
2348
|
-
langchainAdapter_toAIStream as toAIStream,
|
2349
2231
|
toDataStream$1 as toDataStream,
|
2350
2232
|
toDataStreamResponse$1 as toDataStreamResponse,
|
2351
2233
|
};
|
@@ -2370,4 +2252,4 @@ declare namespace llamaindexAdapter {
|
|
2370
2252
|
};
|
2371
2253
|
}
|
2372
2254
|
|
2373
|
-
export { AssistantContent, AssistantResponse, CallWarning,
|
2255
|
+
export { AssistantContent, AssistantResponse, CallWarning, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, ToolCallUnion as CoreToolCallUnion, CoreToolChoice, CoreToolMessage, ToolResultUnion as CoreToolResultUnion, CoreUserMessage, DataContent, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, Experimental_LanguageModelV1Middleware, FilePart, FinishReason, GenerateObjectResult, GenerateTextResult, ImagePart, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidToolArgumentsError, langchainAdapter as LangChainAdapter, LanguageModel, LanguageModelRequestMetadata, LanguageModelResponseMetadata, LanguageModelUsage, llamaindexAdapter as LlamaIndexAdapter, LogProbs, MessageConversionError, NoObjectGeneratedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, Provider, ProviderMetadata, RetryError, StepResult, StreamData, StreamObjectResult, StreamTextResult, TextPart, TextStreamPart, ToolCallPart, ToolContent, ToolResultPart, UserContent, convertToCoreMessages, cosineSimilarity, createStreamDataTransformer, embed, embedMany, experimental_createProviderRegistry, experimental_customProvider, experimental_wrapLanguageModel, generateObject, generateText, streamObject, streamText, tool };
|