ai 4.0.0-canary.5 → 4.0.0-canary.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.ts CHANGED
@@ -46,33 +46,6 @@ type TelemetrySettings = {
46
46
  tracer?: Tracer;
47
47
  };
48
48
 
49
- /**
50
- Represents the number of tokens used in a prompt and completion.
51
- */
52
- type LanguageModelUsage$1 = {
53
- /**
54
- The number of tokens used in the prompt.
55
- */
56
- promptTokens: number;
57
- /**
58
- The number of tokens used in the completion.
59
- */
60
- completionTokens: number;
61
- /**
62
- The total number of tokens used (promptTokens + completionTokens).
63
- */
64
- totalTokens: number;
65
- };
66
- /**
67
- Represents the number of tokens used in an embedding.
68
- */
69
- type EmbeddingModelUsage$1 = {
70
- /**
71
- The number of tokens used in the embedding.
72
- */
73
- tokens: number;
74
- };
75
-
76
49
  /**
77
50
  Embedding model that is used by the AI SDK Core functions.
78
51
  */
@@ -147,10 +120,6 @@ type LanguageModelResponseMetadata = {
147
120
  */
148
121
  headers?: Record<string, string>;
149
122
  };
150
- /**
151
- @deprecated Use `LanguageModelResponseMetadata` instead.
152
- */
153
- type LanguageModelResponseMetadataWithHeaders = LanguageModelResponseMetadata;
154
123
 
155
124
  /**
156
125
  * Provider for language and text embedding models.
@@ -188,19 +157,31 @@ functionality that can be fully encapsulated in the provider.
188
157
  type ProviderMetadata = LanguageModelV1ProviderMetadata;
189
158
 
190
159
  /**
191
- * @deprecated Use LanguageModelUsage instead.
192
- */
193
- type TokenUsage = LanguageModelUsage$1;
194
- /**
195
- * @deprecated Use LanguageModelUsage instead.
160
+ Represents the number of tokens used in a prompt and completion.
196
161
  */
197
- type CompletionTokenUsage = LanguageModelUsage$1;
198
- type LanguageModelUsage = LanguageModelUsage$1;
162
+ type LanguageModelUsage = {
163
+ /**
164
+ The number of tokens used in the prompt.
165
+ */
166
+ promptTokens: number;
167
+ /**
168
+ The number of tokens used in the completion.
169
+ */
170
+ completionTokens: number;
171
+ /**
172
+ The total number of tokens used (promptTokens + completionTokens).
173
+ */
174
+ totalTokens: number;
175
+ };
199
176
  /**
200
- * @deprecated Use EmbeddingModelUsage instead.
177
+ Represents the number of tokens used in an embedding.
201
178
  */
202
- type EmbeddingTokenUsage = EmbeddingModelUsage$1;
203
- type EmbeddingModelUsage = EmbeddingModelUsage$1;
179
+ type EmbeddingModelUsage = {
180
+ /**
181
+ The number of tokens used in the embedding.
182
+ */
183
+ tokens: number;
184
+ };
204
185
 
205
186
  /**
206
187
  The result of an `embed` call.
@@ -218,7 +199,7 @@ interface EmbedResult<VALUE> {
218
199
  /**
219
200
  The embedding token usage.
220
201
  */
221
- readonly usage: EmbeddingModelUsage$1;
202
+ readonly usage: EmbeddingModelUsage;
222
203
  /**
223
204
  Optional raw response data.
224
205
  */
@@ -288,7 +269,7 @@ interface EmbedManyResult<VALUE> {
288
269
  /**
289
270
  The embedding token usage.
290
271
  */
291
- readonly usage: EmbeddingModelUsage$1;
272
+ readonly usage: EmbeddingModelUsage;
292
273
  }
293
274
 
294
275
  /**
@@ -659,7 +640,7 @@ interface GenerateObjectResult<T> {
659
640
  /**
660
641
  The token usage of the generated text.
661
642
  */
662
- readonly usage: LanguageModelUsage$1;
643
+ readonly usage: LanguageModelUsage;
663
644
  /**
664
645
  Warnings from the model provider (e.g. unsupported settings).
665
646
  */
@@ -916,7 +897,7 @@ interface StreamObjectResult<PARTIAL, RESULT, ELEMENT_STREAM> {
916
897
  /**
917
898
  The token usage of the generated response. Resolved when the response is finished.
918
899
  */
919
- readonly usage: Promise<LanguageModelUsage$1>;
900
+ readonly usage: Promise<LanguageModelUsage>;
920
901
  /**
921
902
  Additional provider-specific metadata. They are passed through
922
903
  from the provider to the AI SDK and enable provider-specific
@@ -988,7 +969,7 @@ type ObjectStreamPart<PARTIAL> = {
988
969
  type: 'finish';
989
970
  finishReason: FinishReason;
990
971
  logprobs?: LogProbs;
991
- usage: LanguageModelUsage$1;
972
+ usage: LanguageModelUsage;
992
973
  response: LanguageModelResponseMetadata;
993
974
  providerMetadata?: ProviderMetadata;
994
975
  };
@@ -997,7 +978,7 @@ type OnFinishCallback<RESULT> = (event: {
997
978
  /**
998
979
  The token usage of the generated response.
999
980
  */
1000
- usage: LanguageModelUsage$1;
981
+ usage: LanguageModelUsage;
1001
982
  /**
1002
983
  The generated object. Can be undefined if the final object does not match the schema.
1003
984
  */
@@ -1266,10 +1247,6 @@ declare function tool<PARAMETERS extends Parameters, RESULT>(tool: CoreTool<PARA
1266
1247
  }): CoreTool<PARAMETERS, RESULT> & {
1267
1248
  execute: undefined;
1268
1249
  };
1269
- /**
1270
- * @deprecated Use `CoreTool` instead.
1271
- */
1272
- type ExperimentalTool = CoreTool;
1273
1250
 
1274
1251
  /**
1275
1252
  Converts an array of messages from useChat into an array of CoreMessages that can be used
@@ -1374,7 +1351,7 @@ type StepResult<TOOLS extends Record<string, CoreTool>> = {
1374
1351
  /**
1375
1352
  The token usage of the generated text.
1376
1353
  */
1377
- readonly usage: LanguageModelUsage$1;
1354
+ readonly usage: LanguageModelUsage;
1378
1355
  /**
1379
1356
  Warnings from the model provider (e.g. unsupported settings).
1380
1357
  */
@@ -1440,16 +1417,12 @@ interface GenerateTextResult<TOOLS extends Record<string, CoreTool>> {
1440
1417
  /**
1441
1418
  The token usage of the generated text.
1442
1419
  */
1443
- readonly usage: LanguageModelUsage$1;
1420
+ readonly usage: LanguageModelUsage;
1444
1421
  /**
1445
1422
  Warnings from the model provider (e.g. unsupported settings)
1446
1423
  */
1447
1424
  readonly warnings: CallWarning[] | undefined;
1448
1425
  /**
1449
- @deprecated use `response.messages` instead.
1450
- */
1451
- readonly responseMessages: Array<CoreAssistantMessage | CoreToolMessage>;
1452
- /**
1453
1426
  Details for all steps.
1454
1427
  You can use this to get information about intermediate steps,
1455
1428
  such as the tool calls or the response headers.
@@ -1534,7 +1507,7 @@ If set and supported by the model, calls will generate deterministic results.
1534
1507
  @returns
1535
1508
  A result object that contains the generated text, the results of the tool calls, and additional information.
1536
1509
  */
1537
- declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxSteps, experimental_continuationSteps, experimental_continueSteps: continueSteps, experimental_telemetry: telemetry, experimental_providerMetadata: providerMetadata, experimental_activeTools: activeTools, _internal: { generateId, currentDate, }, onStepFinish, ...settings }: CallSettings & Prompt & {
1510
+ declare function generateText<TOOLS extends Record<string, CoreTool>>({ model, tools, toolChoice, system, prompt, messages, maxRetries, abortSignal, headers, maxSteps, experimental_continueSteps: continueSteps, experimental_telemetry: telemetry, experimental_providerMetadata: providerMetadata, experimental_activeTools: activeTools, _internal: { generateId, currentDate, }, onStepFinish, ...settings }: CallSettings & Prompt & {
1538
1511
  /**
1539
1512
  The language model to use.
1540
1513
  */
@@ -1556,10 +1529,6 @@ By default, it's set to 1, which means that only a single LLM call is made.
1556
1529
  */
1557
1530
  maxSteps?: number;
1558
1531
  /**
1559
- @deprecated Use `experimental_continueSteps` instead.
1560
- */
1561
- experimental_continuationSteps?: boolean;
1562
- /**
1563
1532
  When enabled, the model will perform additional steps if the finish reason is "length" (experimental).
1564
1533
 
1565
1534
  By default, it's set to false.
@@ -1612,11 +1581,6 @@ declare class StreamData {
1612
1581
  * This assumes every chunk is a 'text' chunk.
1613
1582
  */
1614
1583
  declare function createStreamDataTransformer(): TransformStream<any, any>;
1615
- /**
1616
- @deprecated Use `StreamData` instead.
1617
- */
1618
- declare class experimental_StreamData extends StreamData {
1619
- }
1620
1584
 
1621
1585
  /**
1622
1586
  A result object for accessing different stream types and additional information.
@@ -1632,7 +1596,7 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
1632
1596
 
1633
1597
  Resolved when the response is finished.
1634
1598
  */
1635
- readonly usage: Promise<LanguageModelUsage$1>;
1599
+ readonly usage: Promise<LanguageModelUsage>;
1636
1600
  /**
1637
1601
  The reason why the generation finished. Taken from the last step.
1638
1602
 
@@ -1664,10 +1628,6 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
1664
1628
  */
1665
1629
  readonly toolResults: Promise<ToolResultUnion<TOOLS>[]>;
1666
1630
  /**
1667
- @deprecated use `response.messages` instead.
1668
- */
1669
- readonly responseMessages: Promise<Array<CoreAssistantMessage | CoreToolMessage>>;
1670
- /**
1671
1631
  Details for all steps.
1672
1632
  You can use this to get information about intermediate steps,
1673
1633
  such as the tool calls or the response headers.
@@ -1722,11 +1682,14 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
1722
1682
  Writes data stream output to a Node.js response-like object.
1723
1683
 
1724
1684
  @param response A Node.js response-like object (ServerResponse).
1725
- @param options An object with an init property (ResponseInit) and a data property.
1726
- You can also pass in a ResponseInit directly (deprecated).
1685
+ @param options.status The status code.
1686
+ @param options.statusText The status text.
1687
+ @param options.headers The headers.
1688
+ @param options.data The stream data.
1689
+ @param options.getErrorMessage An optional function that converts an error to an error message.
1690
+ @param options.sendUsage Whether to send the usage information to the client. Defaults to true.
1727
1691
  */
1728
- pipeDataStreamToResponse(response: ServerResponse$1, options?: ResponseInit | {
1729
- init?: ResponseInit;
1692
+ pipeDataStreamToResponse(response: ServerResponse$1, options?: ResponseInit & {
1730
1693
  data?: StreamData;
1731
1694
  getErrorMessage?: (error: unknown) => string;
1732
1695
  sendUsage?: boolean;
@@ -1744,13 +1707,16 @@ interface StreamTextResult<TOOLS extends Record<string, CoreTool>> {
1744
1707
  Converts the result to a streamed response object with a stream data part stream.
1745
1708
  It can be used with the `useChat` and `useCompletion` hooks.
1746
1709
 
1747
- @param options An object with an init property (ResponseInit) and a data property.
1748
- You can also pass in a ResponseInit directly (deprecated).
1710
+ @param options.status The status code.
1711
+ @param options.statusText The status text.
1712
+ @param options.headers The headers.
1713
+ @param options.data The stream data.
1714
+ @param options.getErrorMessage An optional function that converts an error to an error message.
1715
+ @param options.sendUsage Whether to send the usage information to the client. Defaults to true.
1749
1716
 
1750
1717
  @return A response object.
1751
1718
  */
1752
- toDataStreamResponse(options?: ResponseInit | {
1753
- init?: ResponseInit;
1719
+ toDataStreamResponse(options?: ResponseInit & {
1754
1720
  data?: StreamData;
1755
1721
  getErrorMessage?: (error: unknown) => string;
1756
1722
  sendUsage?: boolean;
@@ -1784,7 +1750,7 @@ type TextStreamPart<TOOLS extends Record<string, CoreTool>> = {
1784
1750
  type: 'step-finish';
1785
1751
  finishReason: FinishReason;
1786
1752
  logprobs?: LogProbs;
1787
- usage: LanguageModelUsage$1;
1753
+ usage: LanguageModelUsage;
1788
1754
  response: LanguageModelResponseMetadata;
1789
1755
  experimental_providerMetadata?: ProviderMetadata;
1790
1756
  isContinued: boolean;
@@ -1792,7 +1758,7 @@ type TextStreamPart<TOOLS extends Record<string, CoreTool>> = {
1792
1758
  type: 'finish';
1793
1759
  finishReason: FinishReason;
1794
1760
  logprobs?: LogProbs;
1795
- usage: LanguageModelUsage$1;
1761
+ usage: LanguageModelUsage;
1796
1762
  response: LanguageModelResponseMetadata;
1797
1763
  experimental_providerMetadata?: ProviderMetadata;
1798
1764
  } | {
@@ -2034,70 +2000,10 @@ declare class NoSuchProviderError extends NoSuchModelError {
2034
2000
  static isInstance(error: unknown): error is NoSuchProviderError;
2035
2001
  }
2036
2002
 
2037
- /**
2038
- * Provides for language and text embedding models.
2039
- *
2040
- * @deprecated Use `ProviderV1` instead.
2041
- */
2042
- interface experimental_Provider {
2043
- /**
2044
- Returns the language model with the given id in the format `providerId:modelId`.
2045
- The model id is then passed to the provider function to get the model.
2046
-
2047
- @param {string} id - The id of the model to return.
2048
-
2049
- @throws {NoSuchModelError} If no model with the given id exists.
2050
- @throws {NoSuchProviderError} If no provider with the given id exists.
2051
-
2052
- @returns {LanguageModel} The language model associated with the id.
2053
- */
2054
- languageModel?: (modelId: string) => LanguageModel;
2055
- /**
2056
- Returns the text embedding model with the given id in the format `providerId:modelId`.
2057
- The model id is then passed to the provider function to get the model.
2058
-
2059
- @param {string} id - The id of the model to return.
2060
-
2061
- @throws {NoSuchModelError} If no model with the given id exists.
2062
- @throws {NoSuchProviderError} If no provider with the given id exists.
2063
-
2064
- @returns {LanguageModel} The language model associated with the id.
2065
- */
2066
- textEmbeddingModel?: (modelId: string) => EmbeddingModel<string>;
2067
- /**
2068
- Returns the text embedding model with the given id in the format `providerId:modelId`.
2069
- The model id is then passed to the provider function to get the model.
2070
-
2071
- @param {string} id - The id of the model to return.
2072
-
2073
- @throws {NoSuchModelError} If no model with the given id exists.
2074
- @throws {NoSuchProviderError} If no provider with the given id exists.
2075
-
2076
- @returns {LanguageModel} The language model associated with the id.
2077
-
2078
- @deprecated use `textEmbeddingModel` instead.
2079
- */
2080
- textEmbedding?: (modelId: string) => EmbeddingModel<string>;
2081
- }
2082
-
2083
- /**
2084
- Registry for managing models. It enables getting a model with a string id.
2085
-
2086
- @deprecated Use `experimental_Provider` instead.
2087
- */
2088
- type experimental_ProviderRegistry = Provider;
2089
- /**
2090
- * @deprecated Use `experimental_ProviderRegistry` instead.
2091
- */
2092
- type experimental_ModelRegistry = experimental_ProviderRegistry;
2093
2003
  /**
2094
2004
  * Creates a registry for the given providers.
2095
2005
  */
2096
- declare function experimental_createProviderRegistry(providers: Record<string, experimental_Provider | Provider>): Provider;
2097
- /**
2098
- * @deprecated Use `experimental_createProviderRegistry` instead.
2099
- */
2100
- declare const experimental_createModelRegistry: typeof experimental_createProviderRegistry;
2006
+ declare function experimental_createProviderRegistry(providers: Record<string, Provider>): Provider;
2101
2007
 
2102
2008
  /**
2103
2009
  * Calculates the cosine similarity between two vectors. This is a useful metric for
@@ -2246,14 +2152,6 @@ The process parameter is a callback in which you can run the assistant on thread
2246
2152
  */
2247
2153
  type AssistantResponseCallback = (options: {
2248
2154
  /**
2249
- @deprecated use variable from outer scope instead.
2250
- */
2251
- threadId: string;
2252
- /**
2253
- @deprecated use variable from outer scope instead.
2254
- */
2255
- messageId: string;
2256
- /**
2257
2155
  Forwards an assistant message (non-streaming) to the client.
2258
2156
  */
2259
2157
  sendMessage: (message: AssistantMessage) => void;
@@ -2272,10 +2170,6 @@ It is designed to facilitate streaming assistant responses to the `useAssistant`
2272
2170
  It receives an assistant thread and a current message, and can send messages and data messages to the client.
2273
2171
  */
2274
2172
  declare function AssistantResponse({ threadId, messageId }: AssistantResponseSettings, process: AssistantResponseCallback): Response;
2275
- /**
2276
- @deprecated Use `AssistantResponse` instead.
2277
- */
2278
- declare const experimental_AssistantResponse: typeof AssistantResponse;
2279
2173
 
2280
2174
  /**
2281
2175
  * Configuration options and helper callback methods for stream lifecycle events.
@@ -2321,16 +2215,6 @@ type LangChainStreamEvent = {
2321
2215
  /**
2322
2216
  Converts LangChain output streams to AIStream.
2323
2217
 
2324
- The following streams are supported:
2325
- - `LangChainAIMessageChunk` streams (LangChain `model.stream` output)
2326
- - `string` streams (LangChain `StringOutputParser` output)
2327
-
2328
- @deprecated Use `toDataStream` instead.
2329
- */
2330
- declare function toAIStream(stream: ReadableStream<LangChainStreamEvent> | ReadableStream<LangChainAIMessageChunk> | ReadableStream<string>, callbacks?: StreamCallbacks): ReadableStream<any>;
2331
- /**
2332
- Converts LangChain output streams to AIStream.
2333
-
2334
2218
  The following streams are supported:
2335
2219
  - `LangChainAIMessageChunk` streams (LangChain `model.stream` output)
2336
2220
  - `string` streams (LangChain `StringOutputParser` output)
@@ -2342,10 +2226,8 @@ declare function toDataStreamResponse$1(stream: ReadableStream<LangChainStreamEv
2342
2226
  callbacks?: StreamCallbacks;
2343
2227
  }): Response;
2344
2228
 
2345
- declare const langchainAdapter_toAIStream: typeof toAIStream;
2346
2229
  declare namespace langchainAdapter {
2347
2230
  export {
2348
- langchainAdapter_toAIStream as toAIStream,
2349
2231
  toDataStream$1 as toDataStream,
2350
2232
  toDataStreamResponse$1 as toDataStreamResponse,
2351
2233
  };
@@ -2370,4 +2252,4 @@ declare namespace llamaindexAdapter {
2370
2252
  };
2371
2253
  }
2372
2254
 
2373
- export { AssistantContent, AssistantResponse, CallWarning, CompletionTokenUsage, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, ToolCallUnion as CoreToolCallUnion, CoreToolChoice, CoreToolMessage, ToolResultUnion as CoreToolResultUnion, CoreUserMessage, DataContent, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, EmbeddingTokenUsage, ExperimentalTool, Experimental_LanguageModelV1Middleware, FilePart, FinishReason, GenerateObjectResult, GenerateTextResult, ImagePart, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidToolArgumentsError, langchainAdapter as LangChainAdapter, LanguageModel, LanguageModelRequestMetadata, LanguageModelResponseMetadata, LanguageModelResponseMetadataWithHeaders, LanguageModelUsage, llamaindexAdapter as LlamaIndexAdapter, LogProbs, MessageConversionError, NoObjectGeneratedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, Provider, ProviderMetadata, RetryError, StepResult, StreamData, StreamObjectResult, StreamTextResult, TextPart, TextStreamPart, TokenUsage, ToolCallPart, ToolContent, ToolResultPart, UserContent, convertToCoreMessages, cosineSimilarity, createStreamDataTransformer, embed, embedMany, experimental_AssistantResponse, experimental_ModelRegistry, experimental_Provider, experimental_ProviderRegistry, experimental_StreamData, experimental_createModelRegistry, experimental_createProviderRegistry, experimental_customProvider, experimental_wrapLanguageModel, generateObject, generateText, streamObject, streamText, tool };
2255
+ export { AssistantContent, AssistantResponse, CallWarning, CoreAssistantMessage, CoreMessage, CoreSystemMessage, CoreTool, ToolCallUnion as CoreToolCallUnion, CoreToolChoice, CoreToolMessage, ToolResultUnion as CoreToolResultUnion, CoreUserMessage, DataContent, DownloadError, EmbedManyResult, EmbedResult, Embedding, EmbeddingModel, EmbeddingModelUsage, Experimental_LanguageModelV1Middleware, FilePart, FinishReason, GenerateObjectResult, GenerateTextResult, ImagePart, InvalidArgumentError, InvalidDataContentError, InvalidMessageRoleError, InvalidToolArgumentsError, langchainAdapter as LangChainAdapter, LanguageModel, LanguageModelRequestMetadata, LanguageModelResponseMetadata, LanguageModelUsage, llamaindexAdapter as LlamaIndexAdapter, LogProbs, MessageConversionError, NoObjectGeneratedError, NoSuchProviderError, NoSuchToolError, ObjectStreamPart, Provider, ProviderMetadata, RetryError, StepResult, StreamData, StreamObjectResult, StreamTextResult, TextPart, TextStreamPart, ToolCallPart, ToolContent, ToolResultPart, UserContent, convertToCoreMessages, cosineSimilarity, createStreamDataTransformer, embed, embedMany, experimental_createProviderRegistry, experimental_customProvider, experimental_wrapLanguageModel, generateObject, generateText, streamObject, streamText, tool };