ai 4.0.0-canary.6 → 4.0.0-canary.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -1490,11 +1490,14 @@ function standardizePrompt({
1490
1490
  }
1491
1491
 
1492
1492
  // core/types/usage.ts
1493
- function calculateLanguageModelUsage(usage) {
1493
+ function calculateLanguageModelUsage({
1494
+ promptTokens,
1495
+ completionTokens
1496
+ }) {
1494
1497
  return {
1495
- promptTokens: usage.promptTokens,
1496
- completionTokens: usage.completionTokens,
1497
- totalTokens: usage.promptTokens + usage.completionTokens
1498
+ promptTokens,
1499
+ completionTokens,
1500
+ totalTokens: promptTokens + completionTokens
1498
1501
  };
1499
1502
  }
1500
1503
 
@@ -3115,8 +3118,7 @@ async function generateText({
3115
3118
  abortSignal,
3116
3119
  headers,
3117
3120
  maxSteps = 1,
3118
- experimental_continuationSteps,
3119
- experimental_continueSteps: continueSteps = experimental_continuationSteps != null ? experimental_continuationSteps : false,
3121
+ experimental_continueSteps: continueSteps = false,
3120
3122
  experimental_telemetry: telemetry,
3121
3123
  experimental_providerMetadata: providerMetadata,
3122
3124
  experimental_activeTools: activeTools,
@@ -3388,7 +3390,6 @@ async function generateText({
3388
3390
  messages: responseMessages
3389
3391
  },
3390
3392
  logprobs: currentModelResponse.logprobs,
3391
- responseMessages,
3392
3393
  steps,
3393
3394
  providerMetadata: currentModelResponse.providerMetadata
3394
3395
  });
@@ -3465,7 +3466,6 @@ var DefaultGenerateTextResult = class {
3465
3466
  this.warnings = options.warnings;
3466
3467
  this.request = options.request;
3467
3468
  this.response = options.response;
3468
- this.responseMessages = options.responseMessages;
3469
3469
  this.steps = options.steps;
3470
3470
  this.experimental_providerMetadata = options.providerMetadata;
3471
3471
  this.logprobs = options.logprobs;
@@ -4053,11 +4053,6 @@ var DefaultStreamTextResult = class {
4053
4053
  this.request = requestPromise;
4054
4054
  const { resolve: resolveResponse, promise: responsePromise } = createResolvablePromise();
4055
4055
  this.response = responsePromise;
4056
- const {
4057
- resolve: resolveResponseMessages,
4058
- promise: responseMessagesPromise
4059
- } = createResolvablePromise();
4060
- this.responseMessages = responseMessagesPromise;
4061
4056
  const {
4062
4057
  stream: stitchableStream,
4063
4058
  addStream,
@@ -4383,7 +4378,6 @@ var DefaultStreamTextResult = class {
4383
4378
  messages: responseMessages
4384
4379
  });
4385
4380
  resolveSteps(stepResults);
4386
- resolveResponseMessages(responseMessages);
4387
4381
  await (onFinish == null ? void 0 : onFinish({
4388
4382
  finishReason: stepFinishReason,
4389
4383
  logprobs: stepLogProbs,
@@ -4723,15 +4717,11 @@ function experimental_createProviderRegistry(providers) {
4723
4717
  }
4724
4718
  return registry;
4725
4719
  }
4726
- var experimental_createModelRegistry = experimental_createProviderRegistry;
4727
4720
  var DefaultProviderRegistry = class {
4728
4721
  constructor() {
4729
4722
  this.providers = {};
4730
4723
  }
4731
- registerProvider({
4732
- id,
4733
- provider
4734
- }) {
4724
+ registerProvider({ id, provider }) {
4735
4725
  this.providers[id] = provider;
4736
4726
  }
4737
4727
  getProvider(id) {
@@ -4767,10 +4757,10 @@ var DefaultProviderRegistry = class {
4767
4757
  return model;
4768
4758
  }
4769
4759
  textEmbeddingModel(id) {
4770
- var _a11, _b, _c;
4760
+ var _a11;
4771
4761
  const [providerId, modelId] = this.splitId(id, "textEmbeddingModel");
4772
4762
  const provider = this.getProvider(providerId);
4773
- const model = (_c = (_a11 = provider.textEmbeddingModel) == null ? void 0 : _a11.call(provider, modelId)) != null ? _c : "textEmbedding" in provider ? (_b = provider.textEmbedding) == null ? void 0 : _b.call(provider, modelId) : void 0;
4763
+ const model = (_a11 = provider.textEmbeddingModel) == null ? void 0 : _a11.call(provider, modelId);
4774
4764
  if (model == null) {
4775
4765
  throw new NoSuchModelError4({
4776
4766
  modelId: id,
@@ -4882,8 +4872,6 @@ function AssistantResponse({ threadId, messageId }, process2) {
4882
4872
  );
4883
4873
  try {
4884
4874
  await process2({
4885
- threadId,
4886
- messageId,
4887
4875
  sendMessage,
4888
4876
  sendDataMessage,
4889
4877
  forwardStream
@@ -4906,12 +4894,10 @@ function AssistantResponse({ threadId, messageId }, process2) {
4906
4894
  }
4907
4895
  });
4908
4896
  }
4909
- var experimental_AssistantResponse = AssistantResponse;
4910
4897
 
4911
4898
  // streams/langchain-adapter.ts
4912
4899
  var langchain_adapter_exports = {};
4913
4900
  __export(langchain_adapter_exports, {
4914
- toAIStream: () => toAIStream,
4915
4901
  toDataStream: () => toDataStream,
4916
4902
  toDataStreamResponse: () => toDataStreamResponse
4917
4903
  });
@@ -5020,13 +5006,8 @@ function createStreamDataTransformer() {
5020
5006
  }
5021
5007
  });
5022
5008
  }
5023
- var experimental_StreamData = class extends StreamData {
5024
- };
5025
5009
 
5026
5010
  // streams/langchain-adapter.ts
5027
- function toAIStream(stream, callbacks) {
5028
- return toDataStream(stream, callbacks);
5029
- }
5030
5011
  function toDataStream(stream, callbacks) {
5031
5012
  return stream.pipeThrough(
5032
5013
  new TransformStream({
@@ -5151,9 +5132,6 @@ export {
5151
5132
  createStreamDataTransformer,
5152
5133
  embed,
5153
5134
  embedMany,
5154
- experimental_AssistantResponse,
5155
- experimental_StreamData,
5156
- experimental_createModelRegistry,
5157
5135
  experimental_createProviderRegistry,
5158
5136
  experimental_customProvider,
5159
5137
  experimental_wrapLanguageModel,