@ai-sdk/openai 3.0.0-beta.18 → 3.0.0-beta.19

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,11 @@
1
1
  # @ai-sdk/openai
2
2
 
3
+ ## 3.0.0-beta.19
4
+
5
+ ### Patch Changes
6
+
7
+ - 484aa93: Add 'default' as service tier
8
+
3
9
  ## 3.0.0-beta.18
4
10
 
5
11
  ### Patch Changes
package/dist/index.d.mts CHANGED
@@ -14,7 +14,7 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazyValidat
14
14
  metadata?: Record<string, string> | undefined;
15
15
  prediction?: Record<string, any> | undefined;
16
16
  structuredOutputs?: boolean | undefined;
17
- serviceTier?: "auto" | "flex" | "priority" | undefined;
17
+ serviceTier?: "default" | "auto" | "flex" | "priority" | undefined;
18
18
  strictJsonSchema?: boolean | undefined;
19
19
  textVerbosity?: "low" | "medium" | "high" | undefined;
20
20
  promptCacheKey?: string | undefined;
@@ -263,7 +263,7 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyV
263
263
  reasoningEffort?: string | null | undefined;
264
264
  reasoningSummary?: string | null | undefined;
265
265
  safetyIdentifier?: string | null | undefined;
266
- serviceTier?: "auto" | "flex" | "priority" | null | undefined;
266
+ serviceTier?: "default" | "auto" | "flex" | "priority" | null | undefined;
267
267
  store?: boolean | null | undefined;
268
268
  strictJsonSchema?: boolean | null | undefined;
269
269
  textVerbosity?: "low" | "medium" | "high" | null | undefined;
package/dist/index.d.ts CHANGED
@@ -14,7 +14,7 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazyValidat
14
14
  metadata?: Record<string, string> | undefined;
15
15
  prediction?: Record<string, any> | undefined;
16
16
  structuredOutputs?: boolean | undefined;
17
- serviceTier?: "auto" | "flex" | "priority" | undefined;
17
+ serviceTier?: "default" | "auto" | "flex" | "priority" | undefined;
18
18
  strictJsonSchema?: boolean | undefined;
19
19
  textVerbosity?: "low" | "medium" | "high" | undefined;
20
20
  promptCacheKey?: string | undefined;
@@ -263,7 +263,7 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyV
263
263
  reasoningEffort?: string | null | undefined;
264
264
  reasoningSummary?: string | null | undefined;
265
265
  safetyIdentifier?: string | null | undefined;
266
- serviceTier?: "auto" | "flex" | "priority" | null | undefined;
266
+ serviceTier?: "default" | "auto" | "flex" | "priority" | null | undefined;
267
267
  store?: boolean | null | undefined;
268
268
  strictJsonSchema?: boolean | null | undefined;
269
269
  textVerbosity?: "low" | "medium" | "high" | null | undefined;
package/dist/index.js CHANGED
@@ -472,13 +472,15 @@ var openaiChatLanguageModelOptions = (0, import_provider_utils4.lazyValidator)(
472
472
  structuredOutputs: z3.boolean().optional(),
473
473
  /**
474
474
  * Service tier for the request.
475
- * - 'auto': Default service tier
475
+ * - 'auto': Default service tier. The request will be processed with the service tier configured in the
476
+ * Project settings. Unless otherwise configured, the Project will use 'default'.
476
477
  * - 'flex': 50% cheaper processing at the cost of increased latency. Only available for o3 and o4-mini models.
477
478
  * - 'priority': Higher-speed processing with predictably low latency at premium cost. Available for Enterprise customers.
479
+ * - 'default': The request will be processed with the standard pricing and performance for the selected model.
478
480
  *
479
481
  * @default 'auto'
480
482
  */
481
- serviceTier: z3.enum(["auto", "flex", "priority"]).optional(),
483
+ serviceTier: z3.enum(["auto", "flex", "priority", "default"]).optional(),
482
484
  /**
483
485
  * Whether to use strict JSON schema validation.
484
486
  *
@@ -2915,7 +2917,7 @@ var openaiResponsesProviderOptionsSchema = (0, import_provider_utils22.lazyValid
2915
2917
  reasoningEffort: z17.string().nullish(),
2916
2918
  reasoningSummary: z17.string().nullish(),
2917
2919
  safetyIdentifier: z17.string().nullish(),
2918
- serviceTier: z17.enum(["auto", "flex", "priority"]).nullish(),
2920
+ serviceTier: z17.enum(["auto", "flex", "priority", "default"]).nullish(),
2919
2921
  store: z17.boolean().nullish(),
2920
2922
  strictJsonSchema: z17.boolean().nullish(),
2921
2923
  textVerbosity: z17.enum(["low", "medium", "high"]).nullish(),
@@ -4422,7 +4424,7 @@ var OpenAITranscriptionModel = class {
4422
4424
  };
4423
4425
 
4424
4426
  // src/version.ts
4425
- var VERSION = true ? "3.0.0-beta.18" : "0.0.0-test";
4427
+ var VERSION = true ? "3.0.0-beta.19" : "0.0.0-test";
4426
4428
 
4427
4429
  // src/openai-provider.ts
4428
4430
  function createOpenAI(options = {}) {