@yourgpt/llm-sdk 2.0.1 → 2.0.2-beta.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (55) hide show
  1. package/dist/adapters/index.d.mts +27 -4
  2. package/dist/adapters/index.d.ts +27 -4
  3. package/dist/adapters/index.js +395 -25
  4. package/dist/adapters/index.js.map +1 -1
  5. package/dist/adapters/index.mjs +395 -25
  6. package/dist/adapters/index.mjs.map +1 -1
  7. package/dist/index.d.mts +17 -5
  8. package/dist/index.d.ts +17 -5
  9. package/dist/index.js +18 -3
  10. package/dist/index.js.map +1 -1
  11. package/dist/index.mjs +18 -3
  12. package/dist/index.mjs.map +1 -1
  13. package/dist/providers/anthropic/index.d.mts +1 -2
  14. package/dist/providers/anthropic/index.d.ts +1 -2
  15. package/dist/providers/anthropic/index.js +108 -12
  16. package/dist/providers/anthropic/index.js.map +1 -1
  17. package/dist/providers/anthropic/index.mjs +108 -12
  18. package/dist/providers/anthropic/index.mjs.map +1 -1
  19. package/dist/providers/azure/index.d.mts +1 -2
  20. package/dist/providers/azure/index.d.ts +1 -2
  21. package/dist/providers/azure/index.js.map +1 -1
  22. package/dist/providers/azure/index.mjs.map +1 -1
  23. package/dist/providers/google/index.d.mts +1 -2
  24. package/dist/providers/google/index.d.ts +1 -2
  25. package/dist/providers/google/index.js +61 -2
  26. package/dist/providers/google/index.js.map +1 -1
  27. package/dist/providers/google/index.mjs +61 -2
  28. package/dist/providers/google/index.mjs.map +1 -1
  29. package/dist/providers/ollama/index.d.mts +8 -3
  30. package/dist/providers/ollama/index.d.ts +8 -3
  31. package/dist/providers/ollama/index.js +227 -17
  32. package/dist/providers/ollama/index.js.map +1 -1
  33. package/dist/providers/ollama/index.mjs +227 -17
  34. package/dist/providers/ollama/index.mjs.map +1 -1
  35. package/dist/providers/openai/index.d.mts +1 -2
  36. package/dist/providers/openai/index.d.ts +1 -2
  37. package/dist/providers/openai/index.js +57 -3
  38. package/dist/providers/openai/index.js.map +1 -1
  39. package/dist/providers/openai/index.mjs +57 -3
  40. package/dist/providers/openai/index.mjs.map +1 -1
  41. package/dist/providers/openrouter/index.d.mts +56 -3
  42. package/dist/providers/openrouter/index.d.ts +56 -3
  43. package/dist/providers/openrouter/index.js +90 -276
  44. package/dist/providers/openrouter/index.js.map +1 -1
  45. package/dist/providers/openrouter/index.mjs +89 -277
  46. package/dist/providers/openrouter/index.mjs.map +1 -1
  47. package/dist/providers/xai/index.d.mts +1 -2
  48. package/dist/providers/xai/index.d.ts +1 -2
  49. package/dist/providers/xai/index.js.map +1 -1
  50. package/dist/providers/xai/index.mjs.map +1 -1
  51. package/dist/{base-DdxolpKP.d.mts → types-ClJBhmS0.d.mts} +427 -3
  52. package/dist/{base-DdxolpKP.d.ts → types-ClJBhmS0.d.ts} +427 -3
  53. package/package.json +1 -1
  54. package/dist/types-Ck25ZYma.d.mts +0 -323
  55. package/dist/types-Dsz8SpdB.d.ts +0 -323
@@ -342,7 +342,7 @@ interface ResponseOptions {
342
342
  /**
343
343
  * Stream event types
344
344
  */
345
- type StreamEventType = "message:start" | "message:delta" | "message:end" | "thinking:start" | "thinking:delta" | "thinking:end" | "action:start" | "action:args" | "action:end" | "tool_calls" | "tool:result" | "loop:iteration" | "loop:complete" | "error" | "done";
345
+ type StreamEventType = "message:start" | "message:delta" | "message:end" | "thinking:start" | "thinking:delta" | "thinking:end" | "action:start" | "action:args" | "action:end" | "tool_calls" | "tool:result" | "citation" | "loop:iteration" | "loop:complete" | "error" | "done";
346
346
  /**
347
347
  * Base event interface
348
348
  */
@@ -479,6 +479,30 @@ interface LoopCompleteEvent extends BaseEvent {
479
479
  aborted?: boolean;
480
480
  maxIterationsReached?: boolean;
481
481
  }
482
+ /**
483
+ * Citation from web search (unified format for all providers)
484
+ */
485
+ interface Citation {
486
+ /** Unique citation index (1-based) */
487
+ index: number;
488
+ /** Source URL */
489
+ url: string;
490
+ /** Page title */
491
+ title: string;
492
+ /** Cited text snippet (optional) */
493
+ citedText?: string;
494
+ /** Source domain (extracted from URL) */
495
+ domain?: string;
496
+ /** Favicon URL (generated from domain) */
497
+ favicon?: string;
498
+ }
499
+ /**
500
+ * Citation event - web search returned citations
501
+ */
502
+ interface CitationEvent extends BaseEvent {
503
+ type: "citation";
504
+ citations: Citation[];
505
+ }
482
506
  /**
483
507
  * Message format for done event (API format with snake_case)
484
508
  */
@@ -516,7 +540,7 @@ interface DoneEvent extends BaseEvent {
516
540
  /**
517
541
  * Union of all stream events
518
542
  */
519
- type StreamEvent = MessageStartEvent | MessageDeltaEvent | MessageEndEvent | ThinkingStartEvent | ThinkingDeltaEvent | ThinkingEndEvent | ActionStartEvent | ActionArgsEvent | ActionEndEvent | ToolCallsEvent | ToolResultEvent | LoopIterationEvent | LoopCompleteEvent | ErrorEvent | DoneEvent;
543
+ type StreamEvent = MessageStartEvent | MessageDeltaEvent | MessageEndEvent | ThinkingStartEvent | ThinkingDeltaEvent | ThinkingEndEvent | ActionStartEvent | ActionArgsEvent | ActionEndEvent | ToolCallsEvent | ToolResultEvent | CitationEvent | LoopIterationEvent | LoopCompleteEvent | ErrorEvent | DoneEvent;
520
544
  /**
521
545
  * LLM configuration
522
546
  */
@@ -671,6 +695,49 @@ interface AgentLoopConfig {
671
695
  debug?: boolean;
672
696
  enabled?: boolean;
673
697
  }
698
+ /**
699
+ * Web search configuration for native provider search
700
+ *
701
+ * Enables native web search for supported providers:
702
+ * - Anthropic: Uses Claude's built-in web search tool
703
+ * - OpenAI: Uses GPT's web search preview
704
+ * - Google: Uses Gemini's Google Search grounding
705
+ *
706
+ * @example
707
+ * ```typescript
708
+ * const runtime = createRuntime({
709
+ * provider: createAnthropic({ apiKey: '...' }),
710
+ * model: 'claude-sonnet-4-20250514',
711
+ * webSearch: true, // Enable with defaults
712
+ * });
713
+ *
714
+ * // Or with configuration
715
+ * const runtime = createRuntime({
716
+ * provider: createOpenAI({ apiKey: '...' }),
717
+ * model: 'gpt-4o',
718
+ * webSearch: {
719
+ * maxUses: 5,
720
+ * allowedDomains: ['docs.anthropic.com', 'openai.com'],
721
+ * },
722
+ * });
723
+ * ```
724
+ */
725
+ interface WebSearchConfig {
726
+ /** Maximum number of search uses per request (default: unlimited) */
727
+ maxUses?: number;
728
+ /** Only search these domains (provider-specific support) */
729
+ allowedDomains?: string[];
730
+ /** Exclude these domains from search (provider-specific support) */
731
+ blockedDomains?: string[];
732
+ /** User location for localized results (Anthropic only) */
733
+ userLocation?: {
734
+ type: "approximate";
735
+ city?: string;
736
+ region?: string;
737
+ country?: string;
738
+ timezone?: string;
739
+ };
740
+ }
674
741
  /**
675
742
  * Unified tool call format
676
743
  */
@@ -741,6 +808,11 @@ interface ChatCompletionRequest {
741
808
  config?: RequestLLMConfig;
742
809
  /** Abort signal for cancellation */
743
810
  signal?: AbortSignal;
811
+ /**
812
+ * Enable native web search for the provider.
813
+ * When true or configured, the provider's native search is enabled.
814
+ */
815
+ webSearch?: boolean | WebSearchConfig;
744
816
  }
745
817
  /**
746
818
  * Non-streaming completion result
@@ -955,4 +1027,356 @@ type OpenAIMessage = {
955
1027
  */
956
1028
  declare function formatMessagesForOpenAI(messages: Message[], systemPrompt?: string): OpenAIMessage[];
957
1029
 
958
- export { formatTools as $, type ActionDefinition as A, type FinishChunk as B, type CoreMessage as C, type DoneEventMessage as D, type ErrorChunk as E, type FilePart as F, type GenerateTextParams as G, type TokenUsage as H, type ImagePart as I, type FinishReason as J, type KnowledgeBaseConfig as K, type LLMAdapter as L, type Message as M, DEFAULT_CAPABILITIES as N, type ChatCompletionRequest as O, type AdapterFactory as P, type LLMConfig as Q, type ResponseOptions as R, type StreamTextParams as S, type ToolContext$1 as T, type UserMessage as U, type ToolLocation as V, type UnifiedToolCall as W, type UnifiedToolResult as X, type ToolExecution as Y, type CompletionResult as Z, formatMessages as _, type GenerateTextResult as a, formatMessagesForAnthropic as a0, formatMessagesForOpenAI as a1, messageToAnthropicContent as a2, messageToOpenAIContent as a3, hasImageAttachments as a4, hasMediaAttachments as a5, attachmentToAnthropicImage as a6, attachmentToAnthropicDocument as a7, attachmentToOpenAIImage as a8, type AnthropicContentBlock as a9, type OpenAIContentBlock as aa, type StreamTextResult as b, type Tool as c, type ToolDefinition as d, type AgentLoopConfig as e, type StreamEvent as f, type ToolCallInfo as g, type TokenUsageRaw as h, type ToolResponse as i, type LanguageModel as j, type ModelCapabilities as k, type DoGenerateParams as l, type DoGenerateResult as m, type SystemMessage as n, type AssistantMessage as o, type ToolMessage as p, type UserContentPart as q, type TextPart as r, type ToolCall$1 as s, type ToolResult as t, type GenerateStep as u, type StreamPart as v, type StreamChunk as w, type TextDeltaChunk as x, type ToolCallChunk as y, type ToolResultChunk as z };
1030
+ /**
1031
+ * Provider Types
1032
+ *
1033
+ * Defines interfaces for:
1034
+ * 1. Provider Formatters (for tool transformations in agent loop)
1035
+ * 2. Multi-provider architecture (AIProvider, capabilities, configs)
1036
+ */
1037
+
1038
+ /**
1039
+ * Provider formatter interface
1040
+ *
1041
+ * Each provider implements this interface to handle:
1042
+ * - Tool definition transformation
1043
+ * - Tool call parsing from responses
1044
+ * - Tool result formatting
1045
+ * - Stop reason detection
1046
+ */
1047
+ interface ProviderFormatter {
1048
+ /**
1049
+ * Transform unified tool definitions to provider format
1050
+ */
1051
+ transformTools(tools: ToolDefinition[]): unknown[];
1052
+ /**
1053
+ * Parse tool calls from provider response
1054
+ */
1055
+ parseToolCalls(response: unknown): UnifiedToolCall[];
1056
+ /**
1057
+ * Format tool results for provider
1058
+ */
1059
+ formatToolResults(results: UnifiedToolResult[]): unknown[];
1060
+ /**
1061
+ * Check if response indicates tool use is requested
1062
+ */
1063
+ isToolUseStop(response: unknown): boolean;
1064
+ /**
1065
+ * Check if response indicates end of turn
1066
+ */
1067
+ isEndTurnStop(response: unknown): boolean;
1068
+ /**
1069
+ * Get stop reason string from response
1070
+ */
1071
+ getStopReason(response: unknown): string;
1072
+ /**
1073
+ * Extract text content from response
1074
+ */
1075
+ extractTextContent(response: unknown): string;
1076
+ /**
1077
+ * Build assistant message with tool calls for conversation history
1078
+ */
1079
+ buildAssistantToolMessage(toolCalls: UnifiedToolCall[], textContent?: string): unknown;
1080
+ /**
1081
+ * Build user message with tool results for conversation history
1082
+ */
1083
+ buildToolResultMessage(results: UnifiedToolResult[]): unknown;
1084
+ }
1085
+ /**
1086
+ * Anthropic tool definition format
1087
+ */
1088
+ interface AnthropicTool {
1089
+ name: string;
1090
+ description: string;
1091
+ input_schema: {
1092
+ type: "object";
1093
+ properties: Record<string, unknown>;
1094
+ required?: string[];
1095
+ };
1096
+ }
1097
+ /**
1098
+ * Anthropic tool_use block from response
1099
+ */
1100
+ interface AnthropicToolUse {
1101
+ type: "tool_use";
1102
+ id: string;
1103
+ name: string;
1104
+ input: Record<string, unknown>;
1105
+ }
1106
+ /**
1107
+ * Anthropic tool_result block
1108
+ */
1109
+ interface AnthropicToolResult {
1110
+ type: "tool_result";
1111
+ tool_use_id: string;
1112
+ content: string;
1113
+ }
1114
+ /**
1115
+ * OpenAI tool definition format
1116
+ */
1117
+ interface OpenAITool {
1118
+ type: "function";
1119
+ function: {
1120
+ name: string;
1121
+ description: string;
1122
+ parameters: {
1123
+ type: "object";
1124
+ properties: Record<string, unknown>;
1125
+ required?: string[];
1126
+ };
1127
+ };
1128
+ }
1129
+ /**
1130
+ * OpenAI tool call from response
1131
+ */
1132
+ interface OpenAIToolCall {
1133
+ id: string;
1134
+ type: "function";
1135
+ function: {
1136
+ name: string;
1137
+ arguments: string;
1138
+ };
1139
+ }
1140
+ /**
1141
+ * OpenAI tool result message
1142
+ */
1143
+ interface OpenAIToolResult {
1144
+ role: "tool";
1145
+ tool_call_id: string;
1146
+ content: string;
1147
+ }
1148
+ /**
1149
+ * Google Gemini function declaration
1150
+ */
1151
+ interface GeminiFunctionDeclaration {
1152
+ name: string;
1153
+ description: string;
1154
+ parameters?: {
1155
+ type: "object";
1156
+ properties: Record<string, unknown>;
1157
+ required?: string[];
1158
+ };
1159
+ }
1160
+ /**
1161
+ * Gemini function call from response
1162
+ */
1163
+ interface GeminiFunctionCall {
1164
+ name: string;
1165
+ args: Record<string, unknown>;
1166
+ }
1167
+ /**
1168
+ * Gemini function response
1169
+ */
1170
+ interface GeminiFunctionResponse {
1171
+ name: string;
1172
+ response: Record<string, unknown>;
1173
+ }
1174
+ /**
1175
+ * Capabilities of a model for UI feature flags
1176
+ * UI components can use this to enable/disable features
1177
+ */
1178
+ interface ProviderCapabilities {
1179
+ /** Supports image inputs */
1180
+ supportsVision: boolean;
1181
+ /** Supports tool/function calling */
1182
+ supportsTools: boolean;
1183
+ /** Supports extended thinking (Claude, DeepSeek) */
1184
+ supportsThinking: boolean;
1185
+ /** Supports streaming responses */
1186
+ supportsStreaming: boolean;
1187
+ /** Supports PDF document inputs */
1188
+ supportsPDF: boolean;
1189
+ /** Supports audio inputs */
1190
+ supportsAudio: boolean;
1191
+ /** Supports video inputs */
1192
+ supportsVideo: boolean;
1193
+ /** Maximum context tokens */
1194
+ maxTokens: number;
1195
+ /** Supported image MIME types */
1196
+ supportedImageTypes: string[];
1197
+ /** Supported audio MIME types */
1198
+ supportedAudioTypes?: string[];
1199
+ /** Supported video MIME types */
1200
+ supportedVideoTypes?: string[];
1201
+ /** Supports JSON mode / structured output */
1202
+ supportsJsonMode?: boolean;
1203
+ /** Supports system messages */
1204
+ supportsSystemMessages?: boolean;
1205
+ }
1206
+ /**
1207
+ * AI Provider interface (object form)
1208
+ *
1209
+ * Wraps existing LLMAdapter with additional metadata:
1210
+ * - Supported models list
1211
+ * - Per-model capabilities
1212
+ * - Provider name
1213
+ */
1214
+ interface AIProviderObject {
1215
+ /** Provider name (e.g., 'openai', 'anthropic') */
1216
+ readonly name: string;
1217
+ /** List of supported model IDs */
1218
+ readonly supportedModels: string[];
1219
+ /**
1220
+ * Get a language model adapter for the given model ID
1221
+ * Returns the existing LLMAdapter interface - no breaking changes
1222
+ */
1223
+ languageModel(modelId: string): LLMAdapter;
1224
+ /**
1225
+ * Get capabilities for a specific model
1226
+ * UI components use this to enable/disable features
1227
+ */
1228
+ getCapabilities(modelId: string): ProviderCapabilities;
1229
+ /**
1230
+ * Optional: Get an embedding model (future expansion)
1231
+ */
1232
+ embeddingModel?(modelId: string): EmbeddingModel;
1233
+ }
1234
+ /**
1235
+ * Callable AI Provider (Vercel AI SDK style)
1236
+ *
1237
+ * A function that returns a LanguageModel when called with a model ID,
1238
+ * but also has properties for provider metadata and methods.
1239
+ *
1240
+ * @example
1241
+ * ```typescript
1242
+ * const openai = createOpenAI({ apiKey: '...' });
1243
+ *
1244
+ * // Callable - returns LanguageModel directly (Vercel AI SDK style)
1245
+ * const model = openai('gpt-4o');
1246
+ *
1247
+ * // Also supports method calls (backward compatible)
1248
+ * const model2 = openai.languageModel('gpt-4o');
1249
+ *
1250
+ * // Check capabilities
1251
+ * const caps = openai.getCapabilities('gpt-4o');
1252
+ * if (caps.supportsVision) {
1253
+ * // Show image upload button
1254
+ * }
1255
+ * ```
1256
+ */
1257
+ interface AIProvider extends AIProviderObject {
1258
+ /**
1259
+ * Call the provider directly with a model ID to get a LanguageModel
1260
+ * This is the Vercel AI SDK style pattern
1261
+ */
1262
+ (modelId: string): LLMAdapter;
1263
+ }
1264
+ /**
1265
+ * Embedding model interface (for future expansion)
1266
+ */
1267
+ interface EmbeddingModel {
1268
+ readonly provider: string;
1269
+ readonly modelId: string;
1270
+ embed(texts: string[]): Promise<number[][]>;
1271
+ }
1272
+ /**
1273
+ * Base provider configuration
1274
+ */
1275
+ interface BaseProviderConfig {
1276
+ /** API key (falls back to environment variable) */
1277
+ apiKey?: string;
1278
+ /** Custom base URL */
1279
+ baseUrl?: string;
1280
+ /** Request timeout in milliseconds */
1281
+ timeout?: number;
1282
+ /** Custom headers to include */
1283
+ headers?: Record<string, string>;
1284
+ }
1285
+ /**
1286
+ * OpenAI provider configuration
1287
+ */
1288
+ interface OpenAIProviderConfig extends BaseProviderConfig {
1289
+ /** OpenAI organization ID */
1290
+ organization?: string;
1291
+ /** OpenAI project ID */
1292
+ project?: string;
1293
+ /** Vision detail level for images */
1294
+ imageDetail?: "auto" | "low" | "high";
1295
+ }
1296
+ /**
1297
+ * Anthropic provider configuration
1298
+ */
1299
+ interface AnthropicProviderConfig extends BaseProviderConfig {
1300
+ /** Extended thinking budget in tokens (minimum 1024) */
1301
+ thinkingBudget?: number;
1302
+ /** Enable prompt caching */
1303
+ cacheControl?: boolean;
1304
+ }
1305
+ /**
1306
+ * Google provider configuration
1307
+ */
1308
+ interface GoogleProviderConfig extends BaseProviderConfig {
1309
+ /** Safety settings */
1310
+ safetySettings?: GoogleSafetySetting[];
1311
+ /** Grounding configuration (for web search) */
1312
+ groundingConfig?: GoogleGroundingConfig;
1313
+ }
1314
+ /**
1315
+ * Google safety setting
1316
+ */
1317
+ interface GoogleSafetySetting {
1318
+ category: "HARM_CATEGORY_HARASSMENT" | "HARM_CATEGORY_HATE_SPEECH" | "HARM_CATEGORY_SEXUALLY_EXPLICIT" | "HARM_CATEGORY_DANGEROUS_CONTENT";
1319
+ threshold: "BLOCK_NONE" | "BLOCK_LOW_AND_ABOVE" | "BLOCK_MEDIUM_AND_ABOVE" | "BLOCK_HIGH_AND_ABOVE";
1320
+ }
1321
+ /**
1322
+ * Google grounding configuration
1323
+ */
1324
+ interface GoogleGroundingConfig {
1325
+ /** Enable Google Search grounding */
1326
+ googleSearchRetrieval?: boolean;
1327
+ }
1328
+ /**
1329
+ * xAI provider configuration
1330
+ */
1331
+ interface XAIProviderConfig extends BaseProviderConfig {
1332
+ }
1333
+ /**
1334
+ * Azure OpenAI provider configuration
1335
+ */
1336
+ interface AzureProviderConfig extends BaseProviderConfig {
1337
+ /** Azure resource name */
1338
+ resourceName: string;
1339
+ /** Deployment name */
1340
+ deploymentName: string;
1341
+ /** API version (default: 2024-02-15-preview) */
1342
+ apiVersion?: string;
1343
+ }
1344
+ /**
1345
+ * Ollama model-specific options
1346
+ * These map to Ollama's native API options
1347
+ */
1348
+ interface OllamaModelOptions {
1349
+ /** Context window size (default varies by model) */
1350
+ num_ctx?: number;
1351
+ /** Max tokens to predict (-1 = infinite, -2 = fill context) */
1352
+ num_predict?: number;
1353
+ /** Mirostat sampling (0 = disabled, 1 = Mirostat, 2 = Mirostat 2.0) */
1354
+ mirostat?: 0 | 1 | 2;
1355
+ /** Mirostat learning rate (default: 0.1) */
1356
+ mirostat_eta?: number;
1357
+ /** Mirostat target entropy (default: 5.0) */
1358
+ mirostat_tau?: number;
1359
+ /** Repeat penalty (default: 1.1) */
1360
+ repeat_penalty?: number;
1361
+ /** Random seed for reproducibility (-1 = random) */
1362
+ seed?: number;
1363
+ /** Top-k sampling (default: 40) */
1364
+ top_k?: number;
1365
+ /** Top-p (nucleus) sampling (default: 0.9) */
1366
+ top_p?: number;
1367
+ /** Min-p sampling (default: 0.0) */
1368
+ min_p?: number;
1369
+ /** Stop sequences */
1370
+ stop?: string[];
1371
+ /** Temperature override (also available in config) */
1372
+ temperature?: number;
1373
+ }
1374
+ /**
1375
+ * Ollama provider configuration
1376
+ */
1377
+ interface OllamaProviderConfig extends BaseProviderConfig {
1378
+ /** Default Ollama-specific model options */
1379
+ options?: OllamaModelOptions;
1380
+ }
1381
+
1382
+ export { type XAIProviderConfig as $, type AIProvider as A, type ToolResultChunk as B, type CoreMessage as C, type DoneEventMessage as D, type FinishChunk as E, type FilePart as F, type GenerateTextParams as G, type ErrorChunk as H, type ImagePart as I, type TokenUsage as J, type KnowledgeBaseConfig as K, type LLMAdapter as L, type Message as M, type FinishReason as N, DEFAULT_CAPABILITIES as O, type ChatCompletionRequest as P, type AdapterFactory as Q, type ResponseOptions as R, type StreamTextParams as S, type ToolContext$1 as T, type UserMessage as U, type ProviderCapabilities as V, type WebSearchConfig as W, type BaseProviderConfig as X, type OpenAIProviderConfig as Y, type AnthropicProviderConfig as Z, type GoogleProviderConfig as _, type GenerateTextResult as a, type AzureProviderConfig as a0, type OllamaProviderConfig as a1, type OllamaModelOptions as a2, type ProviderFormatter as a3, type AnthropicTool as a4, type AnthropicToolUse as a5, type AnthropicToolResult as a6, type OpenAITool as a7, type OpenAIToolCall as a8, type OpenAIToolResult as a9, type GeminiFunctionDeclaration as aa, type GeminiFunctionCall as ab, type GeminiFunctionResponse as ac, type LLMConfig as ad, type ToolLocation as ae, type UnifiedToolCall as af, type UnifiedToolResult as ag, type ToolExecution as ah, type Citation as ai, type CompletionResult as aj, formatMessages as ak, formatTools as al, formatMessagesForAnthropic as am, formatMessagesForOpenAI as an, messageToAnthropicContent as ao, messageToOpenAIContent as ap, hasImageAttachments as aq, hasMediaAttachments as ar, attachmentToAnthropicImage as as, attachmentToAnthropicDocument as at, attachmentToOpenAIImage as au, type AnthropicContentBlock as av, type OpenAIContentBlock as aw, type StreamTextResult as b, type Tool as c, type ActionDefinition as d, type ToolDefinition as e, type AgentLoopConfig as f, type StreamEvent as g, type ToolCallInfo as h, type TokenUsageRaw as i, type ToolResponse as j, type LanguageModel as k, type ModelCapabilities as l, type DoGenerateParams as m, type DoGenerateResult as n, type SystemMessage as o, type AssistantMessage as p, type ToolMessage as q, type UserContentPart as r, type TextPart as s, type ToolCall$1 as t, type ToolResult as u, type GenerateStep as v, type StreamPart as w, type StreamChunk as x, type TextDeltaChunk as y, type ToolCallChunk as z };
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@yourgpt/llm-sdk",
3
- "version": "2.0.1",
3
+ "version": "2.0.2-beta.1",
4
4
  "description": "AI SDK for building AI Agents with any LLM",
5
5
  "main": "./dist/index.js",
6
6
  "module": "./dist/index.mjs",