weave-typescript 0.19.0 → 0.20.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,588 @@
1
+ import { BinaryReader, BinaryWriter } from "@bufbuild/protobuf/wire";
2
+ import { ProviderKind } from "../../provider/v1/provider.pb";
3
+ export declare const protobufPackage = "weaveapi.inference.v1";
4
+ /** InferenceOperationType identifies one canonical execution capability. */
5
+ export declare enum InferenceOperationType {
6
+ INFERENCE_OPERATION_TYPE_UNSPECIFIED = 0,
7
+ INFERENCE_OPERATION_TYPE_CHAT_COMPLETION = 1,
8
+ INFERENCE_OPERATION_TYPE_STREAM_CHAT_COMPLETION = 2,
9
+ INFERENCE_OPERATION_TYPE_STRUCTURED_OUTPUT = 3,
10
+ INFERENCE_OPERATION_TYPE_TOOL_CALLING = 4,
11
+ INFERENCE_OPERATION_TYPE_EMBEDDINGS = 5,
12
+ INFERENCE_OPERATION_TYPE_IMAGE_GENERATION = 6,
13
+ INFERENCE_OPERATION_TYPE_MCP_CALL = 7,
14
+ UNRECOGNIZED = -1
15
+ }
16
+ export declare function inferenceOperationTypeFromJSON(object: any): InferenceOperationType;
17
+ export declare function inferenceOperationTypeToJSON(object: InferenceOperationType): string;
18
+ /** InferenceMessageRole identifies one participant role in a canonical message. */
19
+ export declare enum InferenceMessageRole {
20
+ INFERENCE_MESSAGE_ROLE_UNSPECIFIED = 0,
21
+ INFERENCE_MESSAGE_ROLE_SYSTEM = 1,
22
+ INFERENCE_MESSAGE_ROLE_USER = 2,
23
+ INFERENCE_MESSAGE_ROLE_ASSISTANT = 3,
24
+ INFERENCE_MESSAGE_ROLE_TOOL = 4,
25
+ INFERENCE_MESSAGE_ROLE_DEVELOPER = 5,
26
+ UNRECOGNIZED = -1
27
+ }
28
+ export declare function inferenceMessageRoleFromJSON(object: any): InferenceMessageRole;
29
+ export declare function inferenceMessageRoleToJSON(object: InferenceMessageRole): string;
30
+ /** InferenceFinishReason describes why one inference operation ended. */
31
+ export declare enum InferenceFinishReason {
32
+ INFERENCE_FINISH_REASON_UNSPECIFIED = 0,
33
+ INFERENCE_FINISH_REASON_STOP = 1,
34
+ INFERENCE_FINISH_REASON_LENGTH = 2,
35
+ INFERENCE_FINISH_REASON_TOOL_CALLS = 3,
36
+ INFERENCE_FINISH_REASON_CONTENT_FILTER = 4,
37
+ INFERENCE_FINISH_REASON_ERROR = 5,
38
+ UNRECOGNIZED = -1
39
+ }
40
+ export declare function inferenceFinishReasonFromJSON(object: any): InferenceFinishReason;
41
+ export declare function inferenceFinishReasonToJSON(object: InferenceFinishReason): string;
42
+ /** ToolChoiceMode controls how tool use should be resolved. */
43
+ export declare enum ToolChoiceMode {
44
+ TOOL_CHOICE_MODE_UNSPECIFIED = 0,
45
+ TOOL_CHOICE_MODE_AUTO = 1,
46
+ TOOL_CHOICE_MODE_NONE = 2,
47
+ TOOL_CHOICE_MODE_REQUIRED = 3,
48
+ UNRECOGNIZED = -1
49
+ }
50
+ export declare function toolChoiceModeFromJSON(object: any): ToolChoiceMode;
51
+ export declare function toolChoiceModeToJSON(object: ToolChoiceMode): string;
52
+ /** ImageDeliveryMode describes how generated images are returned. */
53
+ export declare enum ImageDeliveryMode {
54
+ IMAGE_DELIVERY_MODE_UNSPECIFIED = 0,
55
+ IMAGE_DELIVERY_MODE_INLINE_BYTES = 1,
56
+ IMAGE_DELIVERY_MODE_URL = 2,
57
+ IMAGE_DELIVERY_MODE_BASE64 = 3,
58
+ UNRECOGNIZED = -1
59
+ }
60
+ export declare function imageDeliveryModeFromJSON(object: any): ImageDeliveryMode;
61
+ export declare function imageDeliveryModeToJSON(object: ImageDeliveryMode): string;
62
+ /** InferenceTokenKind classifies one token bucket for usage accounting. */
63
+ export declare enum InferenceTokenKind {
64
+ INFERENCE_TOKEN_KIND_UNSPECIFIED = 0,
65
+ INFERENCE_TOKEN_KIND_INPUT = 1,
66
+ INFERENCE_TOKEN_KIND_INPUT_CACHE_WRITE = 2,
67
+ INFERENCE_TOKEN_KIND_INPUT_CACHE_READ = 3,
68
+ INFERENCE_TOKEN_KIND_INPUT_AUDIO = 4,
69
+ INFERENCE_TOKEN_KIND_INPUT_IMAGE = 5,
70
+ INFERENCE_TOKEN_KIND_OUTPUT = 6,
71
+ INFERENCE_TOKEN_KIND_OUTPUT_REASONING = 7,
72
+ INFERENCE_TOKEN_KIND_OUTPUT_AUDIO = 8,
73
+ UNRECOGNIZED = -1
74
+ }
75
+ export declare function inferenceTokenKindFromJSON(object: any): InferenceTokenKind;
76
+ export declare function inferenceTokenKindToJSON(object: InferenceTokenKind): string;
77
+ /** InferenceRequestContext contains caller-supplied routing hints. */
78
+ export interface InferenceRequestContext {
79
+ organizationId: string;
80
+ approvedModelId: string;
81
+ idempotencyKey: string;
82
+ metadata: {
83
+ [key: string]: any;
84
+ } | undefined;
85
+ }
86
+ /** ResolvedInferenceContext describes the execution route chosen by Weave. */
87
+ export interface ResolvedInferenceContext {
88
+ organizationId: string;
89
+ approvedModelId: string;
90
+ providerConfigurationId: string;
91
+ providerKind: ProviderKind;
92
+ resolvedModel: string;
93
+ operationType: InferenceOperationType;
94
+ requestId: string;
95
+ upstreamRequestId: string;
96
+ requestedAt: Date | undefined;
97
+ }
98
+ /** InferenceTokenUsageBreakdown is one normalized token meter bucket. */
99
+ export interface InferenceTokenUsageBreakdown {
100
+ kind: InferenceTokenKind;
101
+ tokenCount: number;
102
+ providerMetricName: string;
103
+ }
104
+ /** InferenceUsage summarizes token accounting. */
105
+ export interface InferenceUsage {
106
+ inputTokens: number;
107
+ outputTokens: number;
108
+ totalTokens: number;
109
+ tokenBreakdown: InferenceTokenUsageBreakdown[];
110
+ providerUsageRaw: {
111
+ [key: string]: any;
112
+ } | undefined;
113
+ }
114
+ /** InferenceGenerationConfig captures common generation controls. */
115
+ export interface InferenceGenerationConfig {
116
+ temperature: number;
117
+ maxOutputTokens: number;
118
+ stopSequences: string[];
119
+ allowParallelToolCalls: boolean;
120
+ topP: number;
121
+ topK: number;
122
+ seed: number;
123
+ presencePenalty: number;
124
+ frequencyPenalty: number;
125
+ metadata: {
126
+ [key: string]: any;
127
+ } | undefined;
128
+ }
129
+ /** StructuredOutputSchema describes one JSON-schema output contract. */
130
+ export interface StructuredOutputSchema {
131
+ schemaName: string;
132
+ jsonSchema: {
133
+ [key: string]: any;
134
+ } | undefined;
135
+ strict: boolean;
136
+ }
137
+ /** ToolDefinition describes one callable tool. */
138
+ export interface ToolDefinition {
139
+ name: string;
140
+ description: string;
141
+ inputSchema: {
142
+ [key: string]: any;
143
+ } | undefined;
144
+ }
145
+ /** ToolCall is one tool invocation requested by the model. */
146
+ export interface ToolCall {
147
+ id: string;
148
+ name: string;
149
+ arguments: {
150
+ [key: string]: any;
151
+ } | undefined;
152
+ }
153
+ /** ToolResult is one tool result returned to the model. */
154
+ export interface ToolResult {
155
+ toolCallId: string;
156
+ name: string;
157
+ result: {
158
+ [key: string]: any;
159
+ } | undefined;
160
+ isError: boolean;
161
+ }
162
+ /** InferenceContentPart is one canonical content fragment. */
163
+ export interface InferenceContentPart {
164
+ text?: string | undefined;
165
+ imageUrl?: string | undefined;
166
+ binaryData?: Uint8Array | undefined;
167
+ jsonValue?: {
168
+ [key: string]: any;
169
+ } | undefined;
170
+ toolCall?: ToolCall | undefined;
171
+ toolResult?: ToolResult | undefined;
172
+ mimeType: string;
173
+ }
174
+ /** ChatMessage is one canonical input or output message. */
175
+ export interface ChatMessage {
176
+ role: InferenceMessageRole;
177
+ content: InferenceContentPart[];
178
+ participantName: string;
179
+ toolCallId: string;
180
+ }
181
+ /** EmbeddingInput is one embedding payload. */
182
+ export interface EmbeddingInput {
183
+ text?: string | undefined;
184
+ binaryData?: Uint8Array | undefined;
185
+ mimeType: string;
186
+ }
187
+ /** EmbeddingVector is one returned embedding. */
188
+ export interface EmbeddingVector {
189
+ index: number;
190
+ values: number[];
191
+ }
192
+ /** GeneratedImage is one image returned from image generation. */
193
+ export interface GeneratedImage {
194
+ deliveryMode: ImageDeliveryMode;
195
+ url: string;
196
+ imageBytes: Uint8Array;
197
+ base64Data: string;
198
+ mimeType: string;
199
+ revisedPrompt: string;
200
+ }
201
+ export interface CreateChatCompletionRequest {
202
+ context: InferenceRequestContext | undefined;
203
+ model: string;
204
+ messages: ChatMessage[];
205
+ systemInstruction: string;
206
+ tools: ToolDefinition[];
207
+ toolChoiceMode: ToolChoiceMode;
208
+ generationConfig: InferenceGenerationConfig | undefined;
209
+ responseSchema: StructuredOutputSchema | undefined;
210
+ }
211
+ export interface CreateChatCompletionResponse {
212
+ context: ResolvedInferenceContext | undefined;
213
+ message: ChatMessage | undefined;
214
+ toolCalls: ToolCall[];
215
+ finishReason: InferenceFinishReason;
216
+ usage: InferenceUsage | undefined;
217
+ structuredOutput: {
218
+ [key: string]: any;
219
+ } | undefined;
220
+ }
221
+ export interface StreamChatCompletionRequest {
222
+ context: InferenceRequestContext | undefined;
223
+ model: string;
224
+ messages: ChatMessage[];
225
+ systemInstruction: string;
226
+ tools: ToolDefinition[];
227
+ toolChoiceMode: ToolChoiceMode;
228
+ generationConfig: InferenceGenerationConfig | undefined;
229
+ responseSchema: StructuredOutputSchema | undefined;
230
+ }
231
+ export interface StreamChatCompletionResponse {
232
+ context: ResolvedInferenceContext | undefined;
233
+ deltaText: string;
234
+ deltaMessage: ChatMessage | undefined;
235
+ toolCallDelta: ToolCall | undefined;
236
+ finishReason: InferenceFinishReason;
237
+ usage: InferenceUsage | undefined;
238
+ done: boolean;
239
+ structuredOutput: {
240
+ [key: string]: any;
241
+ } | undefined;
242
+ }
243
+ export interface CallToolsRequest {
244
+ context: InferenceRequestContext | undefined;
245
+ model: string;
246
+ messages: ChatMessage[];
247
+ systemInstruction: string;
248
+ tools: ToolDefinition[];
249
+ toolChoiceMode: ToolChoiceMode;
250
+ generationConfig: InferenceGenerationConfig | undefined;
251
+ }
252
+ export interface CallToolsResponse {
253
+ context: ResolvedInferenceContext | undefined;
254
+ message: ChatMessage | undefined;
255
+ toolCalls: ToolCall[];
256
+ finishReason: InferenceFinishReason;
257
+ usage: InferenceUsage | undefined;
258
+ }
259
+ export interface CreateEmbeddingsRequest {
260
+ context: InferenceRequestContext | undefined;
261
+ model: string;
262
+ inputs: EmbeddingInput[];
263
+ dimensions: number;
264
+ outputEncoding: string;
265
+ metadata: {
266
+ [key: string]: any;
267
+ } | undefined;
268
+ }
269
+ export interface CreateEmbeddingsResponse {
270
+ context: ResolvedInferenceContext | undefined;
271
+ embeddings: EmbeddingVector[];
272
+ usage: InferenceUsage | undefined;
273
+ }
274
+ export interface GenerateImagesRequest {
275
+ context: InferenceRequestContext | undefined;
276
+ model: string;
277
+ prompt: string;
278
+ negativePrompt: string;
279
+ imageCount: number;
280
+ size: string;
281
+ quality: string;
282
+ background: string;
283
+ deliveryMode: ImageDeliveryMode;
284
+ metadata: {
285
+ [key: string]: any;
286
+ } | undefined;
287
+ }
288
+ export interface GenerateImagesResponse {
289
+ context: ResolvedInferenceContext | undefined;
290
+ images: GeneratedImage[];
291
+ }
292
+ export interface CallMCPRequest {
293
+ context: InferenceRequestContext | undefined;
294
+ model: string;
295
+ serverName: string;
296
+ toolName: string;
297
+ arguments: {
298
+ [key: string]: any;
299
+ } | undefined;
300
+ metadata: {
301
+ [key: string]: any;
302
+ } | undefined;
303
+ }
304
+ export interface CallMCPResponse {
305
+ context: ResolvedInferenceContext | undefined;
306
+ result: {
307
+ [key: string]: any;
308
+ } | undefined;
309
+ }
310
+ export interface OpenAIChatMessage {
311
+ role: string;
312
+ content: {
313
+ [key: string]: any;
314
+ } | undefined;
315
+ name: string;
316
+ toolCallId: string;
317
+ toolCalls: OpenAIChatToolCall[];
318
+ }
319
+ export interface OpenAIFunctionDefinition {
320
+ name: string;
321
+ description: string;
322
+ parameters: {
323
+ [key: string]: any;
324
+ } | undefined;
325
+ strict: boolean;
326
+ }
327
+ export interface OpenAIToolDefinition {
328
+ type: string;
329
+ function: OpenAIFunctionDefinition | undefined;
330
+ }
331
+ export interface OpenAIResponseFormat {
332
+ type: string;
333
+ schemaName: string;
334
+ jsonSchema: {
335
+ [key: string]: any;
336
+ } | undefined;
337
+ strict: boolean;
338
+ }
339
+ export interface OpenAIFunctionCall {
340
+ name: string;
341
+ arguments: {
342
+ [key: string]: any;
343
+ } | undefined;
344
+ }
345
+ export interface OpenAIChatToolCall {
346
+ id: string;
347
+ type: string;
348
+ function: OpenAIFunctionCall | undefined;
349
+ }
350
+ export interface OpenAICompatibilityServiceCreateChatCompletionRequest {
351
+ model: string;
352
+ messages: OpenAIChatMessage[];
353
+ tools: OpenAIToolDefinition[];
354
+ toolChoice: string;
355
+ responseFormat: OpenAIResponseFormat | undefined;
356
+ temperature: number;
357
+ maxCompletionTokens: number;
358
+ stream: boolean;
359
+ stop: string[];
360
+ n: number;
361
+ metadata: {
362
+ [key: string]: any;
363
+ } | undefined;
364
+ }
365
+ export interface OpenAIChatChoice {
366
+ index: number;
367
+ message: OpenAIChatMessage | undefined;
368
+ finishReason: string;
369
+ }
370
+ export interface OpenAICompatibilityServiceCreateChatCompletionResponse {
371
+ id: string;
372
+ object: string;
373
+ model: string;
374
+ choices: OpenAIChatChoice[];
375
+ usage: InferenceUsage | undefined;
376
+ }
377
+ export interface OpenAIChatDeltaChoice {
378
+ index: number;
379
+ delta: OpenAIChatMessage | undefined;
380
+ finishReason: string;
381
+ }
382
+ export interface OpenAICompatibilityServiceStreamChatCompletionRequest {
383
+ model: string;
384
+ messages: OpenAIChatMessage[];
385
+ tools: OpenAIToolDefinition[];
386
+ toolChoice: string;
387
+ responseFormat: OpenAIResponseFormat | undefined;
388
+ temperature: number;
389
+ maxCompletionTokens: number;
390
+ stream: boolean;
391
+ stop: string[];
392
+ n: number;
393
+ metadata: {
394
+ [key: string]: any;
395
+ } | undefined;
396
+ }
397
+ export interface OpenAICompatibilityServiceStreamChatCompletionResponse {
398
+ id: string;
399
+ object: string;
400
+ model: string;
401
+ choices: OpenAIChatDeltaChoice[];
402
+ usage: InferenceUsage | undefined;
403
+ done: boolean;
404
+ }
405
+ export interface OpenAICompatibilityServiceCreateEmbeddingRequest {
406
+ model: string;
407
+ input: string[];
408
+ dimensions: number;
409
+ encodingFormat: string;
410
+ user: string;
411
+ }
412
+ export interface OpenAIEmbeddingData {
413
+ object: string;
414
+ index: number;
415
+ embedding: number[];
416
+ }
417
+ export interface OpenAICompatibilityServiceCreateEmbeddingResponse {
418
+ object: string;
419
+ data: OpenAIEmbeddingData[];
420
+ model: string;
421
+ usage: InferenceUsage | undefined;
422
+ }
423
+ export interface OpenAICompatibilityServiceGenerateImageRequest {
424
+ model: string;
425
+ prompt: string;
426
+ n: number;
427
+ size: string;
428
+ quality: string;
429
+ background: string;
430
+ responseFormat: string;
431
+ user: string;
432
+ }
433
+ export interface OpenAIImageData {
434
+ url: string;
435
+ b64Json: string;
436
+ revisedPrompt: string;
437
+ }
438
+ export interface OpenAICompatibilityServiceGenerateImageResponse {
439
+ created: number;
440
+ data: OpenAIImageData[];
441
+ }
442
+ export interface AnthropicMessage {
443
+ role: string;
444
+ content: {
445
+ [key: string]: any;
446
+ } | undefined;
447
+ }
448
+ export interface AnthropicToolDefinition {
449
+ name: string;
450
+ description: string;
451
+ inputSchema: {
452
+ [key: string]: any;
453
+ } | undefined;
454
+ }
455
+ export interface AnthropicCompatibilityServiceCreateMessageRequest {
456
+ model: string;
457
+ messages: AnthropicMessage[];
458
+ system: string;
459
+ maxTokens: number;
460
+ temperature: number;
461
+ stream: boolean;
462
+ tools: AnthropicToolDefinition[];
463
+ metadata: {
464
+ [key: string]: any;
465
+ } | undefined;
466
+ }
467
+ export interface AnthropicContentBlock {
468
+ type: string;
469
+ text: string;
470
+ id: string;
471
+ name: string;
472
+ input: {
473
+ [key: string]: any;
474
+ } | undefined;
475
+ toolUseId: string;
476
+ content: {
477
+ [key: string]: any;
478
+ } | undefined;
479
+ isError: boolean;
480
+ }
481
+ export interface AnthropicUsage {
482
+ inputTokens: number;
483
+ outputTokens: number;
484
+ cacheCreationInputTokens: number;
485
+ cacheReadInputTokens: number;
486
+ }
487
+ export interface AnthropicCompatibilityServiceCreateMessageResponse {
488
+ id: string;
489
+ type: string;
490
+ role: string;
491
+ content: AnthropicContentBlock[];
492
+ stopReason: string;
493
+ stopSequence: string;
494
+ usage: AnthropicUsage | undefined;
495
+ model: string;
496
+ }
497
+ export interface AnthropicCompatibilityServiceStreamMessagesRequest {
498
+ model: string;
499
+ messages: AnthropicMessage[];
500
+ system: string;
501
+ maxTokens: number;
502
+ temperature: number;
503
+ stream: boolean;
504
+ tools: AnthropicToolDefinition[];
505
+ metadata: {
506
+ [key: string]: any;
507
+ } | undefined;
508
+ }
509
+ export interface AnthropicCompatibilityServiceStreamMessagesResponse {
510
+ type: string;
511
+ messageId: string;
512
+ delta: {
513
+ [key: string]: any;
514
+ } | undefined;
515
+ usage: AnthropicUsage | undefined;
516
+ done: boolean;
517
+ }
518
+ export declare const InferenceRequestContext: MessageFns<InferenceRequestContext>;
519
+ export declare const ResolvedInferenceContext: MessageFns<ResolvedInferenceContext>;
520
+ export declare const InferenceTokenUsageBreakdown: MessageFns<InferenceTokenUsageBreakdown>;
521
+ export declare const InferenceUsage: MessageFns<InferenceUsage>;
522
+ export declare const InferenceGenerationConfig: MessageFns<InferenceGenerationConfig>;
523
+ export declare const StructuredOutputSchema: MessageFns<StructuredOutputSchema>;
524
+ export declare const ToolDefinition: MessageFns<ToolDefinition>;
525
+ export declare const ToolCall: MessageFns<ToolCall>;
526
+ export declare const ToolResult: MessageFns<ToolResult>;
527
+ export declare const InferenceContentPart: MessageFns<InferenceContentPart>;
528
+ export declare const ChatMessage: MessageFns<ChatMessage>;
529
+ export declare const EmbeddingInput: MessageFns<EmbeddingInput>;
530
+ export declare const EmbeddingVector: MessageFns<EmbeddingVector>;
531
+ export declare const GeneratedImage: MessageFns<GeneratedImage>;
532
+ export declare const CreateChatCompletionRequest: MessageFns<CreateChatCompletionRequest>;
533
+ export declare const CreateChatCompletionResponse: MessageFns<CreateChatCompletionResponse>;
534
+ export declare const StreamChatCompletionRequest: MessageFns<StreamChatCompletionRequest>;
535
+ export declare const StreamChatCompletionResponse: MessageFns<StreamChatCompletionResponse>;
536
+ export declare const CallToolsRequest: MessageFns<CallToolsRequest>;
537
+ export declare const CallToolsResponse: MessageFns<CallToolsResponse>;
538
+ export declare const CreateEmbeddingsRequest: MessageFns<CreateEmbeddingsRequest>;
539
+ export declare const CreateEmbeddingsResponse: MessageFns<CreateEmbeddingsResponse>;
540
+ export declare const GenerateImagesRequest: MessageFns<GenerateImagesRequest>;
541
+ export declare const GenerateImagesResponse: MessageFns<GenerateImagesResponse>;
542
+ export declare const CallMCPRequest: MessageFns<CallMCPRequest>;
543
+ export declare const CallMCPResponse: MessageFns<CallMCPResponse>;
544
+ export declare const OpenAIChatMessage: MessageFns<OpenAIChatMessage>;
545
+ export declare const OpenAIFunctionDefinition: MessageFns<OpenAIFunctionDefinition>;
546
+ export declare const OpenAIToolDefinition: MessageFns<OpenAIToolDefinition>;
547
+ export declare const OpenAIResponseFormat: MessageFns<OpenAIResponseFormat>;
548
+ export declare const OpenAIFunctionCall: MessageFns<OpenAIFunctionCall>;
549
+ export declare const OpenAIChatToolCall: MessageFns<OpenAIChatToolCall>;
550
+ export declare const OpenAICompatibilityServiceCreateChatCompletionRequest: MessageFns<OpenAICompatibilityServiceCreateChatCompletionRequest>;
551
+ export declare const OpenAIChatChoice: MessageFns<OpenAIChatChoice>;
552
+ export declare const OpenAICompatibilityServiceCreateChatCompletionResponse: MessageFns<OpenAICompatibilityServiceCreateChatCompletionResponse>;
553
+ export declare const OpenAIChatDeltaChoice: MessageFns<OpenAIChatDeltaChoice>;
554
+ export declare const OpenAICompatibilityServiceStreamChatCompletionRequest: MessageFns<OpenAICompatibilityServiceStreamChatCompletionRequest>;
555
+ export declare const OpenAICompatibilityServiceStreamChatCompletionResponse: MessageFns<OpenAICompatibilityServiceStreamChatCompletionResponse>;
556
+ export declare const OpenAICompatibilityServiceCreateEmbeddingRequest: MessageFns<OpenAICompatibilityServiceCreateEmbeddingRequest>;
557
+ export declare const OpenAIEmbeddingData: MessageFns<OpenAIEmbeddingData>;
558
+ export declare const OpenAICompatibilityServiceCreateEmbeddingResponse: MessageFns<OpenAICompatibilityServiceCreateEmbeddingResponse>;
559
+ export declare const OpenAICompatibilityServiceGenerateImageRequest: MessageFns<OpenAICompatibilityServiceGenerateImageRequest>;
560
+ export declare const OpenAIImageData: MessageFns<OpenAIImageData>;
561
+ export declare const OpenAICompatibilityServiceGenerateImageResponse: MessageFns<OpenAICompatibilityServiceGenerateImageResponse>;
562
+ export declare const AnthropicMessage: MessageFns<AnthropicMessage>;
563
+ export declare const AnthropicToolDefinition: MessageFns<AnthropicToolDefinition>;
564
+ export declare const AnthropicCompatibilityServiceCreateMessageRequest: MessageFns<AnthropicCompatibilityServiceCreateMessageRequest>;
565
+ export declare const AnthropicContentBlock: MessageFns<AnthropicContentBlock>;
566
+ export declare const AnthropicUsage: MessageFns<AnthropicUsage>;
567
+ export declare const AnthropicCompatibilityServiceCreateMessageResponse: MessageFns<AnthropicCompatibilityServiceCreateMessageResponse>;
568
+ export declare const AnthropicCompatibilityServiceStreamMessagesRequest: MessageFns<AnthropicCompatibilityServiceStreamMessagesRequest>;
569
+ export declare const AnthropicCompatibilityServiceStreamMessagesResponse: MessageFns<AnthropicCompatibilityServiceStreamMessagesResponse>;
570
+ type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined;
571
+ export type DeepPartial<T> = T extends Builtin ? T : T extends globalThis.Array<infer U> ? globalThis.Array<DeepPartial<U>> : T extends ReadonlyArray<infer U> ? ReadonlyArray<DeepPartial<U>> : T extends {} ? {
572
+ [K in keyof T]?: DeepPartial<T[K]>;
573
+ } : Partial<T>;
574
+ type KeysOfUnion<T> = T extends T ? keyof T : never;
575
+ export type Exact<P, I extends P> = P extends Builtin ? P : P & {
576
+ [K in keyof P]: Exact<P[K], I[K]>;
577
+ } & {
578
+ [K in Exclude<keyof I, KeysOfUnion<P>>]: never;
579
+ };
580
+ export interface MessageFns<T> {
581
+ encode(message: T, writer?: BinaryWriter): BinaryWriter;
582
+ decode(input: BinaryReader | Uint8Array, length?: number): T;
583
+ fromJSON(object: any): T;
584
+ toJSON(message: T): unknown;
585
+ create<I extends Exact<DeepPartial<T>, I>>(base?: I): T;
586
+ fromPartial<I extends Exact<DeepPartial<T>, I>>(object: I): T;
587
+ }
588
+ export {};