vectorvein 0.2.7__py3-none-any.whl → 0.2.9__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4,10 +4,19 @@ import asyncio
4
4
  from abc import ABC, abstractmethod
5
5
  from collections import defaultdict
6
6
  from functools import cached_property
7
- from typing import Generator, AsyncGenerator, Any, overload, Literal, Iterable
7
+ from typing import Generator, AsyncGenerator, Any, overload, Literal, Iterable, Optional, Dict, List, Union
8
8
 
9
9
  import httpx
10
10
  from openai import OpenAI, AsyncOpenAI, AzureOpenAI, AsyncAzureOpenAI
11
+ from openai._types import Headers, Query, Body
12
+ from openai.types.shared_params.metadata import Metadata
13
+ from openai.types.chat.completion_create_params import ResponseFormat
14
+ from openai.types.chat.chat_completion_modality import ChatCompletionModality
15
+ from openai.types.chat.chat_completion_audio_param import ChatCompletionAudioParam
16
+ from openai.types.chat.chat_completion_reasoning_effort import ChatCompletionReasoningEffort
17
+ from openai.types.chat.chat_completion_stream_options_param import ChatCompletionStreamOptionsParam
18
+ from openai.types.chat.chat_completion_prediction_content_param import ChatCompletionPredictionContentParam
19
+
11
20
  from anthropic import (
12
21
  Anthropic,
13
22
  AsyncAnthropic,
@@ -16,6 +25,7 @@ from anthropic import (
16
25
  AnthropicBedrock,
17
26
  AsyncAnthropicBedrock,
18
27
  )
28
+ from anthropic.types.thinking_config_param import ThinkingConfigParam
19
29
 
20
30
  from ..settings import settings
21
31
  from ..types import defaults as defs
@@ -25,10 +35,10 @@ from ..types.llm_parameters import (
25
35
  NOT_GIVEN,
26
36
  ToolParam,
27
37
  ToolChoice,
38
+ OpenAINotGiven,
28
39
  EndpointSetting,
29
40
  ChatCompletionMessage,
30
41
  ChatCompletionDeltaMessage,
31
- ChatCompletionStreamOptionsParam,
32
42
  )
33
43
  from ..utilities.rate_limiter import SyncMemoryRateLimiter, SyncRedisRateLimiter, SyncDiskCacheRateLimiter
34
44
  from ..utilities.rate_limiter import AsyncMemoryRateLimiter, AsyncRedisRateLimiter, AsyncDiskCacheRateLimiter
@@ -172,11 +182,33 @@ class BaseChatClient(ABC):
172
182
  max_tokens: int | None = None,
173
183
  tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
174
184
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
175
- response_format: dict | None = None,
185
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
176
186
  stream_options: ChatCompletionStreamOptionsParam | None = None,
177
187
  top_p: float | NotGiven | None = NOT_GIVEN,
178
188
  skip_cutoff: bool = False,
179
- **kwargs,
189
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
190
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
191
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
192
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
193
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
194
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
195
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
196
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
197
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
198
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
199
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
200
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
201
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
202
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
203
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
204
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
205
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
206
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
207
+ user: str | OpenAINotGiven = NOT_GIVEN,
208
+ extra_headers: Headers | None = None,
209
+ extra_query: Query | None = None,
210
+ extra_body: Body | None = None,
211
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
180
212
  ) -> ChatCompletionMessage:
181
213
  pass
182
214
 
@@ -192,11 +224,33 @@ class BaseChatClient(ABC):
192
224
  max_tokens: int | None = None,
193
225
  tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
194
226
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
195
- response_format: dict | None = None,
227
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
196
228
  stream_options: ChatCompletionStreamOptionsParam | None = None,
197
229
  top_p: float | NotGiven | None = NOT_GIVEN,
198
230
  skip_cutoff: bool = False,
199
- **kwargs,
231
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
232
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
233
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
234
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
235
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
236
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
237
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
238
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
239
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
240
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
241
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
242
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
243
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
244
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
245
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
246
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
247
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
248
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
249
+ user: str | OpenAINotGiven = NOT_GIVEN,
250
+ extra_headers: Headers | None = None,
251
+ extra_query: Query | None = None,
252
+ extra_body: Body | None = None,
253
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
200
254
  ) -> Generator[ChatCompletionDeltaMessage, Any, None]:
201
255
  pass
202
256
 
@@ -212,11 +266,33 @@ class BaseChatClient(ABC):
212
266
  max_tokens: int | None = None,
213
267
  tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
214
268
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
215
- response_format: dict | None = None,
269
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
216
270
  stream_options: ChatCompletionStreamOptionsParam | None = None,
217
271
  top_p: float | NotGiven | None = NOT_GIVEN,
218
272
  skip_cutoff: bool = False,
219
- **kwargs,
273
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
274
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
275
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
276
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
277
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
278
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
279
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
280
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
281
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
282
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
283
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
284
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
285
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
286
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
287
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
288
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
289
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
290
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
291
+ user: str | OpenAINotGiven = NOT_GIVEN,
292
+ extra_headers: Headers | None = None,
293
+ extra_query: Query | None = None,
294
+ extra_body: Body | None = None,
295
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
220
296
  ) -> ChatCompletionMessage | Generator[ChatCompletionDeltaMessage, Any, None]:
221
297
  pass
222
298
 
@@ -231,11 +307,33 @@ class BaseChatClient(ABC):
231
307
  max_tokens: int | None = None,
232
308
  tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
233
309
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
234
- response_format: dict | None = None,
310
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
235
311
  stream_options: ChatCompletionStreamOptionsParam | None = None,
236
312
  top_p: float | NotGiven | None = NOT_GIVEN,
237
313
  skip_cutoff: bool = False,
238
- **kwargs,
314
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
315
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
316
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
317
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
318
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
319
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
320
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
321
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
322
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
323
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
324
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
325
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
326
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
327
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
328
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
329
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
330
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
331
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
332
+ user: str | OpenAINotGiven = NOT_GIVEN,
333
+ extra_headers: Headers | None = None,
334
+ extra_query: Query | None = None,
335
+ extra_body: Body | None = None,
336
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
239
337
  ) -> ChatCompletionMessage | Generator[ChatCompletionDeltaMessage, Any, None]:
240
338
  pass
241
339
 
@@ -248,10 +346,33 @@ class BaseChatClient(ABC):
248
346
  max_tokens: int | None = None,
249
347
  tools: Iterable[ToolParam] | NotGiven = NOT_GIVEN,
250
348
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
251
- response_format: dict | None = None,
349
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
252
350
  stream_options: ChatCompletionStreamOptionsParam | None = None,
253
351
  top_p: float | NotGiven | None = NOT_GIVEN,
254
- **kwargs,
352
+ skip_cutoff: bool = False,
353
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
354
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
355
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
356
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
357
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
358
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
359
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
360
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
361
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
362
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
363
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
364
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
365
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
366
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
367
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
368
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
369
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
370
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
371
+ user: str | OpenAINotGiven = NOT_GIVEN,
372
+ extra_headers: Headers | None = None,
373
+ extra_query: Query | None = None,
374
+ extra_body: Body | None = None,
375
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
255
376
  ) -> Generator[ChatCompletionDeltaMessage, Any, None]:
256
377
  return self.create_completion(
257
378
  messages=messages,
@@ -264,7 +385,30 @@ class BaseChatClient(ABC):
264
385
  response_format=response_format,
265
386
  stream_options=stream_options,
266
387
  top_p=top_p,
267
- **kwargs,
388
+ skip_cutoff=skip_cutoff,
389
+ audio=audio,
390
+ frequency_penalty=frequency_penalty,
391
+ logit_bias=logit_bias,
392
+ logprobs=logprobs,
393
+ max_completion_tokens=max_completion_tokens,
394
+ metadata=metadata,
395
+ modalities=modalities,
396
+ n=n,
397
+ parallel_tool_calls=parallel_tool_calls,
398
+ prediction=prediction,
399
+ presence_penalty=presence_penalty,
400
+ reasoning_effort=reasoning_effort,
401
+ thinking=thinking,
402
+ seed=seed,
403
+ service_tier=service_tier,
404
+ stop=stop,
405
+ store=store,
406
+ top_logprobs=top_logprobs,
407
+ user=user,
408
+ extra_headers=extra_headers,
409
+ extra_query=extra_query,
410
+ extra_body=extra_body,
411
+ timeout=timeout,
268
412
  )
269
413
 
270
414
 
@@ -416,11 +560,33 @@ class BaseAsyncChatClient(ABC):
416
560
  max_tokens: int | None = None,
417
561
  tools: list | NotGiven = NOT_GIVEN,
418
562
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
419
- response_format: dict | None = None,
563
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
420
564
  stream_options: ChatCompletionStreamOptionsParam | None = None,
421
565
  top_p: float | NotGiven | None = NOT_GIVEN,
422
566
  skip_cutoff: bool = False,
423
- **kwargs,
567
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
568
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
569
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
570
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
571
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
572
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
573
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
574
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
575
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
576
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
577
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
578
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
579
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
580
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
581
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
582
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
583
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
584
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
585
+ user: str | OpenAINotGiven = NOT_GIVEN,
586
+ extra_headers: Headers | None = None,
587
+ extra_query: Query | None = None,
588
+ extra_body: Body | None = None,
589
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
424
590
  ) -> ChatCompletionMessage:
425
591
  pass
426
592
 
@@ -436,11 +602,33 @@ class BaseAsyncChatClient(ABC):
436
602
  max_tokens: int | None = None,
437
603
  tools: list | NotGiven = NOT_GIVEN,
438
604
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
439
- response_format: dict | None = None,
605
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
440
606
  stream_options: ChatCompletionStreamOptionsParam | None = None,
441
607
  top_p: float | NotGiven | None = NOT_GIVEN,
442
608
  skip_cutoff: bool = False,
443
- **kwargs,
609
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
610
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
611
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
612
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
613
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
614
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
615
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
616
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
617
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
618
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
619
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
620
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
621
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
622
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
623
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
624
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
625
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
626
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
627
+ user: str | OpenAINotGiven = NOT_GIVEN,
628
+ extra_headers: Headers | None = None,
629
+ extra_query: Query | None = None,
630
+ extra_body: Body | None = None,
631
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
444
632
  ) -> AsyncGenerator[ChatCompletionDeltaMessage, None]:
445
633
  pass
446
634
 
@@ -456,11 +644,33 @@ class BaseAsyncChatClient(ABC):
456
644
  max_tokens: int | None = None,
457
645
  tools: list | NotGiven = NOT_GIVEN,
458
646
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
459
- response_format: dict | None = None,
647
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
460
648
  stream_options: ChatCompletionStreamOptionsParam | None = None,
461
649
  top_p: float | NotGiven | None = NOT_GIVEN,
462
650
  skip_cutoff: bool = False,
463
- **kwargs,
651
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
652
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
653
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
654
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
655
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
656
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
657
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
658
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
659
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
660
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
661
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
662
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
663
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
664
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
665
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
666
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
667
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
668
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
669
+ user: str | OpenAINotGiven = NOT_GIVEN,
670
+ extra_headers: Headers | None = None,
671
+ extra_query: Query | None = None,
672
+ extra_body: Body | None = None,
673
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
464
674
  ) -> ChatCompletionMessage | AsyncGenerator[ChatCompletionDeltaMessage, None]:
465
675
  pass
466
676
 
@@ -475,11 +685,33 @@ class BaseAsyncChatClient(ABC):
475
685
  max_tokens: int | None = None,
476
686
  tools: list | NotGiven = NOT_GIVEN,
477
687
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
478
- response_format: dict | None = None,
688
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
479
689
  stream_options: ChatCompletionStreamOptionsParam | None = None,
480
690
  top_p: float | NotGiven | None = NOT_GIVEN,
481
691
  skip_cutoff: bool = False,
482
- **kwargs,
692
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
693
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
694
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
695
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
696
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
697
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
698
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
699
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
700
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
701
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
702
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
703
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
704
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
705
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
706
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
707
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
708
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
709
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
710
+ user: str | OpenAINotGiven = NOT_GIVEN,
711
+ extra_headers: Headers | None = None,
712
+ extra_query: Query | None = None,
713
+ extra_body: Body | None = None,
714
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
483
715
  ) -> ChatCompletionMessage | AsyncGenerator[ChatCompletionDeltaMessage, None]:
484
716
  pass
485
717
 
@@ -492,10 +724,33 @@ class BaseAsyncChatClient(ABC):
492
724
  max_tokens: int | None = None,
493
725
  tools: list | NotGiven = NOT_GIVEN,
494
726
  tool_choice: ToolChoice | NotGiven = NOT_GIVEN,
495
- response_format: dict | None = None,
727
+ response_format: ResponseFormat | NotGiven = NOT_GIVEN,
496
728
  stream_options: ChatCompletionStreamOptionsParam | None = None,
497
729
  top_p: float | NotGiven | None = NOT_GIVEN,
498
- **kwargs,
730
+ skip_cutoff: bool = False,
731
+ audio: Optional[ChatCompletionAudioParam] | OpenAINotGiven = NOT_GIVEN,
732
+ frequency_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
733
+ logit_bias: Optional[Dict[str, int]] | OpenAINotGiven = NOT_GIVEN,
734
+ logprobs: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
735
+ max_completion_tokens: Optional[int] | OpenAINotGiven = NOT_GIVEN,
736
+ metadata: Optional[Metadata] | OpenAINotGiven = NOT_GIVEN,
737
+ modalities: Optional[List[ChatCompletionModality]] | OpenAINotGiven = NOT_GIVEN,
738
+ n: Optional[int] | OpenAINotGiven = NOT_GIVEN,
739
+ parallel_tool_calls: bool | OpenAINotGiven = NOT_GIVEN,
740
+ prediction: Optional[ChatCompletionPredictionContentParam] | OpenAINotGiven = NOT_GIVEN,
741
+ presence_penalty: Optional[float] | OpenAINotGiven = NOT_GIVEN,
742
+ reasoning_effort: Optional[ChatCompletionReasoningEffort] | OpenAINotGiven = NOT_GIVEN,
743
+ thinking: ThinkingConfigParam | None | NotGiven = NOT_GIVEN,
744
+ seed: Optional[int] | OpenAINotGiven = NOT_GIVEN,
745
+ service_tier: Optional[Literal["auto", "default"]] | OpenAINotGiven = NOT_GIVEN,
746
+ stop: Union[Optional[str], List[str]] | OpenAINotGiven = NOT_GIVEN,
747
+ store: Optional[bool] | OpenAINotGiven = NOT_GIVEN,
748
+ top_logprobs: Optional[int] | OpenAINotGiven = NOT_GIVEN,
749
+ user: str | OpenAINotGiven = NOT_GIVEN,
750
+ extra_headers: Headers | None = None,
751
+ extra_query: Query | None = None,
752
+ extra_body: Body | None = None,
753
+ timeout: float | httpx.Timeout | None | OpenAINotGiven = NOT_GIVEN,
499
754
  ) -> AsyncGenerator[ChatCompletionDeltaMessage, None]:
500
755
  return await self.create_completion(
501
756
  messages=messages,
@@ -508,5 +763,28 @@ class BaseAsyncChatClient(ABC):
508
763
  response_format=response_format,
509
764
  stream_options=stream_options,
510
765
  top_p=top_p,
511
- **kwargs,
766
+ skip_cutoff=skip_cutoff,
767
+ audio=audio,
768
+ frequency_penalty=frequency_penalty,
769
+ logit_bias=logit_bias,
770
+ logprobs=logprobs,
771
+ max_completion_tokens=max_completion_tokens,
772
+ metadata=metadata,
773
+ modalities=modalities,
774
+ n=n,
775
+ parallel_tool_calls=parallel_tool_calls,
776
+ prediction=prediction,
777
+ presence_penalty=presence_penalty,
778
+ reasoning_effort=reasoning_effort,
779
+ thinking=thinking,
780
+ seed=seed,
781
+ service_tier=service_tier,
782
+ stop=stop,
783
+ store=store,
784
+ top_logprobs=top_logprobs,
785
+ user=user,
786
+ extra_headers=extra_headers,
787
+ extra_query=extra_query,
788
+ extra_body=extra_body,
789
+ timeout=timeout,
512
790
  )