xai-review 0.33.0__py3-none-any.whl → 0.35.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of xai-review might be problematic. Click here for more details.

Files changed (56) hide show
  1. ai_review/clients/claude/client.py +1 -1
  2. ai_review/clients/claude/schema.py +2 -2
  3. ai_review/clients/gemini/client.py +2 -1
  4. ai_review/clients/gemini/schema.py +2 -2
  5. ai_review/clients/ollama/client.py +1 -1
  6. ai_review/clients/openai/v1/__init__.py +0 -0
  7. ai_review/clients/openai/{client.py → v1/client.py} +9 -9
  8. ai_review/clients/openai/{schema.py → v1/schema.py} +1 -1
  9. ai_review/clients/openai/v1/types.py +8 -0
  10. ai_review/clients/openai/v2/__init__.py +0 -0
  11. ai_review/clients/openai/v2/client.py +46 -0
  12. ai_review/clients/openai/v2/schema.py +47 -0
  13. ai_review/clients/openai/v2/types.py +11 -0
  14. ai_review/clients/openrouter/__init__.py +0 -0
  15. ai_review/clients/openrouter/client.py +50 -0
  16. ai_review/clients/openrouter/schema.py +36 -0
  17. ai_review/clients/openrouter/types.py +11 -0
  18. ai_review/libs/config/llm/base.py +8 -1
  19. ai_review/libs/config/llm/meta.py +2 -2
  20. ai_review/libs/config/llm/openai.py +4 -0
  21. ai_review/libs/config/llm/openrouter.py +12 -0
  22. ai_review/libs/constants/llm_provider.py +1 -0
  23. ai_review/resources/pricing.yaml +39 -1
  24. ai_review/services/llm/factory.py +3 -0
  25. ai_review/services/llm/openai/client.py +37 -9
  26. ai_review/services/llm/openrouter/__init__.py +0 -0
  27. ai_review/services/llm/openrouter/client.py +31 -0
  28. ai_review/tests/fixtures/clients/openai.py +84 -12
  29. ai_review/tests/fixtures/clients/openrouter.py +72 -0
  30. ai_review/tests/suites/clients/openai/v1/__init__.py +0 -0
  31. ai_review/tests/suites/clients/openai/v1/test_client.py +12 -0
  32. ai_review/tests/suites/clients/openai/{test_schema.py → v1/test_schema.py} +1 -1
  33. ai_review/tests/suites/clients/openai/v2/__init__.py +0 -0
  34. ai_review/tests/suites/clients/openai/v2/test_client.py +12 -0
  35. ai_review/tests/suites/clients/openai/v2/test_schema.py +80 -0
  36. ai_review/tests/suites/clients/openrouter/__init__.py +0 -0
  37. ai_review/tests/suites/clients/openrouter/test_client.py +12 -0
  38. ai_review/tests/suites/clients/openrouter/test_schema.py +57 -0
  39. ai_review/tests/suites/libs/config/llm/__init__.py +0 -0
  40. ai_review/tests/suites/libs/config/llm/test_openai.py +28 -0
  41. ai_review/tests/suites/services/llm/openai/test_client.py +23 -6
  42. ai_review/tests/suites/services/llm/openrouter/__init__.py +0 -0
  43. ai_review/tests/suites/services/llm/openrouter/test_client.py +22 -0
  44. ai_review/tests/suites/services/llm/test_factory.py +8 -1
  45. ai_review/tests/suites/services/review/gateway/test_review_comment_gateway.py +5 -5
  46. ai_review/tests/suites/services/review/gateway/test_review_dry_run_comment_gateway.py +5 -5
  47. ai_review/tests/suites/services/review/gateway/test_review_llm_gateway.py +2 -2
  48. ai_review/tests/suites/services/review/test_service.py +2 -2
  49. {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/METADATA +11 -10
  50. {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/RECORD +54 -30
  51. ai_review/clients/openai/types.py +0 -8
  52. ai_review/tests/suites/clients/openai/test_client.py +0 -12
  53. {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/WHEEL +0 -0
  54. {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/entry_points.txt +0 -0
  55. {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/licenses/LICENSE +0 -0
  56. {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/top_level.txt +0 -0
@@ -17,7 +17,7 @@ class ClaudeHTTPClientError(HTTPClientError):
17
17
  class ClaudeHTTPClient(HTTPClient, ClaudeHTTPClientProtocol):
18
18
  @handle_http_error(client="ClaudeHTTPClient", exception=ClaudeHTTPClientError)
19
19
  async def chat_api(self, request: ClaudeChatRequestSchema) -> Response:
20
- return await self.post("/v1/messages", json=request.model_dump())
20
+ return await self.post("/v1/messages", json=request.model_dump(exclude_none=True))
21
21
 
22
22
  async def chat(self, request: ClaudeChatRequestSchema) -> ClaudeChatResponseSchema:
23
23
  response = await self.chat_api(request)
@@ -12,8 +12,8 @@ class ClaudeChatRequestSchema(BaseModel):
12
12
  model: str
13
13
  system: str | None = None
14
14
  messages: list[ClaudeMessageSchema]
15
- max_tokens: int
16
- temperature: float
15
+ max_tokens: int | None = None
16
+ temperature: float | None = None
17
17
 
18
18
 
19
19
  class ClaudeContentSchema(BaseModel):
@@ -19,7 +19,8 @@ class GeminiHTTPClient(HTTPClient, GeminiHTTPClientProtocol):
19
19
  async def chat_api(self, request: GeminiChatRequestSchema) -> Response:
20
20
  meta = settings.llm.meta
21
21
  return await self.post(
22
- f"/v1beta/models/{meta.model}:generateContent", json=request.model_dump()
22
+ f"/v1beta/models/{meta.model}:generateContent",
23
+ json=request.model_dump(exclude_none=True)
23
24
  )
24
25
 
25
26
  async def chat(self, request: GeminiChatRequestSchema) -> GeminiChatResponseSchema:
@@ -45,8 +45,8 @@ class GeminiCandidateSchema(BaseModel):
45
45
  class GeminiGenerationConfigSchema(BaseModel):
46
46
  model_config = ConfigDict(populate_by_name=True)
47
47
 
48
- temperature: float
49
- max_output_tokens: int = Field(alias="maxOutputTokens")
48
+ temperature: float | None = None
49
+ max_output_tokens: int | None = Field(alias="maxOutputTokens", default=None)
50
50
 
51
51
 
52
52
  class GeminiChatRequestSchema(BaseModel):
@@ -17,7 +17,7 @@ class OllamaHTTPClientError(HTTPClientError):
17
17
  class OllamaHTTPClient(HTTPClient, OllamaHTTPClientProtocol):
18
18
  @handle_http_error(client="OllamaHTTPClient", exception=OllamaHTTPClientError)
19
19
  async def chat_api(self, request: OllamaChatRequestSchema) -> Response:
20
- return await self.post("/api/chat", json=request.model_dump())
20
+ return await self.post("/api/chat", json=request.model_dump(exclude_none=True))
21
21
 
22
22
  async def chat(self, request: OllamaChatRequestSchema) -> OllamaChatResponseSchema:
23
23
  response = await self.chat_api(request)
File without changes
@@ -1,7 +1,7 @@
1
1
  from httpx import Response, AsyncHTTPTransport, AsyncClient
2
2
 
3
- from ai_review.clients.openai.schema import OpenAIChatRequestSchema, OpenAIChatResponseSchema
4
- from ai_review.clients.openai.types import OpenAIHTTPClientProtocol
3
+ from ai_review.clients.openai.v1.schema import OpenAIChatRequestSchema, OpenAIChatResponseSchema
4
+ from ai_review.clients.openai.v1.types import OpenAIV1HTTPClientProtocol
5
5
  from ai_review.config import settings
6
6
  from ai_review.libs.http.client import HTTPClient
7
7
  from ai_review.libs.http.event_hooks.logger import LoggerEventHook
@@ -10,22 +10,22 @@ from ai_review.libs.http.transports.retry import RetryTransport
10
10
  from ai_review.libs.logger import get_logger
11
11
 
12
12
 
13
- class OpenAIHTTPClientError(HTTPClientError):
13
+ class OpenAIV1HTTPClientError(HTTPClientError):
14
14
  pass
15
15
 
16
16
 
17
- class OpenAIHTTPClient(HTTPClient, OpenAIHTTPClientProtocol):
18
- @handle_http_error(client='OpenAIHTTPClient', exception=OpenAIHTTPClientError)
17
+ class OpenAIV1HTTPClient(HTTPClient, OpenAIV1HTTPClientProtocol):
18
+ @handle_http_error(client='OpenAIV1HTTPClient', exception=OpenAIV1HTTPClientError)
19
19
  async def chat_api(self, request: OpenAIChatRequestSchema) -> Response:
20
- return await self.post("/chat/completions", json=request.model_dump())
20
+ return await self.post("/chat/completions", json=request.model_dump(exclude_none=True))
21
21
 
22
22
  async def chat(self, request: OpenAIChatRequestSchema) -> OpenAIChatResponseSchema:
23
23
  response = await self.chat_api(request)
24
24
  return OpenAIChatResponseSchema.model_validate_json(response.text)
25
25
 
26
26
 
27
- def get_openai_http_client() -> OpenAIHTTPClient:
28
- logger = get_logger("OPENAI_HTTP_CLIENT")
27
+ def get_openai_v1_http_client() -> OpenAIV1HTTPClient:
28
+ logger = get_logger("OPENAI_V1_HTTP_CLIENT")
29
29
  logger_event_hook = LoggerEventHook(logger=logger)
30
30
  retry_transport = RetryTransport(logger=logger, transport=AsyncHTTPTransport())
31
31
 
@@ -40,4 +40,4 @@ def get_openai_http_client() -> OpenAIHTTPClient:
40
40
  }
41
41
  )
42
42
 
43
- return OpenAIHTTPClient(client=client)
43
+ return OpenAIV1HTTPClient(client=client)
@@ -21,7 +21,7 @@ class OpenAIChoiceSchema(BaseModel):
21
21
  class OpenAIChatRequestSchema(BaseModel):
22
22
  model: str
23
23
  messages: list[OpenAIMessageSchema]
24
- max_tokens: int
24
+ max_tokens: int | None = None
25
25
  temperature: float
26
26
 
27
27
 
@@ -0,0 +1,8 @@
1
+ from typing import Protocol
2
+
3
+ from ai_review.clients.openai.v1.schema import OpenAIChatRequestSchema, OpenAIChatResponseSchema
4
+
5
+
6
+ class OpenAIV1HTTPClientProtocol(Protocol):
7
+ async def chat(self, request: OpenAIChatRequestSchema) -> OpenAIChatResponseSchema:
8
+ ...
File without changes
@@ -0,0 +1,46 @@
1
+ from httpx import Response, AsyncClient, AsyncHTTPTransport
2
+
3
+ from ai_review.clients.openai.v2.schema import (
4
+ OpenAIResponsesRequestSchema,
5
+ OpenAIResponsesResponseSchema
6
+ )
7
+ from ai_review.clients.openai.v2.types import OpenAIV2HTTPClientProtocol
8
+ from ai_review.config import settings
9
+ from ai_review.libs.http.client import HTTPClient
10
+ from ai_review.libs.http.event_hooks.logger import LoggerEventHook
11
+ from ai_review.libs.http.handlers import HTTPClientError, handle_http_error
12
+ from ai_review.libs.http.transports.retry import RetryTransport
13
+ from ai_review.libs.logger import get_logger
14
+
15
+
16
+ class OpenAIV2HTTPClientError(HTTPClientError):
17
+ pass
18
+
19
+
20
+ class OpenAIV2HTTPClient(HTTPClient, OpenAIV2HTTPClientProtocol):
21
+ @handle_http_error(client='OpenAIV2HTTPClient', exception=OpenAIV2HTTPClientError)
22
+ async def chat_api(self, request: OpenAIResponsesRequestSchema) -> Response:
23
+ return await self.post("/responses", json=request.model_dump(exclude_none=True))
24
+
25
+ async def chat(self, request: OpenAIResponsesRequestSchema) -> OpenAIResponsesResponseSchema:
26
+ response = await self.chat_api(request)
27
+ return OpenAIResponsesResponseSchema.model_validate_json(response.text)
28
+
29
+
30
+ def get_openai_v2_http_client() -> OpenAIV2HTTPClient:
31
+ logger = get_logger("OPENAI_V2_HTTP_CLIENT")
32
+ logger_event_hook = LoggerEventHook(logger=logger)
33
+ retry_transport = RetryTransport(logger=logger, transport=AsyncHTTPTransport())
34
+
35
+ client = AsyncClient(
36
+ timeout=settings.llm.http_client.timeout,
37
+ headers={"Authorization": f"Bearer {settings.llm.http_client.api_token_value}"},
38
+ base_url=settings.llm.http_client.api_url_value,
39
+ transport=retry_transport,
40
+ event_hooks={
41
+ 'request': [logger_event_hook.request],
42
+ 'response': [logger_event_hook.response]
43
+ }
44
+ )
45
+
46
+ return OpenAIV2HTTPClient(client=client)
@@ -0,0 +1,47 @@
1
+ from pydantic import BaseModel
2
+
3
+
4
+ class OpenAIResponseUsageSchema(BaseModel):
5
+ total_tokens: int
6
+ input_tokens: int
7
+ output_tokens: int
8
+
9
+
10
+ class OpenAIInputMessageSchema(BaseModel):
11
+ role: str
12
+ content: str
13
+
14
+
15
+ class OpenAIResponseContentSchema(BaseModel):
16
+ type: str
17
+ text: str | None = None
18
+
19
+
20
+ class OpenAIResponseOutputSchema(BaseModel):
21
+ type: str
22
+ role: str | None = None
23
+ content: list[OpenAIResponseContentSchema] | None = None
24
+
25
+
26
+ class OpenAIResponsesRequestSchema(BaseModel):
27
+ model: str
28
+ input: list[OpenAIInputMessageSchema]
29
+ temperature: float | None = None
30
+ instructions: str | None = None
31
+ max_output_tokens: int | None = None
32
+
33
+
34
+ class OpenAIResponsesResponseSchema(BaseModel):
35
+ usage: OpenAIResponseUsageSchema
36
+ output: list[OpenAIResponseOutputSchema]
37
+
38
+ @property
39
+ def first_text(self) -> str:
40
+ results: list[str] = []
41
+ for block in self.output:
42
+ if block.type == "message" and block.content:
43
+ for content in block.content:
44
+ if content.type == "output_text" and content.text:
45
+ results.append(content.text)
46
+
47
+ return "".join(results).strip()
@@ -0,0 +1,11 @@
1
+ from typing import Protocol
2
+
3
+ from ai_review.clients.openai.v2.schema import (
4
+ OpenAIResponsesRequestSchema,
5
+ OpenAIResponsesResponseSchema
6
+ )
7
+
8
+
9
+ class OpenAIV2HTTPClientProtocol(Protocol):
10
+ async def chat(self, request: OpenAIResponsesRequestSchema) -> OpenAIResponsesResponseSchema:
11
+ ...
File without changes
@@ -0,0 +1,50 @@
1
+ from httpx import AsyncClient, Response, AsyncHTTPTransport
2
+
3
+ from ai_review.clients.openrouter.schema import OpenRouterChatRequestSchema, OpenRouterChatResponseSchema
4
+ from ai_review.clients.openrouter.types import OpenRouterHTTPClientProtocol
5
+ from ai_review.config import settings
6
+ from ai_review.libs.http.client import HTTPClient
7
+ from ai_review.libs.http.event_hooks.logger import LoggerEventHook
8
+ from ai_review.libs.http.handlers import HTTPClientError, handle_http_error
9
+ from ai_review.libs.http.transports.retry import RetryTransport
10
+ from ai_review.libs.logger import get_logger
11
+
12
+
13
+ class OpenRouterHTTPClientError(HTTPClientError):
14
+ pass
15
+
16
+
17
+ class OpenRouterHTTPClient(HTTPClient, OpenRouterHTTPClientProtocol):
18
+ @handle_http_error(client="OpenRouterHTTPClient", exception=OpenRouterHTTPClientError)
19
+ async def chat_api(self, request: OpenRouterChatRequestSchema) -> Response:
20
+ return await self.post("/chat/completions", json=request.model_dump(exclude_none=True))
21
+
22
+ async def chat(self, request: OpenRouterChatRequestSchema) -> OpenRouterChatResponseSchema:
23
+ response = await self.chat_api(request)
24
+ return OpenRouterChatResponseSchema.model_validate_json(response.text)
25
+
26
+
27
+ def get_openrouter_http_client() -> OpenRouterHTTPClient:
28
+ logger = get_logger("OPENROUTER_HTTP_CLIENT")
29
+ logger_event_hook = LoggerEventHook(logger=logger)
30
+ retry_transport = RetryTransport(logger=logger, transport=AsyncHTTPTransport())
31
+
32
+ headers = {"Authorization": f"Bearer {settings.llm.http_client.api_token_value}"}
33
+ if settings.llm.meta.title:
34
+ headers["X-Title"] = settings.llm.meta.title
35
+
36
+ if settings.llm.meta.referer:
37
+ headers["Referer"] = settings.llm.meta.referer
38
+
39
+ client = AsyncClient(
40
+ timeout=settings.llm.http_client.timeout,
41
+ headers=headers,
42
+ base_url=settings.llm.http_client.api_url_value,
43
+ transport=retry_transport,
44
+ event_hooks={
45
+ "request": [logger_event_hook.request],
46
+ "response": [logger_event_hook.response],
47
+ },
48
+ )
49
+
50
+ return OpenRouterHTTPClient(client=client)
@@ -0,0 +1,36 @@
1
+ from typing import Literal
2
+
3
+ from pydantic import BaseModel
4
+
5
+
6
+ class OpenRouterUsageSchema(BaseModel):
7
+ total_tokens: int
8
+ prompt_tokens: int
9
+ completion_tokens: int
10
+
11
+
12
+ class OpenRouterMessageSchema(BaseModel):
13
+ role: Literal["system", "user", "assistant"]
14
+ content: str
15
+
16
+
17
+ class OpenRouterChoiceSchema(BaseModel):
18
+ message: OpenRouterMessageSchema
19
+
20
+
21
+ class OpenRouterChatRequestSchema(BaseModel):
22
+ model: str
23
+ messages: list[OpenRouterMessageSchema]
24
+ max_tokens: int | None = None
25
+ temperature: float | None = None
26
+
27
+
28
+ class OpenRouterChatResponseSchema(BaseModel):
29
+ usage: OpenRouterUsageSchema
30
+ choices: list[OpenRouterChoiceSchema]
31
+
32
+ @property
33
+ def first_text(self) -> str:
34
+ if not self.choices:
35
+ return ""
36
+ return (self.choices[0].message.content or "").strip()
@@ -0,0 +1,11 @@
1
+ from typing import Protocol
2
+
3
+ from ai_review.clients.openrouter.schema import (
4
+ OpenRouterChatRequestSchema,
5
+ OpenRouterChatResponseSchema
6
+ )
7
+
8
+
9
+ class OpenRouterHTTPClientProtocol(Protocol):
10
+ async def chat(self, request: OpenRouterChatRequestSchema) -> OpenRouterChatResponseSchema:
11
+ ...
@@ -8,6 +8,7 @@ from ai_review.libs.config.llm.claude import ClaudeHTTPClientConfig, ClaudeMetaC
8
8
  from ai_review.libs.config.llm.gemini import GeminiHTTPClientConfig, GeminiMetaConfig
9
9
  from ai_review.libs.config.llm.ollama import OllamaHTTPClientConfig, OllamaMetaConfig
10
10
  from ai_review.libs.config.llm.openai import OpenAIHTTPClientConfig, OpenAIMetaConfig
11
+ from ai_review.libs.config.llm.openrouter import OpenRouterHTTPClientConfig, OpenRouterMetaConfig
11
12
  from ai_review.libs.constants.llm_provider import LLMProvider
12
13
  from ai_review.libs.resources import load_resource
13
14
 
@@ -62,7 +63,13 @@ class OllamaLLMConfig(LLMConfigBase):
62
63
  http_client: OllamaHTTPClientConfig
63
64
 
64
65
 
66
+ class OpenRouterLLMConfig(LLMConfigBase):
67
+ meta: OpenRouterMetaConfig
68
+ provider: Literal[LLMProvider.OPENROUTER]
69
+ http_client: OpenRouterHTTPClientConfig
70
+
71
+
65
72
  LLMConfig = Annotated[
66
- OpenAILLMConfig | GeminiLLMConfig | ClaudeLLMConfig | OllamaLLMConfig,
73
+ OpenAILLMConfig | GeminiLLMConfig | ClaudeLLMConfig | OllamaLLMConfig | OpenRouterLLMConfig,
67
74
  Field(discriminator="provider")
68
75
  ]
@@ -3,5 +3,5 @@ from pydantic import BaseModel, Field
3
3
 
4
4
  class LLMMetaConfig(BaseModel):
5
5
  model: str
6
- max_tokens: int = Field(default=5000, ge=1)
7
- temperature: float = Field(default=0.3, ge=0.0, le=2.0)
6
+ max_tokens: int | None = Field(default=None, ge=1)
7
+ temperature: float | None = Field(default=None, ge=0.0, le=2.0)
@@ -5,6 +5,10 @@ from ai_review.libs.config.llm.meta import LLMMetaConfig
5
5
  class OpenAIMetaConfig(LLMMetaConfig):
6
6
  model: str = "gpt-4o-mini"
7
7
 
8
+ @property
9
+ def is_v2_model(self) -> bool:
10
+ return any(self.model.startswith(model) for model in ("gpt-5", "gpt-4.1"))
11
+
8
12
 
9
13
  class OpenAIHTTPClientConfig(HTTPClientWithTokenConfig):
10
14
  pass
@@ -0,0 +1,12 @@
1
+ from ai_review.libs.config.http import HTTPClientWithTokenConfig
2
+ from ai_review.libs.config.llm.meta import LLMMetaConfig
3
+
4
+
5
+ class OpenRouterMetaConfig(LLMMetaConfig):
6
+ model: str = "openai/gpt-4o-mini"
7
+ title: str | None = None
8
+ referer: str | None = None
9
+
10
+
11
+ class OpenRouterHTTPClientConfig(HTTPClientWithTokenConfig):
12
+ pass
@@ -6,3 +6,4 @@ class LLMProvider(StrEnum):
6
6
  GEMINI = "GEMINI"
7
7
  CLAUDE = "CLAUDE"
8
8
  OLLAMA = "OLLAMA"
9
+ OPENROUTER = "OPENROUTER"
@@ -1,3 +1,14 @@
1
+ # ===============================
2
+ # 📊 Pricing per 1 token (USD)
3
+ # ===============================
4
+ # NOTE:
5
+ # - Prices are per 1 token and can change over time.
6
+ # - For the latest OpenAI and Anthropic pricing, see:
7
+ # - https://openai.com/api/pricing
8
+ # - https://www.anthropic.com/pricing
9
+ # - For OpenRouter models, prices vary by provider: https://openrouter.ai
10
+
11
+ # --- OpenAI ---
1
12
  gpt-4o-mini:
2
13
  input: 0.15e-6
3
14
  output: 0.60e-6
@@ -14,17 +25,22 @@ gpt-4.1:
14
25
  input: 5.00e-6
15
26
  output: 15.00e-6
16
27
 
28
+ gpt-5:
29
+ input: 10.00e-6
30
+ output: 30.00e-6
31
+
17
32
  gpt-3.5-turbo:
18
33
  input: 0.50e-6
19
34
  output: 1.50e-6
20
35
 
36
+ # --- Google Gemini ---
21
37
  gemini-2.5-flash-lite:
22
38
  input: 0.10e-6
23
39
  output: 0.40e-6
24
40
 
25
41
  gemini-2.0-flash-lite:
26
42
  input: 0.019e-6
27
- output: 0.1e-6
43
+ output: 0.10e-6
28
44
 
29
45
  gemini-2.0-pro:
30
46
  input: 0.125e-6
@@ -38,6 +54,7 @@ gemini-2.5-pro-long-context:
38
54
  input: 2.50e-6
39
55
  output: 15.00e-6
40
56
 
57
+ # --- Anthropic Claude ---
41
58
  claude-3.5-sonnet:
42
59
  input: 3.00e-6
43
60
  output: 15.00e-6
@@ -53,3 +70,24 @@ claude-3-sonnet:
53
70
  claude-3-haiku:
54
71
  input: 0.25e-6
55
72
  output: 1.25e-6
73
+
74
+ # --- OpenRouter ---
75
+ o3-mini:
76
+ input: 1.10e-6
77
+ output: 4.40e-6
78
+
79
+ openai/chatgpt-4o-latest:
80
+ input: 5.00e-6
81
+ output: 15.00e-6
82
+
83
+ mistralai/mixtral-8x7b:
84
+ input: 0.24e-6
85
+ output: 0.48e-6
86
+
87
+ meta-llama/llama-3-8b-instruct:
88
+ input: 0.20e-6
89
+ output: 0.40e-6
90
+
91
+ meta-llama/llama-3-70b-instruct:
92
+ input: 0.80e-6
93
+ output: 1.60e-6
@@ -4,6 +4,7 @@ from ai_review.services.llm.claude.client import ClaudeLLMClient
4
4
  from ai_review.services.llm.gemini.client import GeminiLLMClient
5
5
  from ai_review.services.llm.ollama.client import OllamaLLMClient
6
6
  from ai_review.services.llm.openai.client import OpenAILLMClient
7
+ from ai_review.services.llm.openrouter.client import OpenRouterLLMClient
7
8
  from ai_review.services.llm.types import LLMClientProtocol
8
9
 
9
10
 
@@ -17,5 +18,7 @@ def get_llm_client() -> LLMClientProtocol:
17
18
  return ClaudeLLMClient()
18
19
  case LLMProvider.OLLAMA:
19
20
  return OllamaLLMClient()
21
+ case LLMProvider.OPENROUTER:
22
+ return OpenRouterLLMClient()
20
23
  case _:
21
24
  raise ValueError(f"Unsupported LLM provider: {settings.llm.provider}")
@@ -1,28 +1,56 @@
1
- from ai_review.clients.openai.client import get_openai_http_client
2
- from ai_review.clients.openai.schema import OpenAIChatRequestSchema, OpenAIMessageSchema
1
+ from ai_review.clients.openai.v1.client import get_openai_v1_http_client
2
+ from ai_review.clients.openai.v1.schema import OpenAIChatRequestSchema, OpenAIMessageSchema
3
+ from ai_review.clients.openai.v2.client import get_openai_v2_http_client
4
+ from ai_review.clients.openai.v2.schema import OpenAIInputMessageSchema, OpenAIResponsesRequestSchema
3
5
  from ai_review.config import settings
4
6
  from ai_review.services.llm.types import LLMClientProtocol, ChatResultSchema
5
7
 
6
8
 
7
9
  class OpenAILLMClient(LLMClientProtocol):
8
10
  def __init__(self):
9
- self.http_client = get_openai_http_client()
11
+ self.meta = settings.llm.meta
10
12
 
11
- async def chat(self, prompt: str, prompt_system: str) -> ChatResultSchema:
12
- meta = settings.llm.meta
13
+ self.http_client_v1 = get_openai_v1_http_client()
14
+ self.http_client_v2 = get_openai_v2_http_client()
15
+
16
+ async def chat_v1(self, prompt: str, prompt_system: str) -> ChatResultSchema:
13
17
  request = OpenAIChatRequestSchema(
14
- model=meta.model,
18
+ model=self.meta.model,
15
19
  messages=[
16
20
  OpenAIMessageSchema(role="system", content=prompt_system),
17
21
  OpenAIMessageSchema(role="user", content=prompt),
18
22
  ],
19
- max_tokens=meta.max_tokens,
20
- temperature=meta.temperature,
23
+ max_tokens=self.meta.max_tokens,
24
+ temperature=self.meta.temperature,
21
25
  )
22
- response = await self.http_client.chat(request)
26
+ response = await self.http_client_v1.chat(request)
23
27
  return ChatResultSchema(
24
28
  text=response.first_text,
25
29
  total_tokens=response.usage.total_tokens,
26
30
  prompt_tokens=response.usage.prompt_tokens,
27
31
  completion_tokens=response.usage.completion_tokens,
28
32
  )
33
+
34
+ async def chat_v2(self, prompt: str, prompt_system: str) -> ChatResultSchema:
35
+ request = OpenAIResponsesRequestSchema(
36
+ model=self.meta.model,
37
+ input=[
38
+ OpenAIInputMessageSchema(role="system", content=prompt_system),
39
+ OpenAIInputMessageSchema(role="user", content=prompt),
40
+ ],
41
+ temperature=self.meta.temperature,
42
+ max_output_tokens=self.meta.max_tokens,
43
+ )
44
+ response = await self.http_client_v2.chat(request)
45
+ return ChatResultSchema(
46
+ text=response.first_text,
47
+ total_tokens=response.usage.total_tokens,
48
+ prompt_tokens=response.usage.input_tokens,
49
+ completion_tokens=response.usage.output_tokens,
50
+ )
51
+
52
+ async def chat(self, prompt: str, prompt_system: str) -> ChatResultSchema:
53
+ if self.meta.is_v2_model:
54
+ return await self.chat_v2(prompt, prompt_system)
55
+
56
+ return await self.chat_v1(prompt, prompt_system)
File without changes
@@ -0,0 +1,31 @@
1
+ from ai_review.clients.openrouter.client import get_openrouter_http_client
2
+ from ai_review.clients.openrouter.schema import (
3
+ OpenRouterMessageSchema,
4
+ OpenRouterChatRequestSchema,
5
+ )
6
+ from ai_review.config import settings
7
+ from ai_review.services.llm.types import LLMClientProtocol, ChatResultSchema
8
+
9
+
10
+ class OpenRouterLLMClient(LLMClientProtocol):
11
+ def __init__(self):
12
+ self.http_client = get_openrouter_http_client()
13
+
14
+ async def chat(self, prompt: str, prompt_system: str) -> ChatResultSchema:
15
+ meta = settings.llm.meta
16
+ request = OpenRouterChatRequestSchema(
17
+ model=meta.model,
18
+ messages=[
19
+ OpenRouterMessageSchema(role="system", content=prompt_system),
20
+ OpenRouterMessageSchema(role="user", content=prompt),
21
+ ],
22
+ max_tokens=meta.max_tokens,
23
+ temperature=meta.temperature,
24
+ )
25
+ response = await self.http_client.chat(request)
26
+ return ChatResultSchema(
27
+ text=response.first_text,
28
+ total_tokens=response.usage.total_tokens,
29
+ prompt_tokens=response.usage.prompt_tokens,
30
+ completion_tokens=response.usage.completion_tokens,
31
+ )