xai-review 0.33.0__py3-none-any.whl → 0.35.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of xai-review might be problematic. Click here for more details.
- ai_review/clients/claude/client.py +1 -1
- ai_review/clients/claude/schema.py +2 -2
- ai_review/clients/gemini/client.py +2 -1
- ai_review/clients/gemini/schema.py +2 -2
- ai_review/clients/ollama/client.py +1 -1
- ai_review/clients/openai/v1/__init__.py +0 -0
- ai_review/clients/openai/{client.py → v1/client.py} +9 -9
- ai_review/clients/openai/{schema.py → v1/schema.py} +1 -1
- ai_review/clients/openai/v1/types.py +8 -0
- ai_review/clients/openai/v2/__init__.py +0 -0
- ai_review/clients/openai/v2/client.py +46 -0
- ai_review/clients/openai/v2/schema.py +47 -0
- ai_review/clients/openai/v2/types.py +11 -0
- ai_review/clients/openrouter/__init__.py +0 -0
- ai_review/clients/openrouter/client.py +50 -0
- ai_review/clients/openrouter/schema.py +36 -0
- ai_review/clients/openrouter/types.py +11 -0
- ai_review/libs/config/llm/base.py +8 -1
- ai_review/libs/config/llm/meta.py +2 -2
- ai_review/libs/config/llm/openai.py +4 -0
- ai_review/libs/config/llm/openrouter.py +12 -0
- ai_review/libs/constants/llm_provider.py +1 -0
- ai_review/resources/pricing.yaml +39 -1
- ai_review/services/llm/factory.py +3 -0
- ai_review/services/llm/openai/client.py +37 -9
- ai_review/services/llm/openrouter/__init__.py +0 -0
- ai_review/services/llm/openrouter/client.py +31 -0
- ai_review/tests/fixtures/clients/openai.py +84 -12
- ai_review/tests/fixtures/clients/openrouter.py +72 -0
- ai_review/tests/suites/clients/openai/v1/__init__.py +0 -0
- ai_review/tests/suites/clients/openai/v1/test_client.py +12 -0
- ai_review/tests/suites/clients/openai/{test_schema.py → v1/test_schema.py} +1 -1
- ai_review/tests/suites/clients/openai/v2/__init__.py +0 -0
- ai_review/tests/suites/clients/openai/v2/test_client.py +12 -0
- ai_review/tests/suites/clients/openai/v2/test_schema.py +80 -0
- ai_review/tests/suites/clients/openrouter/__init__.py +0 -0
- ai_review/tests/suites/clients/openrouter/test_client.py +12 -0
- ai_review/tests/suites/clients/openrouter/test_schema.py +57 -0
- ai_review/tests/suites/libs/config/llm/__init__.py +0 -0
- ai_review/tests/suites/libs/config/llm/test_openai.py +28 -0
- ai_review/tests/suites/services/llm/openai/test_client.py +23 -6
- ai_review/tests/suites/services/llm/openrouter/__init__.py +0 -0
- ai_review/tests/suites/services/llm/openrouter/test_client.py +22 -0
- ai_review/tests/suites/services/llm/test_factory.py +8 -1
- ai_review/tests/suites/services/review/gateway/test_review_comment_gateway.py +5 -5
- ai_review/tests/suites/services/review/gateway/test_review_dry_run_comment_gateway.py +5 -5
- ai_review/tests/suites/services/review/gateway/test_review_llm_gateway.py +2 -2
- ai_review/tests/suites/services/review/test_service.py +2 -2
- {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/METADATA +11 -10
- {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/RECORD +54 -30
- ai_review/clients/openai/types.py +0 -8
- ai_review/tests/suites/clients/openai/test_client.py +0 -12
- {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/WHEEL +0 -0
- {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/entry_points.txt +0 -0
- {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/licenses/LICENSE +0 -0
- {xai_review-0.33.0.dist-info → xai_review-0.35.0.dist-info}/top_level.txt +0 -0
|
@@ -17,7 +17,7 @@ class ClaudeHTTPClientError(HTTPClientError):
|
|
|
17
17
|
class ClaudeHTTPClient(HTTPClient, ClaudeHTTPClientProtocol):
|
|
18
18
|
@handle_http_error(client="ClaudeHTTPClient", exception=ClaudeHTTPClientError)
|
|
19
19
|
async def chat_api(self, request: ClaudeChatRequestSchema) -> Response:
|
|
20
|
-
return await self.post("/v1/messages", json=request.model_dump())
|
|
20
|
+
return await self.post("/v1/messages", json=request.model_dump(exclude_none=True))
|
|
21
21
|
|
|
22
22
|
async def chat(self, request: ClaudeChatRequestSchema) -> ClaudeChatResponseSchema:
|
|
23
23
|
response = await self.chat_api(request)
|
|
@@ -12,8 +12,8 @@ class ClaudeChatRequestSchema(BaseModel):
|
|
|
12
12
|
model: str
|
|
13
13
|
system: str | None = None
|
|
14
14
|
messages: list[ClaudeMessageSchema]
|
|
15
|
-
max_tokens: int
|
|
16
|
-
temperature: float
|
|
15
|
+
max_tokens: int | None = None
|
|
16
|
+
temperature: float | None = None
|
|
17
17
|
|
|
18
18
|
|
|
19
19
|
class ClaudeContentSchema(BaseModel):
|
|
@@ -19,7 +19,8 @@ class GeminiHTTPClient(HTTPClient, GeminiHTTPClientProtocol):
|
|
|
19
19
|
async def chat_api(self, request: GeminiChatRequestSchema) -> Response:
|
|
20
20
|
meta = settings.llm.meta
|
|
21
21
|
return await self.post(
|
|
22
|
-
f"/v1beta/models/{meta.model}:generateContent",
|
|
22
|
+
f"/v1beta/models/{meta.model}:generateContent",
|
|
23
|
+
json=request.model_dump(exclude_none=True)
|
|
23
24
|
)
|
|
24
25
|
|
|
25
26
|
async def chat(self, request: GeminiChatRequestSchema) -> GeminiChatResponseSchema:
|
|
@@ -45,8 +45,8 @@ class GeminiCandidateSchema(BaseModel):
|
|
|
45
45
|
class GeminiGenerationConfigSchema(BaseModel):
|
|
46
46
|
model_config = ConfigDict(populate_by_name=True)
|
|
47
47
|
|
|
48
|
-
temperature: float
|
|
49
|
-
max_output_tokens: int = Field(alias="maxOutputTokens")
|
|
48
|
+
temperature: float | None = None
|
|
49
|
+
max_output_tokens: int | None = Field(alias="maxOutputTokens", default=None)
|
|
50
50
|
|
|
51
51
|
|
|
52
52
|
class GeminiChatRequestSchema(BaseModel):
|
|
@@ -17,7 +17,7 @@ class OllamaHTTPClientError(HTTPClientError):
|
|
|
17
17
|
class OllamaHTTPClient(HTTPClient, OllamaHTTPClientProtocol):
|
|
18
18
|
@handle_http_error(client="OllamaHTTPClient", exception=OllamaHTTPClientError)
|
|
19
19
|
async def chat_api(self, request: OllamaChatRequestSchema) -> Response:
|
|
20
|
-
return await self.post("/api/chat", json=request.model_dump())
|
|
20
|
+
return await self.post("/api/chat", json=request.model_dump(exclude_none=True))
|
|
21
21
|
|
|
22
22
|
async def chat(self, request: OllamaChatRequestSchema) -> OllamaChatResponseSchema:
|
|
23
23
|
response = await self.chat_api(request)
|
|
File without changes
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
from httpx import Response, AsyncHTTPTransport, AsyncClient
|
|
2
2
|
|
|
3
|
-
from ai_review.clients.openai.schema import OpenAIChatRequestSchema, OpenAIChatResponseSchema
|
|
4
|
-
from ai_review.clients.openai.types import
|
|
3
|
+
from ai_review.clients.openai.v1.schema import OpenAIChatRequestSchema, OpenAIChatResponseSchema
|
|
4
|
+
from ai_review.clients.openai.v1.types import OpenAIV1HTTPClientProtocol
|
|
5
5
|
from ai_review.config import settings
|
|
6
6
|
from ai_review.libs.http.client import HTTPClient
|
|
7
7
|
from ai_review.libs.http.event_hooks.logger import LoggerEventHook
|
|
@@ -10,22 +10,22 @@ from ai_review.libs.http.transports.retry import RetryTransport
|
|
|
10
10
|
from ai_review.libs.logger import get_logger
|
|
11
11
|
|
|
12
12
|
|
|
13
|
-
class
|
|
13
|
+
class OpenAIV1HTTPClientError(HTTPClientError):
|
|
14
14
|
pass
|
|
15
15
|
|
|
16
16
|
|
|
17
|
-
class
|
|
18
|
-
@handle_http_error(client='
|
|
17
|
+
class OpenAIV1HTTPClient(HTTPClient, OpenAIV1HTTPClientProtocol):
|
|
18
|
+
@handle_http_error(client='OpenAIV1HTTPClient', exception=OpenAIV1HTTPClientError)
|
|
19
19
|
async def chat_api(self, request: OpenAIChatRequestSchema) -> Response:
|
|
20
|
-
return await self.post("/chat/completions", json=request.model_dump())
|
|
20
|
+
return await self.post("/chat/completions", json=request.model_dump(exclude_none=True))
|
|
21
21
|
|
|
22
22
|
async def chat(self, request: OpenAIChatRequestSchema) -> OpenAIChatResponseSchema:
|
|
23
23
|
response = await self.chat_api(request)
|
|
24
24
|
return OpenAIChatResponseSchema.model_validate_json(response.text)
|
|
25
25
|
|
|
26
26
|
|
|
27
|
-
def
|
|
28
|
-
logger = get_logger("
|
|
27
|
+
def get_openai_v1_http_client() -> OpenAIV1HTTPClient:
|
|
28
|
+
logger = get_logger("OPENAI_V1_HTTP_CLIENT")
|
|
29
29
|
logger_event_hook = LoggerEventHook(logger=logger)
|
|
30
30
|
retry_transport = RetryTransport(logger=logger, transport=AsyncHTTPTransport())
|
|
31
31
|
|
|
@@ -40,4 +40,4 @@ def get_openai_http_client() -> OpenAIHTTPClient:
|
|
|
40
40
|
}
|
|
41
41
|
)
|
|
42
42
|
|
|
43
|
-
return
|
|
43
|
+
return OpenAIV1HTTPClient(client=client)
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
from typing import Protocol
|
|
2
|
+
|
|
3
|
+
from ai_review.clients.openai.v1.schema import OpenAIChatRequestSchema, OpenAIChatResponseSchema
|
|
4
|
+
|
|
5
|
+
|
|
6
|
+
class OpenAIV1HTTPClientProtocol(Protocol):
|
|
7
|
+
async def chat(self, request: OpenAIChatRequestSchema) -> OpenAIChatResponseSchema:
|
|
8
|
+
...
|
|
File without changes
|
|
@@ -0,0 +1,46 @@
|
|
|
1
|
+
from httpx import Response, AsyncClient, AsyncHTTPTransport
|
|
2
|
+
|
|
3
|
+
from ai_review.clients.openai.v2.schema import (
|
|
4
|
+
OpenAIResponsesRequestSchema,
|
|
5
|
+
OpenAIResponsesResponseSchema
|
|
6
|
+
)
|
|
7
|
+
from ai_review.clients.openai.v2.types import OpenAIV2HTTPClientProtocol
|
|
8
|
+
from ai_review.config import settings
|
|
9
|
+
from ai_review.libs.http.client import HTTPClient
|
|
10
|
+
from ai_review.libs.http.event_hooks.logger import LoggerEventHook
|
|
11
|
+
from ai_review.libs.http.handlers import HTTPClientError, handle_http_error
|
|
12
|
+
from ai_review.libs.http.transports.retry import RetryTransport
|
|
13
|
+
from ai_review.libs.logger import get_logger
|
|
14
|
+
|
|
15
|
+
|
|
16
|
+
class OpenAIV2HTTPClientError(HTTPClientError):
|
|
17
|
+
pass
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
class OpenAIV2HTTPClient(HTTPClient, OpenAIV2HTTPClientProtocol):
|
|
21
|
+
@handle_http_error(client='OpenAIV2HTTPClient', exception=OpenAIV2HTTPClientError)
|
|
22
|
+
async def chat_api(self, request: OpenAIResponsesRequestSchema) -> Response:
|
|
23
|
+
return await self.post("/responses", json=request.model_dump(exclude_none=True))
|
|
24
|
+
|
|
25
|
+
async def chat(self, request: OpenAIResponsesRequestSchema) -> OpenAIResponsesResponseSchema:
|
|
26
|
+
response = await self.chat_api(request)
|
|
27
|
+
return OpenAIResponsesResponseSchema.model_validate_json(response.text)
|
|
28
|
+
|
|
29
|
+
|
|
30
|
+
def get_openai_v2_http_client() -> OpenAIV2HTTPClient:
|
|
31
|
+
logger = get_logger("OPENAI_V2_HTTP_CLIENT")
|
|
32
|
+
logger_event_hook = LoggerEventHook(logger=logger)
|
|
33
|
+
retry_transport = RetryTransport(logger=logger, transport=AsyncHTTPTransport())
|
|
34
|
+
|
|
35
|
+
client = AsyncClient(
|
|
36
|
+
timeout=settings.llm.http_client.timeout,
|
|
37
|
+
headers={"Authorization": f"Bearer {settings.llm.http_client.api_token_value}"},
|
|
38
|
+
base_url=settings.llm.http_client.api_url_value,
|
|
39
|
+
transport=retry_transport,
|
|
40
|
+
event_hooks={
|
|
41
|
+
'request': [logger_event_hook.request],
|
|
42
|
+
'response': [logger_event_hook.response]
|
|
43
|
+
}
|
|
44
|
+
)
|
|
45
|
+
|
|
46
|
+
return OpenAIV2HTTPClient(client=client)
|
|
@@ -0,0 +1,47 @@
|
|
|
1
|
+
from pydantic import BaseModel
|
|
2
|
+
|
|
3
|
+
|
|
4
|
+
class OpenAIResponseUsageSchema(BaseModel):
|
|
5
|
+
total_tokens: int
|
|
6
|
+
input_tokens: int
|
|
7
|
+
output_tokens: int
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
class OpenAIInputMessageSchema(BaseModel):
|
|
11
|
+
role: str
|
|
12
|
+
content: str
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
class OpenAIResponseContentSchema(BaseModel):
|
|
16
|
+
type: str
|
|
17
|
+
text: str | None = None
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
class OpenAIResponseOutputSchema(BaseModel):
|
|
21
|
+
type: str
|
|
22
|
+
role: str | None = None
|
|
23
|
+
content: list[OpenAIResponseContentSchema] | None = None
|
|
24
|
+
|
|
25
|
+
|
|
26
|
+
class OpenAIResponsesRequestSchema(BaseModel):
|
|
27
|
+
model: str
|
|
28
|
+
input: list[OpenAIInputMessageSchema]
|
|
29
|
+
temperature: float | None = None
|
|
30
|
+
instructions: str | None = None
|
|
31
|
+
max_output_tokens: int | None = None
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
class OpenAIResponsesResponseSchema(BaseModel):
|
|
35
|
+
usage: OpenAIResponseUsageSchema
|
|
36
|
+
output: list[OpenAIResponseOutputSchema]
|
|
37
|
+
|
|
38
|
+
@property
|
|
39
|
+
def first_text(self) -> str:
|
|
40
|
+
results: list[str] = []
|
|
41
|
+
for block in self.output:
|
|
42
|
+
if block.type == "message" and block.content:
|
|
43
|
+
for content in block.content:
|
|
44
|
+
if content.type == "output_text" and content.text:
|
|
45
|
+
results.append(content.text)
|
|
46
|
+
|
|
47
|
+
return "".join(results).strip()
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
from typing import Protocol
|
|
2
|
+
|
|
3
|
+
from ai_review.clients.openai.v2.schema import (
|
|
4
|
+
OpenAIResponsesRequestSchema,
|
|
5
|
+
OpenAIResponsesResponseSchema
|
|
6
|
+
)
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class OpenAIV2HTTPClientProtocol(Protocol):
|
|
10
|
+
async def chat(self, request: OpenAIResponsesRequestSchema) -> OpenAIResponsesResponseSchema:
|
|
11
|
+
...
|
|
File without changes
|
|
@@ -0,0 +1,50 @@
|
|
|
1
|
+
from httpx import AsyncClient, Response, AsyncHTTPTransport
|
|
2
|
+
|
|
3
|
+
from ai_review.clients.openrouter.schema import OpenRouterChatRequestSchema, OpenRouterChatResponseSchema
|
|
4
|
+
from ai_review.clients.openrouter.types import OpenRouterHTTPClientProtocol
|
|
5
|
+
from ai_review.config import settings
|
|
6
|
+
from ai_review.libs.http.client import HTTPClient
|
|
7
|
+
from ai_review.libs.http.event_hooks.logger import LoggerEventHook
|
|
8
|
+
from ai_review.libs.http.handlers import HTTPClientError, handle_http_error
|
|
9
|
+
from ai_review.libs.http.transports.retry import RetryTransport
|
|
10
|
+
from ai_review.libs.logger import get_logger
|
|
11
|
+
|
|
12
|
+
|
|
13
|
+
class OpenRouterHTTPClientError(HTTPClientError):
|
|
14
|
+
pass
|
|
15
|
+
|
|
16
|
+
|
|
17
|
+
class OpenRouterHTTPClient(HTTPClient, OpenRouterHTTPClientProtocol):
|
|
18
|
+
@handle_http_error(client="OpenRouterHTTPClient", exception=OpenRouterHTTPClientError)
|
|
19
|
+
async def chat_api(self, request: OpenRouterChatRequestSchema) -> Response:
|
|
20
|
+
return await self.post("/chat/completions", json=request.model_dump(exclude_none=True))
|
|
21
|
+
|
|
22
|
+
async def chat(self, request: OpenRouterChatRequestSchema) -> OpenRouterChatResponseSchema:
|
|
23
|
+
response = await self.chat_api(request)
|
|
24
|
+
return OpenRouterChatResponseSchema.model_validate_json(response.text)
|
|
25
|
+
|
|
26
|
+
|
|
27
|
+
def get_openrouter_http_client() -> OpenRouterHTTPClient:
|
|
28
|
+
logger = get_logger("OPENROUTER_HTTP_CLIENT")
|
|
29
|
+
logger_event_hook = LoggerEventHook(logger=logger)
|
|
30
|
+
retry_transport = RetryTransport(logger=logger, transport=AsyncHTTPTransport())
|
|
31
|
+
|
|
32
|
+
headers = {"Authorization": f"Bearer {settings.llm.http_client.api_token_value}"}
|
|
33
|
+
if settings.llm.meta.title:
|
|
34
|
+
headers["X-Title"] = settings.llm.meta.title
|
|
35
|
+
|
|
36
|
+
if settings.llm.meta.referer:
|
|
37
|
+
headers["Referer"] = settings.llm.meta.referer
|
|
38
|
+
|
|
39
|
+
client = AsyncClient(
|
|
40
|
+
timeout=settings.llm.http_client.timeout,
|
|
41
|
+
headers=headers,
|
|
42
|
+
base_url=settings.llm.http_client.api_url_value,
|
|
43
|
+
transport=retry_transport,
|
|
44
|
+
event_hooks={
|
|
45
|
+
"request": [logger_event_hook.request],
|
|
46
|
+
"response": [logger_event_hook.response],
|
|
47
|
+
},
|
|
48
|
+
)
|
|
49
|
+
|
|
50
|
+
return OpenRouterHTTPClient(client=client)
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
from typing import Literal
|
|
2
|
+
|
|
3
|
+
from pydantic import BaseModel
|
|
4
|
+
|
|
5
|
+
|
|
6
|
+
class OpenRouterUsageSchema(BaseModel):
|
|
7
|
+
total_tokens: int
|
|
8
|
+
prompt_tokens: int
|
|
9
|
+
completion_tokens: int
|
|
10
|
+
|
|
11
|
+
|
|
12
|
+
class OpenRouterMessageSchema(BaseModel):
|
|
13
|
+
role: Literal["system", "user", "assistant"]
|
|
14
|
+
content: str
|
|
15
|
+
|
|
16
|
+
|
|
17
|
+
class OpenRouterChoiceSchema(BaseModel):
|
|
18
|
+
message: OpenRouterMessageSchema
|
|
19
|
+
|
|
20
|
+
|
|
21
|
+
class OpenRouterChatRequestSchema(BaseModel):
|
|
22
|
+
model: str
|
|
23
|
+
messages: list[OpenRouterMessageSchema]
|
|
24
|
+
max_tokens: int | None = None
|
|
25
|
+
temperature: float | None = None
|
|
26
|
+
|
|
27
|
+
|
|
28
|
+
class OpenRouterChatResponseSchema(BaseModel):
|
|
29
|
+
usage: OpenRouterUsageSchema
|
|
30
|
+
choices: list[OpenRouterChoiceSchema]
|
|
31
|
+
|
|
32
|
+
@property
|
|
33
|
+
def first_text(self) -> str:
|
|
34
|
+
if not self.choices:
|
|
35
|
+
return ""
|
|
36
|
+
return (self.choices[0].message.content or "").strip()
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
from typing import Protocol
|
|
2
|
+
|
|
3
|
+
from ai_review.clients.openrouter.schema import (
|
|
4
|
+
OpenRouterChatRequestSchema,
|
|
5
|
+
OpenRouterChatResponseSchema
|
|
6
|
+
)
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
class OpenRouterHTTPClientProtocol(Protocol):
|
|
10
|
+
async def chat(self, request: OpenRouterChatRequestSchema) -> OpenRouterChatResponseSchema:
|
|
11
|
+
...
|
|
@@ -8,6 +8,7 @@ from ai_review.libs.config.llm.claude import ClaudeHTTPClientConfig, ClaudeMetaC
|
|
|
8
8
|
from ai_review.libs.config.llm.gemini import GeminiHTTPClientConfig, GeminiMetaConfig
|
|
9
9
|
from ai_review.libs.config.llm.ollama import OllamaHTTPClientConfig, OllamaMetaConfig
|
|
10
10
|
from ai_review.libs.config.llm.openai import OpenAIHTTPClientConfig, OpenAIMetaConfig
|
|
11
|
+
from ai_review.libs.config.llm.openrouter import OpenRouterHTTPClientConfig, OpenRouterMetaConfig
|
|
11
12
|
from ai_review.libs.constants.llm_provider import LLMProvider
|
|
12
13
|
from ai_review.libs.resources import load_resource
|
|
13
14
|
|
|
@@ -62,7 +63,13 @@ class OllamaLLMConfig(LLMConfigBase):
|
|
|
62
63
|
http_client: OllamaHTTPClientConfig
|
|
63
64
|
|
|
64
65
|
|
|
66
|
+
class OpenRouterLLMConfig(LLMConfigBase):
|
|
67
|
+
meta: OpenRouterMetaConfig
|
|
68
|
+
provider: Literal[LLMProvider.OPENROUTER]
|
|
69
|
+
http_client: OpenRouterHTTPClientConfig
|
|
70
|
+
|
|
71
|
+
|
|
65
72
|
LLMConfig = Annotated[
|
|
66
|
-
OpenAILLMConfig | GeminiLLMConfig | ClaudeLLMConfig | OllamaLLMConfig,
|
|
73
|
+
OpenAILLMConfig | GeminiLLMConfig | ClaudeLLMConfig | OllamaLLMConfig | OpenRouterLLMConfig,
|
|
67
74
|
Field(discriminator="provider")
|
|
68
75
|
]
|
|
@@ -3,5 +3,5 @@ from pydantic import BaseModel, Field
|
|
|
3
3
|
|
|
4
4
|
class LLMMetaConfig(BaseModel):
|
|
5
5
|
model: str
|
|
6
|
-
max_tokens: int = Field(default=
|
|
7
|
-
temperature: float = Field(default=
|
|
6
|
+
max_tokens: int | None = Field(default=None, ge=1)
|
|
7
|
+
temperature: float | None = Field(default=None, ge=0.0, le=2.0)
|
|
@@ -5,6 +5,10 @@ from ai_review.libs.config.llm.meta import LLMMetaConfig
|
|
|
5
5
|
class OpenAIMetaConfig(LLMMetaConfig):
|
|
6
6
|
model: str = "gpt-4o-mini"
|
|
7
7
|
|
|
8
|
+
@property
|
|
9
|
+
def is_v2_model(self) -> bool:
|
|
10
|
+
return any(self.model.startswith(model) for model in ("gpt-5", "gpt-4.1"))
|
|
11
|
+
|
|
8
12
|
|
|
9
13
|
class OpenAIHTTPClientConfig(HTTPClientWithTokenConfig):
|
|
10
14
|
pass
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
from ai_review.libs.config.http import HTTPClientWithTokenConfig
|
|
2
|
+
from ai_review.libs.config.llm.meta import LLMMetaConfig
|
|
3
|
+
|
|
4
|
+
|
|
5
|
+
class OpenRouterMetaConfig(LLMMetaConfig):
|
|
6
|
+
model: str = "openai/gpt-4o-mini"
|
|
7
|
+
title: str | None = None
|
|
8
|
+
referer: str | None = None
|
|
9
|
+
|
|
10
|
+
|
|
11
|
+
class OpenRouterHTTPClientConfig(HTTPClientWithTokenConfig):
|
|
12
|
+
pass
|
ai_review/resources/pricing.yaml
CHANGED
|
@@ -1,3 +1,14 @@
|
|
|
1
|
+
# ===============================
|
|
2
|
+
# 📊 Pricing per 1 token (USD)
|
|
3
|
+
# ===============================
|
|
4
|
+
# NOTE:
|
|
5
|
+
# - Prices are per 1 token and can change over time.
|
|
6
|
+
# - For the latest OpenAI and Anthropic pricing, see:
|
|
7
|
+
# - https://openai.com/api/pricing
|
|
8
|
+
# - https://www.anthropic.com/pricing
|
|
9
|
+
# - For OpenRouter models, prices vary by provider: https://openrouter.ai
|
|
10
|
+
|
|
11
|
+
# --- OpenAI ---
|
|
1
12
|
gpt-4o-mini:
|
|
2
13
|
input: 0.15e-6
|
|
3
14
|
output: 0.60e-6
|
|
@@ -14,17 +25,22 @@ gpt-4.1:
|
|
|
14
25
|
input: 5.00e-6
|
|
15
26
|
output: 15.00e-6
|
|
16
27
|
|
|
28
|
+
gpt-5:
|
|
29
|
+
input: 10.00e-6
|
|
30
|
+
output: 30.00e-6
|
|
31
|
+
|
|
17
32
|
gpt-3.5-turbo:
|
|
18
33
|
input: 0.50e-6
|
|
19
34
|
output: 1.50e-6
|
|
20
35
|
|
|
36
|
+
# --- Google Gemini ---
|
|
21
37
|
gemini-2.5-flash-lite:
|
|
22
38
|
input: 0.10e-6
|
|
23
39
|
output: 0.40e-6
|
|
24
40
|
|
|
25
41
|
gemini-2.0-flash-lite:
|
|
26
42
|
input: 0.019e-6
|
|
27
|
-
output: 0.
|
|
43
|
+
output: 0.10e-6
|
|
28
44
|
|
|
29
45
|
gemini-2.0-pro:
|
|
30
46
|
input: 0.125e-6
|
|
@@ -38,6 +54,7 @@ gemini-2.5-pro-long-context:
|
|
|
38
54
|
input: 2.50e-6
|
|
39
55
|
output: 15.00e-6
|
|
40
56
|
|
|
57
|
+
# --- Anthropic Claude ---
|
|
41
58
|
claude-3.5-sonnet:
|
|
42
59
|
input: 3.00e-6
|
|
43
60
|
output: 15.00e-6
|
|
@@ -53,3 +70,24 @@ claude-3-sonnet:
|
|
|
53
70
|
claude-3-haiku:
|
|
54
71
|
input: 0.25e-6
|
|
55
72
|
output: 1.25e-6
|
|
73
|
+
|
|
74
|
+
# --- OpenRouter ---
|
|
75
|
+
o3-mini:
|
|
76
|
+
input: 1.10e-6
|
|
77
|
+
output: 4.40e-6
|
|
78
|
+
|
|
79
|
+
openai/chatgpt-4o-latest:
|
|
80
|
+
input: 5.00e-6
|
|
81
|
+
output: 15.00e-6
|
|
82
|
+
|
|
83
|
+
mistralai/mixtral-8x7b:
|
|
84
|
+
input: 0.24e-6
|
|
85
|
+
output: 0.48e-6
|
|
86
|
+
|
|
87
|
+
meta-llama/llama-3-8b-instruct:
|
|
88
|
+
input: 0.20e-6
|
|
89
|
+
output: 0.40e-6
|
|
90
|
+
|
|
91
|
+
meta-llama/llama-3-70b-instruct:
|
|
92
|
+
input: 0.80e-6
|
|
93
|
+
output: 1.60e-6
|
|
@@ -4,6 +4,7 @@ from ai_review.services.llm.claude.client import ClaudeLLMClient
|
|
|
4
4
|
from ai_review.services.llm.gemini.client import GeminiLLMClient
|
|
5
5
|
from ai_review.services.llm.ollama.client import OllamaLLMClient
|
|
6
6
|
from ai_review.services.llm.openai.client import OpenAILLMClient
|
|
7
|
+
from ai_review.services.llm.openrouter.client import OpenRouterLLMClient
|
|
7
8
|
from ai_review.services.llm.types import LLMClientProtocol
|
|
8
9
|
|
|
9
10
|
|
|
@@ -17,5 +18,7 @@ def get_llm_client() -> LLMClientProtocol:
|
|
|
17
18
|
return ClaudeLLMClient()
|
|
18
19
|
case LLMProvider.OLLAMA:
|
|
19
20
|
return OllamaLLMClient()
|
|
21
|
+
case LLMProvider.OPENROUTER:
|
|
22
|
+
return OpenRouterLLMClient()
|
|
20
23
|
case _:
|
|
21
24
|
raise ValueError(f"Unsupported LLM provider: {settings.llm.provider}")
|
|
@@ -1,28 +1,56 @@
|
|
|
1
|
-
from ai_review.clients.openai.client import
|
|
2
|
-
from ai_review.clients.openai.schema import OpenAIChatRequestSchema, OpenAIMessageSchema
|
|
1
|
+
from ai_review.clients.openai.v1.client import get_openai_v1_http_client
|
|
2
|
+
from ai_review.clients.openai.v1.schema import OpenAIChatRequestSchema, OpenAIMessageSchema
|
|
3
|
+
from ai_review.clients.openai.v2.client import get_openai_v2_http_client
|
|
4
|
+
from ai_review.clients.openai.v2.schema import OpenAIInputMessageSchema, OpenAIResponsesRequestSchema
|
|
3
5
|
from ai_review.config import settings
|
|
4
6
|
from ai_review.services.llm.types import LLMClientProtocol, ChatResultSchema
|
|
5
7
|
|
|
6
8
|
|
|
7
9
|
class OpenAILLMClient(LLMClientProtocol):
|
|
8
10
|
def __init__(self):
|
|
9
|
-
self.
|
|
11
|
+
self.meta = settings.llm.meta
|
|
10
12
|
|
|
11
|
-
|
|
12
|
-
|
|
13
|
+
self.http_client_v1 = get_openai_v1_http_client()
|
|
14
|
+
self.http_client_v2 = get_openai_v2_http_client()
|
|
15
|
+
|
|
16
|
+
async def chat_v1(self, prompt: str, prompt_system: str) -> ChatResultSchema:
|
|
13
17
|
request = OpenAIChatRequestSchema(
|
|
14
|
-
model=meta.model,
|
|
18
|
+
model=self.meta.model,
|
|
15
19
|
messages=[
|
|
16
20
|
OpenAIMessageSchema(role="system", content=prompt_system),
|
|
17
21
|
OpenAIMessageSchema(role="user", content=prompt),
|
|
18
22
|
],
|
|
19
|
-
max_tokens=meta.max_tokens,
|
|
20
|
-
temperature=meta.temperature,
|
|
23
|
+
max_tokens=self.meta.max_tokens,
|
|
24
|
+
temperature=self.meta.temperature,
|
|
21
25
|
)
|
|
22
|
-
response = await self.
|
|
26
|
+
response = await self.http_client_v1.chat(request)
|
|
23
27
|
return ChatResultSchema(
|
|
24
28
|
text=response.first_text,
|
|
25
29
|
total_tokens=response.usage.total_tokens,
|
|
26
30
|
prompt_tokens=response.usage.prompt_tokens,
|
|
27
31
|
completion_tokens=response.usage.completion_tokens,
|
|
28
32
|
)
|
|
33
|
+
|
|
34
|
+
async def chat_v2(self, prompt: str, prompt_system: str) -> ChatResultSchema:
|
|
35
|
+
request = OpenAIResponsesRequestSchema(
|
|
36
|
+
model=self.meta.model,
|
|
37
|
+
input=[
|
|
38
|
+
OpenAIInputMessageSchema(role="system", content=prompt_system),
|
|
39
|
+
OpenAIInputMessageSchema(role="user", content=prompt),
|
|
40
|
+
],
|
|
41
|
+
temperature=self.meta.temperature,
|
|
42
|
+
max_output_tokens=self.meta.max_tokens,
|
|
43
|
+
)
|
|
44
|
+
response = await self.http_client_v2.chat(request)
|
|
45
|
+
return ChatResultSchema(
|
|
46
|
+
text=response.first_text,
|
|
47
|
+
total_tokens=response.usage.total_tokens,
|
|
48
|
+
prompt_tokens=response.usage.input_tokens,
|
|
49
|
+
completion_tokens=response.usage.output_tokens,
|
|
50
|
+
)
|
|
51
|
+
|
|
52
|
+
async def chat(self, prompt: str, prompt_system: str) -> ChatResultSchema:
|
|
53
|
+
if self.meta.is_v2_model:
|
|
54
|
+
return await self.chat_v2(prompt, prompt_system)
|
|
55
|
+
|
|
56
|
+
return await self.chat_v1(prompt, prompt_system)
|
|
File without changes
|
|
@@ -0,0 +1,31 @@
|
|
|
1
|
+
from ai_review.clients.openrouter.client import get_openrouter_http_client
|
|
2
|
+
from ai_review.clients.openrouter.schema import (
|
|
3
|
+
OpenRouterMessageSchema,
|
|
4
|
+
OpenRouterChatRequestSchema,
|
|
5
|
+
)
|
|
6
|
+
from ai_review.config import settings
|
|
7
|
+
from ai_review.services.llm.types import LLMClientProtocol, ChatResultSchema
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
class OpenRouterLLMClient(LLMClientProtocol):
|
|
11
|
+
def __init__(self):
|
|
12
|
+
self.http_client = get_openrouter_http_client()
|
|
13
|
+
|
|
14
|
+
async def chat(self, prompt: str, prompt_system: str) -> ChatResultSchema:
|
|
15
|
+
meta = settings.llm.meta
|
|
16
|
+
request = OpenRouterChatRequestSchema(
|
|
17
|
+
model=meta.model,
|
|
18
|
+
messages=[
|
|
19
|
+
OpenRouterMessageSchema(role="system", content=prompt_system),
|
|
20
|
+
OpenRouterMessageSchema(role="user", content=prompt),
|
|
21
|
+
],
|
|
22
|
+
max_tokens=meta.max_tokens,
|
|
23
|
+
temperature=meta.temperature,
|
|
24
|
+
)
|
|
25
|
+
response = await self.http_client.chat(request)
|
|
26
|
+
return ChatResultSchema(
|
|
27
|
+
text=response.first_text,
|
|
28
|
+
total_tokens=response.usage.total_tokens,
|
|
29
|
+
prompt_tokens=response.usage.prompt_tokens,
|
|
30
|
+
completion_tokens=response.usage.completion_tokens,
|
|
31
|
+
)
|