vllm-judge 0.1.4__py3-none-any.whl → 0.1.6__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
vllm_judge/__init__.py CHANGED
@@ -5,7 +5,7 @@ A lightweight library for evaluating text responses using self-hosted language m
5
5
  via vLLM's OpenAI-compatible API.
6
6
  """
7
7
 
8
- __version__ = "0.1.4"
8
+ __version__ = "0.1.6"
9
9
 
10
10
  from vllm_judge.judge import Judge
11
11
  from vllm_judge.models import (
@@ -17,17 +17,18 @@ from vllm_judge.models import (
17
17
  ModelSpecificMetric
18
18
  )
19
19
  from vllm_judge.templating import TemplateProcessor
20
- from vllm_judge.metrics import (
20
+ from vllm_judge.builtin_metrics import (
21
21
  # General metrics
22
22
  HELPFULNESS,
23
23
  ACCURACY,
24
24
  CLARITY,
25
25
  CONCISENESS,
26
26
  RELEVANCE,
27
-
27
+ COHERENCE,
28
28
  # Safety metrics
29
29
  SAFETY,
30
30
  TOXICITY,
31
+ BIAS_DETECTION,
31
32
  LLAMA_GUARD_3_SAFETY,
32
33
 
33
34
  # Code metrics
@@ -61,6 +62,12 @@ from vllm_judge.metrics import (
61
62
  PRODUCT_REVIEW_TEMPLATE,
62
63
  MEDICAL_INFO_TEMPLATE,
63
64
  API_DOCS_TEMPLATE,
65
+ RAG_EVALUATION_TEMPLATE,
66
+ AGENT_PERFORMANCE_TEMPLATE,
67
+
68
+ # NLP metrics
69
+ TRANSLATION_QUALITY,
70
+ SUMMARIZATION_QUALITY,
64
71
 
65
72
  )
66
73
  from vllm_judge.exceptions import (
@@ -91,8 +98,10 @@ __all__ = [
91
98
  "CLARITY",
92
99
  "CONCISENESS",
93
100
  "RELEVANCE",
101
+ "COHERENCE",
94
102
  "SAFETY",
95
103
  "TOXICITY",
104
+ "BIAS_DETECTION",
96
105
  "LLAMA_GUARD_3_SAFETY",
97
106
  "CODE_QUALITY",
98
107
  "CODE_SECURITY",
@@ -112,6 +121,11 @@ __all__ = [
112
121
  "PRODUCT_REVIEW_TEMPLATE",
113
122
  "MEDICAL_INFO_TEMPLATE",
114
123
  "API_DOCS_TEMPLATE",
124
+ "RAG_EVALUATION_TEMPLATE",
125
+ "AGENT_PERFORMANCE_TEMPLATE",
126
+ "TRANSLATION_QUALITY",
127
+ "SUMMARIZATION_QUALITY",
128
+
115
129
  # Exceptions
116
130
  "VLLMJudgeError",
117
131
  "ConfigurationError",
@@ -1,6 +1,3 @@
1
- """
2
- API module for vLLM Judge.
3
- """
4
1
  from vllm_judge.api.server import app, create_app, start_server
5
2
  from vllm_judge.api.client import JudgeClient
6
3
  from vllm_judge.api.models import (
vllm_judge/api/client.py CHANGED
@@ -1,6 +1,3 @@
1
- """
2
- HTTP client for vLLM Judge API.
3
- """
4
1
  import asyncio
5
2
  from typing import Union, Dict, List, Optional, Tuple, Any, AsyncIterator
6
3
  import httpx
vllm_judge/api/server.py CHANGED
@@ -1,7 +1,3 @@
1
- """
2
- FastAPI server for vLLM Judge API.
3
- """
4
- import asyncio
5
1
  import time
6
2
  import uuid
7
3
  from datetime import datetime
@@ -14,7 +10,7 @@ import uvicorn
14
10
 
15
11
  from vllm_judge.judge import Judge
16
12
  from vllm_judge.models import EvaluationResult, JudgeConfig
17
- from vllm_judge.metrics import BUILTIN_METRICS
13
+ from vllm_judge.builtin_metrics import BUILTIN_METRICS
18
14
  from vllm_judge.exceptions import VLLMJudgeError
19
15
  from vllm_judge.api.models import (
20
16
  EvaluateRequest,
vllm_judge/batch.py CHANGED
@@ -17,7 +17,8 @@ class BatchProcessor:
17
17
  max_concurrent: Maximum concurrent requests
18
18
  """
19
19
  self.judge = judge
20
- self.semaphore = asyncio.Semaphore(max_concurrent)
20
+ self.max_concurrent = max_concurrent
21
+ self.semaphore = asyncio.Semaphore(self.max_concurrent)
21
22
  self.progress_lock = asyncio.Lock()
22
23
  self.completed = 0
23
24