judgeval 0.0.16__tar.gz → 0.0.17__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {judgeval-0.0.16 → judgeval-0.0.17}/PKG-INFO +1 -1
- {judgeval-0.0.16 → judgeval-0.0.17}/pyproject.toml +1 -1
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/common/tracer.py +21 -21
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/constants.py +1 -1
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/datasets/eval_dataset_client.py +10 -5
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/judgment_client.py +8 -4
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/run_evaluation.py +6 -3
- {judgeval-0.0.16 → judgeval-0.0.17}/.github/workflows/ci.yaml +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/.gitignore +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/LICENSE.md +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/Pipfile +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/Pipfile.lock +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/README.md +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/README.md +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/api_reference/judgment_client.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/api_reference/trace.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/development.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/essentials/code.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/essentials/images.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/essentials/markdown.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/essentials/navigation.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/essentials/reusable-snippets.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/essentials/settings.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/data_datasets.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/data_examples.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/introduction.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/judges.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/answer_correctness.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/answer_relevancy.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/classifier_scorer.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/contextual_precision.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/contextual_recall.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/contextual_relevancy.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/custom_scorers.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/faithfulness.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/hallucination.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/introduction.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/json_correctness.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/summarization.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/scorers/tool_correctness.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/evaluation/unit_testing.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/favicon.svg +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/getting_started.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/basic_trace_example.png +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/checks-passed.png +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/create_aggressive_scorer.png +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/create_scorer.png +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/evaluation_diagram.png +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/hero-dark.svg +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/hero-light.svg +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/images/trace_screenshot.png +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/introduction.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/judgment/introduction.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/logo/dark.svg +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/logo/light.svg +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/mint.json +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/monitoring/introduction.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/monitoring/production_insights.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/monitoring/tracing.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/notebooks/create_dataset.ipynb +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/notebooks/create_scorer.ipynb +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/notebooks/demo.ipynb +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/notebooks/prompt_scorer.ipynb +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/notebooks/quickstart.ipynb +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/quickstart.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/docs/snippets/snippet-intro.mdx +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/pytest.ini +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/anime_chatbot_agent/animeChatBot.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/ci_testing/ci_testing.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/ci_testing/travel_response.txt +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/custom_scorers/competitor_mentions.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/custom_scorers/text2sql.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/langchain_basic_rag/basic_agentic_rag.ipynb +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/langchain_basic_rag/tesla_q3.pdf +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/langchain_sales/example_product_price_id_mapping.json +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/langchain_sales/sales_agent_with_context.ipynb +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/langchain_sales/sample_product_catalog.txt +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/new_bot/basic_bot.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/openai_travel_agent/agent.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/openai_travel_agent/populate_db.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/openai_travel_agent/tools.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/rules_alerts/rules_bot.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/rules_alerts/rules_demo.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/rules_alerts/utils_helper.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/customer_use/cstone/basic_test.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/customer_use/cstone/cstone_data.csv +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/customer_use/cstone/data.csv +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/customer_use/cstone/faithfulness_testing.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/customer_use/cstone/galen_data.csv +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/customer_use/cstone/playground.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/demo/customer_use/cstone/results.csv +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/clients.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/common/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/common/exceptions.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/common/logger.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/common/utils.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/api_example.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/datasets/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/datasets/dataset.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/datasets/ground_truth.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/datasets/utils.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/example.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/result.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/data/scorer_data.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/evaluation_run.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/judges/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/judges/base_judge.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/judges/litellm_judge.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/judges/mixture_of_judges.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/judges/together_judge.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/judges/utils.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/rules.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/api_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/base_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/exceptions.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/answer_correctness.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/answer_relevancy.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/contextual_precision.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/contextual_recall.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/contextual_relevancy.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/faithfulness.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/hallucination.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/json_correctness.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/summarization.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/tool_correctness.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/classifiers/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/classifiers/text2sql/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/classifiers/text2sql/text2sql_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_correctness/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_correctness/answer_correctness_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_correctness/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_relevancy/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_relevancy/answer_relevancy_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_relevancy/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_precision/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_precision/contextual_precision_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_precision/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_recall/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_recall/contextual_recall_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_recall/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_relevancy/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_relevancy/contextual_relevancy_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_relevancy/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/faithfulness/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/faithfulness/faithfulness_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/faithfulness/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/hallucination/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/hallucination/hallucination_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/hallucination/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/json_correctness/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/json_correctness/json_correctness_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/summarization/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/summarization/prompts.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/summarization/summarization_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/tool_correctness/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/tool_correctness/tool_correctness_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/prompt_scorer.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/score.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/utils.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/tracer/__init__.py +0 -0
- {judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/utils/alerts.py +0 -0
@@ -37,7 +37,7 @@ from http import HTTPStatus
|
|
37
37
|
import pika
|
38
38
|
import os
|
39
39
|
|
40
|
-
from judgeval.constants import JUDGMENT_TRACES_SAVE_API_URL, JUDGMENT_TRACES_FETCH_API_URL, RABBITMQ_HOST, RABBITMQ_PORT, RABBITMQ_QUEUE, JUDGMENT_TRACES_DELETE_API_URL
|
40
|
+
from judgeval.constants import JUDGMENT_TRACES_SAVE_API_URL, JUDGMENT_TRACES_FETCH_API_URL, RABBITMQ_HOST, RABBITMQ_PORT, RABBITMQ_QUEUE, JUDGMENT_TRACES_DELETE_API_URL,JUDGMENT_TRACES_ADD_TO_EVAL_QUEUE_API_URL
|
41
41
|
from judgeval.judgment_client import JudgmentClient
|
42
42
|
from judgeval.data import Example
|
43
43
|
from judgeval.scorers import APIJudgmentScorer, JudgevalScorer, ScorerWrapper
|
@@ -207,7 +207,8 @@ class TraceManagerClient:
|
|
207
207
|
"Content-Type": "application/json",
|
208
208
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
209
209
|
"X-Organization-Id": self.organization_id
|
210
|
-
}
|
210
|
+
},
|
211
|
+
verify=True
|
211
212
|
)
|
212
213
|
|
213
214
|
if response.status_code != HTTPStatus.OK:
|
@@ -231,7 +232,8 @@ class TraceManagerClient:
|
|
231
232
|
"Content-Type": "application/json",
|
232
233
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
233
234
|
"X-Organization-Id": self.organization_id
|
234
|
-
}
|
235
|
+
},
|
236
|
+
verify=True
|
235
237
|
)
|
236
238
|
|
237
239
|
if response.status_code == HTTPStatus.BAD_REQUEST:
|
@@ -617,25 +619,23 @@ class TraceClient:
|
|
617
619
|
}
|
618
620
|
# Execute asynchrous evaluation in the background
|
619
621
|
if not empty_save: # Only send to RabbitMQ if the trace is not empty
|
620
|
-
|
621
|
-
|
622
|
-
|
623
|
-
|
624
|
-
|
625
|
-
trace_data["judgment_api_key"] = self.tracer.api_key
|
626
|
-
trace_data["organization_id"] = self.tracer.organization_id
|
627
|
-
channel.basic_publish(
|
628
|
-
exchange='',
|
629
|
-
routing_key=RABBITMQ_QUEUE,
|
630
|
-
body=json.dumps(trace_data),
|
631
|
-
properties=pika.BasicProperties(
|
632
|
-
delivery_mode=pika.DeliveryMode.Transient, # Changed from Persistent to Transient
|
622
|
+
# Send trace data to evaluation queue via API
|
623
|
+
try:
|
624
|
+
response = requests.post(
|
625
|
+
JUDGMENT_TRACES_ADD_TO_EVAL_QUEUE_API_URL,
|
626
|
+
json=trace_data,
|
633
627
|
headers={
|
634
|
-
|
635
|
-
|
636
|
-
|
637
|
-
|
638
|
-
|
628
|
+
"Content-Type": "application/json",
|
629
|
+
"Authorization": f"Bearer {self.tracer.api_key}",
|
630
|
+
"X-Organization-Id": self.tracer.organization_id
|
631
|
+
},
|
632
|
+
verify=True
|
633
|
+
)
|
634
|
+
|
635
|
+
if response.status_code != HTTPStatus.OK:
|
636
|
+
warnings.warn(f"Failed to add trace to evaluation queue: {response.text}")
|
637
|
+
except Exception as e:
|
638
|
+
warnings.warn(f"Error sending trace to evaluation queue: {str(e)}")
|
639
639
|
|
640
640
|
self.trace_manager_client.save_trace(trace_data, empty_save)
|
641
641
|
|
@@ -46,7 +46,7 @@ JUDGMENT_EVAL_DELETE_PROJECT_API_URL = f"{ROOT_API}/delete_eval_results_by_proje
|
|
46
46
|
JUDGMENT_TRACES_FETCH_API_URL = f"{ROOT_API}/traces/fetch/"
|
47
47
|
JUDGMENT_TRACES_SAVE_API_URL = f"{ROOT_API}/traces/save/"
|
48
48
|
JUDGMENT_TRACES_DELETE_API_URL = f"{ROOT_API}/traces/delete/"
|
49
|
-
|
49
|
+
JUDGMENT_TRACES_ADD_TO_EVAL_QUEUE_API_URL = f"{ROOT_API}/traces/add_to_eval_queue/"
|
50
50
|
# RabbitMQ
|
51
51
|
RABBITMQ_HOST = os.getenv("RABBITMQ_HOST", "rabbitmq-networklb-faa155df16ec9085.elb.us-west-1.amazonaws.com")
|
52
52
|
RABBITMQ_PORT = os.getenv("RABBITMQ_PORT", 5672)
|
@@ -68,7 +68,8 @@ class EvalDatasetClient:
|
|
68
68
|
"Content-Type": "application/json",
|
69
69
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
70
70
|
"X-Organization-Id": self.organization_id
|
71
|
-
}
|
71
|
+
},
|
72
|
+
verify=True
|
72
73
|
)
|
73
74
|
if response.status_code == 500:
|
74
75
|
error(f"Server error during push: {content.get('message')}")
|
@@ -132,7 +133,8 @@ class EvalDatasetClient:
|
|
132
133
|
"Content-Type": "application/json",
|
133
134
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
134
135
|
"X-Organization-Id": self.organization_id
|
135
|
-
}
|
136
|
+
},
|
137
|
+
verify=True
|
136
138
|
)
|
137
139
|
response.raise_for_status()
|
138
140
|
except requests.exceptions.RequestException as e:
|
@@ -190,7 +192,8 @@ class EvalDatasetClient:
|
|
190
192
|
"Content-Type": "application/json",
|
191
193
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
192
194
|
"X-Organization-Id": self.organization_id
|
193
|
-
}
|
195
|
+
},
|
196
|
+
verify=True
|
194
197
|
)
|
195
198
|
response.raise_for_status()
|
196
199
|
except requests.exceptions.RequestException as e:
|
@@ -243,7 +246,8 @@ class EvalDatasetClient:
|
|
243
246
|
"Content-Type": "application/json",
|
244
247
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
245
248
|
"X-Organization-Id": self.organization_id
|
246
|
-
}
|
249
|
+
},
|
250
|
+
verify=True
|
247
251
|
)
|
248
252
|
response.raise_for_status()
|
249
253
|
except requests.exceptions.RequestException as e:
|
@@ -274,7 +278,8 @@ class EvalDatasetClient:
|
|
274
278
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
275
279
|
"X-Organization-Id": self.organization_id
|
276
280
|
},
|
277
|
-
stream=True
|
281
|
+
stream=True,
|
282
|
+
verify=True
|
278
283
|
)
|
279
284
|
response.raise_for_status()
|
280
285
|
except requests.exceptions.HTTPError as err:
|
@@ -306,7 +306,8 @@ class JudgmentClient:
|
|
306
306
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
307
307
|
"X-Organization-Id": self.organization_id
|
308
308
|
},
|
309
|
-
json=eval_run_request_body.model_dump()
|
309
|
+
json=eval_run_request_body.model_dump(),
|
310
|
+
verify=True)
|
310
311
|
if eval_run.status_code != requests.codes.ok:
|
311
312
|
raise ValueError(f"Error fetching eval results: {eval_run.json()}")
|
312
313
|
|
@@ -378,7 +379,8 @@ class JudgmentClient:
|
|
378
379
|
"Content-Type": "application/json",
|
379
380
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
380
381
|
},
|
381
|
-
json={} # Empty body now
|
382
|
+
json={}, # Empty body now
|
383
|
+
verify=True
|
382
384
|
)
|
383
385
|
if response.status_code == 200:
|
384
386
|
return True, response.json()
|
@@ -409,7 +411,8 @@ class JudgmentClient:
|
|
409
411
|
"Content-Type": "application/json",
|
410
412
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
411
413
|
"X-Organization-Id": self.organization_id
|
412
|
-
}
|
414
|
+
},
|
415
|
+
verify=True
|
413
416
|
)
|
414
417
|
|
415
418
|
if response.status_code == 500:
|
@@ -452,7 +455,8 @@ class JudgmentClient:
|
|
452
455
|
"Content-Type": "application/json",
|
453
456
|
"Authorization": f"Bearer {self.judgment_api_key}",
|
454
457
|
"X-Organization-Id": self.organization_id
|
455
|
-
}
|
458
|
+
},
|
459
|
+
verify=True
|
456
460
|
)
|
457
461
|
|
458
462
|
if response.status_code == 500:
|
@@ -55,7 +55,8 @@ def execute_api_eval(evaluation_run: EvaluationRun) -> List[Dict]:
|
|
55
55
|
"Authorization": f"Bearer {evaluation_run.judgment_api_key}",
|
56
56
|
"X-Organization-Id": evaluation_run.organization_id
|
57
57
|
},
|
58
|
-
json=payload
|
58
|
+
json=payload,
|
59
|
+
verify=True)
|
59
60
|
response_data = response.json()
|
60
61
|
except Exception as e:
|
61
62
|
error(f"Error: {e}")
|
@@ -168,7 +169,8 @@ def check_eval_run_name_exists(eval_name: str, project_name: str, judgment_api_k
|
|
168
169
|
"eval_name": eval_name,
|
169
170
|
"project_name": project_name,
|
170
171
|
"judgment_api_key": judgment_api_key,
|
171
|
-
}
|
172
|
+
},
|
173
|
+
verify=True
|
172
174
|
)
|
173
175
|
|
174
176
|
if response.status_code == 409:
|
@@ -210,7 +212,8 @@ def log_evaluation_results(merged_results: List[ScoringResult], evaluation_run:
|
|
210
212
|
"results": [result.to_dict() for result in merged_results],
|
211
213
|
"project_name": evaluation_run.project_name,
|
212
214
|
"eval_name": evaluation_run.eval_name,
|
213
|
-
}
|
215
|
+
},
|
216
|
+
verify=True
|
214
217
|
)
|
215
218
|
|
216
219
|
if not res.ok:
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
{judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/custom_scorers/competitor_mentions.py
RENAMED
File without changes
|
File without changes
|
{judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/langchain_basic_rag/basic_agentic_rag.ipynb
RENAMED
File without changes
|
File without changes
|
File without changes
|
File without changes
|
{judgeval-0.0.16 → judgeval-0.0.17}/src/demo/cookbooks/langchain_sales/sample_product_catalog.txt
RENAMED
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
{judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/__init__.py
RENAMED
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
{judgeval-0.0.16 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/classifiers/__init__.py
RENAMED
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|