judgeval 0.0.15__tar.gz → 0.0.17__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (168) hide show
  1. {judgeval-0.0.15 → judgeval-0.0.17}/PKG-INFO +1 -1
  2. {judgeval-0.0.15 → judgeval-0.0.17}/pyproject.toml +1 -1
  3. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/common/tracer.py +19 -21
  4. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/constants.py +1 -1
  5. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/datasets/eval_dataset_client.py +5 -5
  6. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/judgment_client.py +4 -4
  7. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/run_evaluation.py +3 -3
  8. {judgeval-0.0.15 → judgeval-0.0.17}/.github/workflows/ci.yaml +0 -0
  9. {judgeval-0.0.15 → judgeval-0.0.17}/.gitignore +0 -0
  10. {judgeval-0.0.15 → judgeval-0.0.17}/LICENSE.md +0 -0
  11. {judgeval-0.0.15 → judgeval-0.0.17}/Pipfile +0 -0
  12. {judgeval-0.0.15 → judgeval-0.0.17}/Pipfile.lock +0 -0
  13. {judgeval-0.0.15 → judgeval-0.0.17}/README.md +0 -0
  14. {judgeval-0.0.15 → judgeval-0.0.17}/docs/README.md +0 -0
  15. {judgeval-0.0.15 → judgeval-0.0.17}/docs/api_reference/judgment_client.mdx +0 -0
  16. {judgeval-0.0.15 → judgeval-0.0.17}/docs/api_reference/trace.mdx +0 -0
  17. {judgeval-0.0.15 → judgeval-0.0.17}/docs/development.mdx +0 -0
  18. {judgeval-0.0.15 → judgeval-0.0.17}/docs/essentials/code.mdx +0 -0
  19. {judgeval-0.0.15 → judgeval-0.0.17}/docs/essentials/images.mdx +0 -0
  20. {judgeval-0.0.15 → judgeval-0.0.17}/docs/essentials/markdown.mdx +0 -0
  21. {judgeval-0.0.15 → judgeval-0.0.17}/docs/essentials/navigation.mdx +0 -0
  22. {judgeval-0.0.15 → judgeval-0.0.17}/docs/essentials/reusable-snippets.mdx +0 -0
  23. {judgeval-0.0.15 → judgeval-0.0.17}/docs/essentials/settings.mdx +0 -0
  24. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/data_datasets.mdx +0 -0
  25. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/data_examples.mdx +0 -0
  26. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/introduction.mdx +0 -0
  27. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/judges.mdx +0 -0
  28. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/answer_correctness.mdx +0 -0
  29. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/answer_relevancy.mdx +0 -0
  30. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/classifier_scorer.mdx +0 -0
  31. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/contextual_precision.mdx +0 -0
  32. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/contextual_recall.mdx +0 -0
  33. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/contextual_relevancy.mdx +0 -0
  34. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/custom_scorers.mdx +0 -0
  35. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/faithfulness.mdx +0 -0
  36. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/hallucination.mdx +0 -0
  37. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/introduction.mdx +0 -0
  38. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/json_correctness.mdx +0 -0
  39. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/summarization.mdx +0 -0
  40. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/scorers/tool_correctness.mdx +0 -0
  41. {judgeval-0.0.15 → judgeval-0.0.17}/docs/evaluation/unit_testing.mdx +0 -0
  42. {judgeval-0.0.15 → judgeval-0.0.17}/docs/favicon.svg +0 -0
  43. {judgeval-0.0.15 → judgeval-0.0.17}/docs/getting_started.mdx +0 -0
  44. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/basic_trace_example.png +0 -0
  45. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/checks-passed.png +0 -0
  46. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/create_aggressive_scorer.png +0 -0
  47. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/create_scorer.png +0 -0
  48. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/evaluation_diagram.png +0 -0
  49. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/hero-dark.svg +0 -0
  50. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/hero-light.svg +0 -0
  51. {judgeval-0.0.15 → judgeval-0.0.17}/docs/images/trace_screenshot.png +0 -0
  52. {judgeval-0.0.15 → judgeval-0.0.17}/docs/introduction.mdx +0 -0
  53. {judgeval-0.0.15 → judgeval-0.0.17}/docs/judgment/introduction.mdx +0 -0
  54. {judgeval-0.0.15 → judgeval-0.0.17}/docs/logo/dark.svg +0 -0
  55. {judgeval-0.0.15 → judgeval-0.0.17}/docs/logo/light.svg +0 -0
  56. {judgeval-0.0.15 → judgeval-0.0.17}/docs/mint.json +0 -0
  57. {judgeval-0.0.15 → judgeval-0.0.17}/docs/monitoring/introduction.mdx +0 -0
  58. {judgeval-0.0.15 → judgeval-0.0.17}/docs/monitoring/production_insights.mdx +0 -0
  59. {judgeval-0.0.15 → judgeval-0.0.17}/docs/monitoring/tracing.mdx +0 -0
  60. {judgeval-0.0.15 → judgeval-0.0.17}/docs/notebooks/create_dataset.ipynb +0 -0
  61. {judgeval-0.0.15 → judgeval-0.0.17}/docs/notebooks/create_scorer.ipynb +0 -0
  62. {judgeval-0.0.15 → judgeval-0.0.17}/docs/notebooks/demo.ipynb +0 -0
  63. {judgeval-0.0.15 → judgeval-0.0.17}/docs/notebooks/prompt_scorer.ipynb +0 -0
  64. {judgeval-0.0.15 → judgeval-0.0.17}/docs/notebooks/quickstart.ipynb +0 -0
  65. {judgeval-0.0.15 → judgeval-0.0.17}/docs/quickstart.mdx +0 -0
  66. {judgeval-0.0.15 → judgeval-0.0.17}/docs/snippets/snippet-intro.mdx +0 -0
  67. {judgeval-0.0.15 → judgeval-0.0.17}/pytest.ini +0 -0
  68. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/anime_chatbot_agent/animeChatBot.py +0 -0
  69. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/ci_testing/ci_testing.py +0 -0
  70. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/ci_testing/travel_response.txt +0 -0
  71. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/custom_scorers/competitor_mentions.py +0 -0
  72. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/custom_scorers/text2sql.py +0 -0
  73. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/langchain_basic_rag/basic_agentic_rag.ipynb +0 -0
  74. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/langchain_basic_rag/tesla_q3.pdf +0 -0
  75. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/langchain_sales/example_product_price_id_mapping.json +0 -0
  76. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/langchain_sales/sales_agent_with_context.ipynb +0 -0
  77. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/langchain_sales/sample_product_catalog.txt +0 -0
  78. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/new_bot/basic_bot.py +0 -0
  79. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/openai_travel_agent/agent.py +0 -0
  80. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/openai_travel_agent/populate_db.py +0 -0
  81. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/openai_travel_agent/tools.py +0 -0
  82. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/rules_alerts/rules_bot.py +0 -0
  83. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/rules_alerts/rules_demo.py +0 -0
  84. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/cookbooks/rules_alerts/utils_helper.py +0 -0
  85. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/customer_use/cstone/basic_test.py +0 -0
  86. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/customer_use/cstone/cstone_data.csv +0 -0
  87. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/customer_use/cstone/data.csv +0 -0
  88. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/customer_use/cstone/faithfulness_testing.py +0 -0
  89. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/customer_use/cstone/galen_data.csv +0 -0
  90. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/customer_use/cstone/playground.py +0 -0
  91. {judgeval-0.0.15 → judgeval-0.0.17}/src/demo/customer_use/cstone/results.csv +0 -0
  92. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/__init__.py +0 -0
  93. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/clients.py +0 -0
  94. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/common/__init__.py +0 -0
  95. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/common/exceptions.py +0 -0
  96. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/common/logger.py +0 -0
  97. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/common/utils.py +0 -0
  98. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/__init__.py +0 -0
  99. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/api_example.py +0 -0
  100. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/datasets/__init__.py +0 -0
  101. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/datasets/dataset.py +0 -0
  102. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/datasets/ground_truth.py +0 -0
  103. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/datasets/utils.py +0 -0
  104. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/example.py +0 -0
  105. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/result.py +0 -0
  106. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/data/scorer_data.py +0 -0
  107. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/evaluation_run.py +0 -0
  108. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/judges/__init__.py +0 -0
  109. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/judges/base_judge.py +0 -0
  110. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/judges/litellm_judge.py +0 -0
  111. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/judges/mixture_of_judges.py +0 -0
  112. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/judges/together_judge.py +0 -0
  113. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/judges/utils.py +0 -0
  114. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/rules.py +0 -0
  115. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/__init__.py +0 -0
  116. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/api_scorer.py +0 -0
  117. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/base_scorer.py +0 -0
  118. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/exceptions.py +0 -0
  119. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorer.py +0 -0
  120. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/__init__.py +0 -0
  121. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/__init__.py +0 -0
  122. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/answer_correctness.py +0 -0
  123. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/answer_relevancy.py +0 -0
  124. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/contextual_precision.py +0 -0
  125. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/contextual_recall.py +0 -0
  126. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/contextual_relevancy.py +0 -0
  127. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/faithfulness.py +0 -0
  128. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/hallucination.py +0 -0
  129. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/json_correctness.py +0 -0
  130. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/summarization.py +0 -0
  131. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/api_scorers/tool_correctness.py +0 -0
  132. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/classifiers/__init__.py +0 -0
  133. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/classifiers/text2sql/__init__.py +0 -0
  134. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/classifiers/text2sql/text2sql_scorer.py +0 -0
  135. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/__init__.py +0 -0
  136. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_correctness/__init__.py +0 -0
  137. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_correctness/answer_correctness_scorer.py +0 -0
  138. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_correctness/prompts.py +0 -0
  139. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_relevancy/__init__.py +0 -0
  140. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_relevancy/answer_relevancy_scorer.py +0 -0
  141. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/answer_relevancy/prompts.py +0 -0
  142. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_precision/__init__.py +0 -0
  143. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_precision/contextual_precision_scorer.py +0 -0
  144. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_precision/prompts.py +0 -0
  145. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_recall/__init__.py +0 -0
  146. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_recall/contextual_recall_scorer.py +0 -0
  147. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_recall/prompts.py +0 -0
  148. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_relevancy/__init__.py +0 -0
  149. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_relevancy/contextual_relevancy_scorer.py +0 -0
  150. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/contextual_relevancy/prompts.py +0 -0
  151. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/faithfulness/__init__.py +0 -0
  152. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/faithfulness/faithfulness_scorer.py +0 -0
  153. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/faithfulness/prompts.py +0 -0
  154. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/hallucination/__init__.py +0 -0
  155. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/hallucination/hallucination_scorer.py +0 -0
  156. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/hallucination/prompts.py +0 -0
  157. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/json_correctness/__init__.py +0 -0
  158. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/json_correctness/json_correctness_scorer.py +0 -0
  159. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/summarization/__init__.py +0 -0
  160. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/summarization/prompts.py +0 -0
  161. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/summarization/summarization_scorer.py +0 -0
  162. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/tool_correctness/__init__.py +0 -0
  163. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/judgeval_scorers/local_implementations/tool_correctness/tool_correctness_scorer.py +0 -0
  164. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/prompt_scorer.py +0 -0
  165. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/score.py +0 -0
  166. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/scorers/utils.py +0 -0
  167. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/tracer/__init__.py +0 -0
  168. {judgeval-0.0.15 → judgeval-0.0.17}/src/judgeval/utils/alerts.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: judgeval
3
- Version: 0.0.15
3
+ Version: 0.0.17
4
4
  Summary: Judgeval Package
5
5
  Project-URL: Homepage, https://github.com/JudgmentLabs/judgeval
6
6
  Project-URL: Issues, https://github.com/JudgmentLabs/judgeval/issues
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "judgeval"
3
- version = "0.0.15"
3
+ version = "0.0.17"
4
4
  authors = [
5
5
  { name="Andrew Li", email="andrew@judgmentlabs.ai" },
6
6
  { name="Alex Shan", email="alex@judgmentlabs.ai" },
@@ -37,7 +37,7 @@ from http import HTTPStatus
37
37
  import pika
38
38
  import os
39
39
 
40
- from judgeval.constants import JUDGMENT_TRACES_SAVE_API_URL, JUDGMENT_TRACES_FETCH_API_URL, RABBITMQ_HOST, RABBITMQ_PORT, RABBITMQ_QUEUE, JUDGMENT_TRACES_DELETE_API_URL
40
+ from judgeval.constants import JUDGMENT_TRACES_SAVE_API_URL, JUDGMENT_TRACES_FETCH_API_URL, RABBITMQ_HOST, RABBITMQ_PORT, RABBITMQ_QUEUE, JUDGMENT_TRACES_DELETE_API_URL,JUDGMENT_TRACES_ADD_TO_EVAL_QUEUE_API_URL
41
41
  from judgeval.judgment_client import JudgmentClient
42
42
  from judgeval.data import Example
43
43
  from judgeval.scorers import APIJudgmentScorer, JudgevalScorer, ScorerWrapper
@@ -208,7 +208,7 @@ class TraceManagerClient:
208
208
  "Authorization": f"Bearer {self.judgment_api_key}",
209
209
  "X-Organization-Id": self.organization_id
210
210
  },
211
- verify=False
211
+ verify=True
212
212
  )
213
213
 
214
214
  if response.status_code != HTTPStatus.OK:
@@ -233,7 +233,7 @@ class TraceManagerClient:
233
233
  "Authorization": f"Bearer {self.judgment_api_key}",
234
234
  "X-Organization-Id": self.organization_id
235
235
  },
236
- verify=False
236
+ verify=True
237
237
  )
238
238
 
239
239
  if response.status_code == HTTPStatus.BAD_REQUEST:
@@ -619,25 +619,23 @@ class TraceClient:
619
619
  }
620
620
  # Execute asynchrous evaluation in the background
621
621
  if not empty_save: # Only send to RabbitMQ if the trace is not empty
622
- connection = pika.BlockingConnection(
623
- pika.ConnectionParameters(host=RABBITMQ_HOST, port=RABBITMQ_PORT))
624
- channel = connection.channel()
625
-
626
- channel.queue_declare(queue=RABBITMQ_QUEUE, durable=True)
627
- trace_data["judgment_api_key"] = self.tracer.api_key
628
- trace_data["organization_id"] = self.tracer.organization_id
629
- channel.basic_publish(
630
- exchange='',
631
- routing_key=RABBITMQ_QUEUE,
632
- body=json.dumps(trace_data),
633
- properties=pika.BasicProperties(
634
- delivery_mode=pika.DeliveryMode.Transient, # Changed from Persistent to Transient
622
+ # Send trace data to evaluation queue via API
623
+ try:
624
+ response = requests.post(
625
+ JUDGMENT_TRACES_ADD_TO_EVAL_QUEUE_API_URL,
626
+ json=trace_data,
635
627
  headers={
636
- 'api_key': self.tracer.api_key,
637
- 'organization_id': self.tracer.organization_id
638
- }
639
- ))
640
- connection.close()
628
+ "Content-Type": "application/json",
629
+ "Authorization": f"Bearer {self.tracer.api_key}",
630
+ "X-Organization-Id": self.tracer.organization_id
631
+ },
632
+ verify=True
633
+ )
634
+
635
+ if response.status_code != HTTPStatus.OK:
636
+ warnings.warn(f"Failed to add trace to evaluation queue: {response.text}")
637
+ except Exception as e:
638
+ warnings.warn(f"Error sending trace to evaluation queue: {str(e)}")
641
639
 
642
640
  self.trace_manager_client.save_trace(trace_data, empty_save)
643
641
 
@@ -46,7 +46,7 @@ JUDGMENT_EVAL_DELETE_PROJECT_API_URL = f"{ROOT_API}/delete_eval_results_by_proje
46
46
  JUDGMENT_TRACES_FETCH_API_URL = f"{ROOT_API}/traces/fetch/"
47
47
  JUDGMENT_TRACES_SAVE_API_URL = f"{ROOT_API}/traces/save/"
48
48
  JUDGMENT_TRACES_DELETE_API_URL = f"{ROOT_API}/traces/delete/"
49
-
49
+ JUDGMENT_TRACES_ADD_TO_EVAL_QUEUE_API_URL = f"{ROOT_API}/traces/add_to_eval_queue/"
50
50
  # RabbitMQ
51
51
  RABBITMQ_HOST = os.getenv("RABBITMQ_HOST", "rabbitmq-networklb-faa155df16ec9085.elb.us-west-1.amazonaws.com")
52
52
  RABBITMQ_PORT = os.getenv("RABBITMQ_PORT", 5672)
@@ -69,7 +69,7 @@ class EvalDatasetClient:
69
69
  "Authorization": f"Bearer {self.judgment_api_key}",
70
70
  "X-Organization-Id": self.organization_id
71
71
  },
72
- verify=False
72
+ verify=True
73
73
  )
74
74
  if response.status_code == 500:
75
75
  error(f"Server error during push: {content.get('message')}")
@@ -134,7 +134,7 @@ class EvalDatasetClient:
134
134
  "Authorization": f"Bearer {self.judgment_api_key}",
135
135
  "X-Organization-Id": self.organization_id
136
136
  },
137
- verify=False
137
+ verify=True
138
138
  )
139
139
  response.raise_for_status()
140
140
  except requests.exceptions.RequestException as e:
@@ -193,7 +193,7 @@ class EvalDatasetClient:
193
193
  "Authorization": f"Bearer {self.judgment_api_key}",
194
194
  "X-Organization-Id": self.organization_id
195
195
  },
196
- verify=False
196
+ verify=True
197
197
  )
198
198
  response.raise_for_status()
199
199
  except requests.exceptions.RequestException as e:
@@ -247,7 +247,7 @@ class EvalDatasetClient:
247
247
  "Authorization": f"Bearer {self.judgment_api_key}",
248
248
  "X-Organization-Id": self.organization_id
249
249
  },
250
- verify=False
250
+ verify=True
251
251
  )
252
252
  response.raise_for_status()
253
253
  except requests.exceptions.RequestException as e:
@@ -279,7 +279,7 @@ class EvalDatasetClient:
279
279
  "X-Organization-Id": self.organization_id
280
280
  },
281
281
  stream=True,
282
- verify=False
282
+ verify=True
283
283
  )
284
284
  response.raise_for_status()
285
285
  except requests.exceptions.HTTPError as err:
@@ -307,7 +307,7 @@ class JudgmentClient:
307
307
  "X-Organization-Id": self.organization_id
308
308
  },
309
309
  json=eval_run_request_body.model_dump(),
310
- verify=False)
310
+ verify=True)
311
311
  if eval_run.status_code != requests.codes.ok:
312
312
  raise ValueError(f"Error fetching eval results: {eval_run.json()}")
313
313
 
@@ -380,7 +380,7 @@ class JudgmentClient:
380
380
  "Authorization": f"Bearer {self.judgment_api_key}",
381
381
  },
382
382
  json={}, # Empty body now
383
- verify=False
383
+ verify=True
384
384
  )
385
385
  if response.status_code == 200:
386
386
  return True, response.json()
@@ -412,7 +412,7 @@ class JudgmentClient:
412
412
  "Authorization": f"Bearer {self.judgment_api_key}",
413
413
  "X-Organization-Id": self.organization_id
414
414
  },
415
- verify=False
415
+ verify=True
416
416
  )
417
417
 
418
418
  if response.status_code == 500:
@@ -456,7 +456,7 @@ class JudgmentClient:
456
456
  "Authorization": f"Bearer {self.judgment_api_key}",
457
457
  "X-Organization-Id": self.organization_id
458
458
  },
459
- verify=False
459
+ verify=True
460
460
  )
461
461
 
462
462
  if response.status_code == 500:
@@ -56,7 +56,7 @@ def execute_api_eval(evaluation_run: EvaluationRun) -> List[Dict]:
56
56
  "X-Organization-Id": evaluation_run.organization_id
57
57
  },
58
58
  json=payload,
59
- verify=False)
59
+ verify=True)
60
60
  response_data = response.json()
61
61
  except Exception as e:
62
62
  error(f"Error: {e}")
@@ -170,7 +170,7 @@ def check_eval_run_name_exists(eval_name: str, project_name: str, judgment_api_k
170
170
  "project_name": project_name,
171
171
  "judgment_api_key": judgment_api_key,
172
172
  },
173
- verify=False
173
+ verify=True
174
174
  )
175
175
 
176
176
  if response.status_code == 409:
@@ -213,7 +213,7 @@ def log_evaluation_results(merged_results: List[ScoringResult], evaluation_run:
213
213
  "project_name": evaluation_run.project_name,
214
214
  "eval_name": evaluation_run.eval_name,
215
215
  },
216
- verify=False
216
+ verify=True
217
217
  )
218
218
 
219
219
  if not res.ok:
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes
File without changes