fleet-python 0.2.115__tar.gz → 0.2.117__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (126) hide show
  1. {fleet_python-0.2.115/fleet_python.egg-info → fleet_python-0.2.117}/PKG-INFO +1 -1
  2. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/__init__.py +5 -1
  3. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/__init__.py +1 -1
  4. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/base.py +1 -1
  5. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/client.py +39 -3
  6. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/base.py +1 -1
  7. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/client.py +36 -3
  8. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/judge.py +54 -0
  9. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/__init__.py +3 -0
  10. fleet_python-0.2.117/fleet/verifiers/local_executor.py +247 -0
  11. {fleet_python-0.2.115 → fleet_python-0.2.117/fleet_python.egg-info}/PKG-INFO +1 -1
  12. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet_python.egg-info/SOURCES.txt +2 -0
  13. {fleet_python-0.2.115 → fleet_python-0.2.117}/pyproject.toml +1 -1
  14. fleet_python-0.2.117/tests/test_judge_criteria_markers.py +192 -0
  15. {fleet_python-0.2.115 → fleet_python-0.2.117}/LICENSE +0 -0
  16. {fleet_python-0.2.115 → fleet_python-0.2.117}/README.md +0 -0
  17. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/diff_example.py +0 -0
  18. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/dsl_example.py +0 -0
  19. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example.py +0 -0
  20. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/exampleResume.py +0 -0
  21. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_account.py +0 -0
  22. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_action_log.py +0 -0
  23. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_client.py +0 -0
  24. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_mcp_anthropic.py +0 -0
  25. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_mcp_openai.py +0 -0
  26. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_sync.py +0 -0
  27. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_task.py +0 -0
  28. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_tasks.py +0 -0
  29. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/example_verifier.py +0 -0
  30. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/export_tasks.py +0 -0
  31. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/export_tasks_filtered.py +0 -0
  32. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/fetch_tasks.py +0 -0
  33. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/gemini_example.py +0 -0
  34. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/import_tasks.py +0 -0
  35. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/iterate_verifiers.py +0 -0
  36. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/json_tasks_example.py +0 -0
  37. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/nova_act_example.py +0 -0
  38. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/openai_example.py +0 -0
  39. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/openai_simple_example.py +0 -0
  40. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/query_builder_example.py +0 -0
  41. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/quickstart.py +0 -0
  42. {fleet_python-0.2.115 → fleet_python-0.2.117}/examples/test_cdp_logging.py +0 -0
  43. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/env/__init__.py +0 -0
  44. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/env/client.py +0 -0
  45. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/exceptions.py +0 -0
  46. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/global_client.py +0 -0
  47. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/instance/__init__.py +0 -0
  48. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/instance/base.py +0 -0
  49. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/instance/client.py +0 -0
  50. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/judge.py +0 -0
  51. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/models.py +0 -0
  52. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/resources/__init__.py +0 -0
  53. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/resources/api.py +0 -0
  54. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/resources/base.py +0 -0
  55. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/resources/browser.py +0 -0
  56. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/resources/filesystem.py +0 -0
  57. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/resources/mcp.py +0 -0
  58. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/resources/sqlite.py +0 -0
  59. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/tasks.py +0 -0
  60. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/verifiers/__init__.py +0 -0
  61. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/verifiers/bundler.py +0 -0
  62. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/_async/verifiers/verifier.py +0 -0
  63. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/__init__.py +0 -0
  64. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/Dockerfile +0 -0
  65. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/__init__.py +0 -0
  66. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/agent.py +0 -0
  67. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/mcp/main.py +0 -0
  68. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/mcp_server/__init__.py +0 -0
  69. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/mcp_server/main.py +0 -0
  70. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/mcp_server/tools.py +0 -0
  71. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/requirements.txt +0 -0
  72. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/gemini_cua/start.sh +0 -0
  73. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/orchestrator.py +0 -0
  74. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/types.py +0 -0
  75. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/agent/utils.py +0 -0
  76. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/cli.py +0 -0
  77. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/config.py +0 -0
  78. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/env/__init__.py +0 -0
  79. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/env/client.py +0 -0
  80. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/eval/__init__.py +0 -0
  81. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/eval/uploader.py +0 -0
  82. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/exceptions.py +0 -0
  83. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/global_client.py +0 -0
  84. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/instance/__init__.py +0 -0
  85. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/instance/base.py +0 -0
  86. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/instance/client.py +0 -0
  87. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/instance/models.py +0 -0
  88. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/models.py +0 -0
  89. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/proxy/__init__.py +0 -0
  90. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/proxy/proxy.py +0 -0
  91. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/proxy/whitelist.py +0 -0
  92. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/resources/__init__.py +0 -0
  93. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/resources/api.py +0 -0
  94. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/resources/base.py +0 -0
  95. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/resources/browser.py +0 -0
  96. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/resources/filesystem.py +0 -0
  97. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/resources/mcp.py +0 -0
  98. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/resources/sqlite.py +0 -0
  99. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/tasks.py +0 -0
  100. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/types.py +0 -0
  101. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/utils/__init__.py +0 -0
  102. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/utils/http_logging.py +0 -0
  103. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/utils/logging.py +0 -0
  104. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/utils/playwright.py +0 -0
  105. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/bundler.py +0 -0
  106. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/code.py +0 -0
  107. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/db.py +0 -0
  108. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/decorator.py +0 -0
  109. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/parse.py +0 -0
  110. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/sql_differ.py +0 -0
  111. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet/verifiers/verifier.py +0 -0
  112. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet_python.egg-info/dependency_links.txt +0 -0
  113. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet_python.egg-info/entry_points.txt +0 -0
  114. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet_python.egg-info/requires.txt +0 -0
  115. {fleet_python-0.2.115 → fleet_python-0.2.117}/fleet_python.egg-info/top_level.txt +0 -0
  116. {fleet_python-0.2.115 → fleet_python-0.2.117}/scripts/fix_sync_imports.py +0 -0
  117. {fleet_python-0.2.115 → fleet_python-0.2.117}/scripts/unasync.py +0 -0
  118. {fleet_python-0.2.115 → fleet_python-0.2.117}/setup.cfg +0 -0
  119. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/__init__.py +0 -0
  120. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/test_app_method.py +0 -0
  121. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/test_expect_exactly.py +0 -0
  122. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/test_expect_only.py +0 -0
  123. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/test_instance_dispatch.py +0 -0
  124. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/test_sqlite_resource_dual_mode.py +0 -0
  125. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/test_sqlite_shared_memory_behavior.py +0 -0
  126. {fleet_python-0.2.115 → fleet_python-0.2.117}/tests/test_verifier_from_string.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: fleet-python
3
- Version: 0.2.115
3
+ Version: 0.2.117
4
4
  Summary: Python SDK for Fleet environments
5
5
  Author-email: Fleet AI <nic@fleet.so>
6
6
  License: Apache-2.0
@@ -38,6 +38,8 @@ from .verifiers import (
38
38
  SnapshotDiff,
39
39
  TASK_FAILED_SCORE,
40
40
  TASK_SUCCESSFUL_SCORE,
41
+ execute_verifier_local,
42
+ LocalEnvironment,
41
43
  )
42
44
 
43
45
  # Import async verifiers (default verifier is async for modern usage)
@@ -76,7 +78,7 @@ from . import env
76
78
  from . import global_client as _global_client
77
79
  from ._async import global_client as _async_global_client
78
80
 
79
- __version__ = "0.2.115"
81
+ __version__ = "0.2.117"
80
82
 
81
83
  __all__ = [
82
84
  # Core classes
@@ -114,6 +116,8 @@ __all__ = [
114
116
  "SnapshotDiff",
115
117
  "TASK_FAILED_SCORE",
116
118
  "TASK_SUCCESSFUL_SCORE",
119
+ "execute_verifier_local",
120
+ "LocalEnvironment",
117
121
  # Environment module
118
122
  "env",
119
123
  # Global client helpers
@@ -44,7 +44,7 @@ from ..types import VerifierFunction
44
44
  from .. import env
45
45
  from . import global_client as _async_global_client
46
46
 
47
- __version__ = "0.2.115"
47
+ __version__ = "0.2.117"
48
48
 
49
49
  __all__ = [
50
50
  # Core classes
@@ -26,7 +26,7 @@ from .exceptions import (
26
26
  try:
27
27
  from .. import __version__
28
28
  except ImportError:
29
- __version__ = "0.2.115"
29
+ __version__ = "0.2.117"
30
30
 
31
31
  logger = logging.getLogger(__name__)
32
32
 
@@ -601,7 +601,6 @@ class AsyncFleet:
601
601
  )
602
602
 
603
603
  instance = AsyncEnv(client=self.client, **response.json())
604
- await instance.instance.load()
605
604
  return instance
606
605
 
607
606
  async def make_for_task(self, task: Task) -> AsyncEnv:
@@ -653,7 +652,6 @@ class AsyncFleet:
653
652
  else:
654
653
  response = await self.client.request("GET", f"/v1/env/instances/{instance_id}")
655
654
  instance = AsyncEnv(client=self.client, **response.json())
656
- await instance.instance.load()
657
655
  return instance
658
656
 
659
657
  def _create_url_instance(self, base_url: str) -> AsyncEnv:
@@ -834,7 +832,45 @@ class AsyncFleet:
834
832
  At least one of run_id or profile_id must be provided.
835
833
  """
836
834
  return await _delete_instances_batch(self.client, run_id=run_id, profile_id=profile_id)
837
-
835
+
836
+ @staticmethod
837
+ async def execute_verifier_local(
838
+ verifier_func: str,
839
+ seed_db: str,
840
+ current_db: str,
841
+ final_answer: Optional[str] = None,
842
+ ) -> Dict[str, Any]:
843
+ """Execute a verifier function locally against SQLite database files.
844
+
845
+ No authentication or remote server required. The verifier code is executed
846
+ in an isolated namespace with the same helpers available in production
847
+ (``normalized_contains``, ``IgnoreConfig``, ``DatabaseSnapshot``, etc.).
848
+
849
+ Args:
850
+ verifier_func: Python source code containing the verifier function definition.
851
+ seed_db: Path to the seed (before) SQLite database file.
852
+ current_db: Path to the current (after) SQLite database file.
853
+ final_answer: Optional final answer string passed to the verifier.
854
+
855
+ Returns:
856
+ Dict with keys ``success``, ``result``, ``error``, and ``stdout``.
857
+
858
+ Example::
859
+
860
+ result = await fleet.execute_verifier_local(
861
+ verifier_func=verifier_code_string,
862
+ seed_db="./seed.db",
863
+ current_db="./current.db",
864
+ )
865
+ print(result["result"]) # 1 (TASK_SUCCESSFUL_SCORE) or 0
866
+ """
867
+ import asyncio
868
+ from ..verifiers.local_executor import execute_verifier_local
869
+
870
+ return await asyncio.to_thread(
871
+ execute_verifier_local, verifier_func, seed_db, current_db, final_answer
872
+ )
873
+
838
874
  async def list_runs(
839
875
  self, profile_id: Optional[str] = None, status: Optional[str] = "active"
840
876
  ) -> List[Run]:
@@ -27,7 +27,7 @@ from .exceptions import (
27
27
  try:
28
28
  from . import __version__
29
29
  except ImportError:
30
- __version__ = "0.2.115"
30
+ __version__ = "0.2.117"
31
31
 
32
32
  logger = logging.getLogger(__name__)
33
33
 
@@ -613,7 +613,6 @@ class Fleet:
613
613
  )
614
614
 
615
615
  instance = SyncEnv(client=self.client, **response.json())
616
- instance.instance.load()
617
616
  return instance
618
617
 
619
618
  def make_for_task(self, task: Task) -> SyncEnv:
@@ -665,7 +664,6 @@ class Fleet:
665
664
  else:
666
665
  response = self.client.request("GET", f"/v1/env/instances/{instance_id}")
667
666
  instance = SyncEnv(client=self.client, **response.json())
668
- instance.instance.load()
669
667
  return instance
670
668
 
671
669
  def _create_url_instance(self, base_url: str) -> SyncEnv:
@@ -846,7 +844,42 @@ class Fleet:
846
844
  At least one of run_id or profile_id must be provided.
847
845
  """
848
846
  return _delete_instances_batch(self.client, run_id=run_id, profile_id=profile_id)
849
-
847
+
848
+ @staticmethod
849
+ def execute_verifier_local(
850
+ verifier_func: str,
851
+ seed_db: str,
852
+ current_db: str,
853
+ final_answer: Optional[str] = None,
854
+ ) -> Dict[str, Any]:
855
+ """Execute a verifier function locally against SQLite database files.
856
+
857
+ No authentication or remote server required. The verifier code is executed
858
+ in an isolated namespace with the same helpers available in production
859
+ (``normalized_contains``, ``IgnoreConfig``, ``DatabaseSnapshot``, etc.).
860
+
861
+ Args:
862
+ verifier_func: Python source code containing the verifier function definition.
863
+ seed_db: Path to the seed (before) SQLite database file.
864
+ current_db: Path to the current (after) SQLite database file.
865
+ final_answer: Optional final answer string passed to the verifier.
866
+
867
+ Returns:
868
+ Dict with keys ``success``, ``result``, ``error``, and ``stdout``.
869
+
870
+ Example::
871
+
872
+ result = fleet.execute_verifier_local(
873
+ verifier_func=verifier_code_string,
874
+ seed_db="./seed.db",
875
+ current_db="./current.db",
876
+ )
877
+ print(result["result"]) # 1 (TASK_SUCCESSFUL_SCORE) or 0
878
+ """
879
+ from .verifiers.local_executor import execute_verifier_local
880
+
881
+ return execute_verifier_local(verifier_func, seed_db, current_db, final_answer)
882
+
850
883
  def list_runs(
851
884
  self, profile_id: Optional[str] = None, status: Optional[str] = "active"
852
885
  ) -> List[Run]:
@@ -823,6 +823,54 @@ def _parse_grade_response(data: dict) -> JudgeResult:
823
823
  return JudgeResult(score, details=data)
824
824
 
825
825
 
826
+ def _print_criteria_markers(criteria: list) -> None:
827
+ """Emit ``>>> CRITERIA >>>`` stdout markers for structured criteria display.
828
+
829
+ The orchestrator (theseus PR #1967) scans verifier stdout for these
830
+ markers and wraps the execution result so the client (client PR #1737)
831
+ can render an expandable rubric breakdown.
832
+
833
+ Converts from the orchestrator judge-response format::
834
+
835
+ {"name": str, "score": int, "max_score": int, "reasoning": str}
836
+
837
+ to the client-expected marker format::
838
+
839
+ {"criteria": str, "score": float, "score_out_of": float, "description"?: str}
840
+
841
+ Each criterion's score is normalised to a 0.0–1.0 float using its own
842
+ ``max_score``.
843
+ """
844
+ marker_criteria = []
845
+ for c in criteria:
846
+ name = c.get("name", "")
847
+ cscore = c.get("score", 0)
848
+ cmax = c.get("max_score", 0)
849
+
850
+ # Normalise per-criterion score to 0.0–1.0
851
+ if cmax and float(cmax) > 0:
852
+ norm_score = float(cscore) / float(cmax)
853
+ else:
854
+ norm_score = float(cscore)
855
+
856
+ entry: dict = {
857
+ "criteria": name,
858
+ "score": round(norm_score, 4),
859
+ "score_out_of": 1.0,
860
+ }
861
+
862
+ reasoning = c.get("reasoning", "")
863
+ if reasoning:
864
+ entry["description"] = reasoning
865
+
866
+ marker_criteria.append(entry)
867
+
868
+ if marker_criteria:
869
+ print(">>> CRITERIA >>>")
870
+ print(json.dumps(marker_criteria))
871
+ print("<<< CRITERIA <<<")
872
+
873
+
826
874
  def _print_judge_result(data: dict) -> None:
827
875
  """Print detailed judge grading result for verifier stdout capture."""
828
876
  model = data.get("model_used", "unknown")
@@ -848,6 +896,12 @@ def _print_judge_result(data: dict) -> None:
848
896
  if len(reasoning) > 200:
849
897
  reasoning = reasoning[:200] + "..."
850
898
  print(f"[C] {name}: {cscore}/{cmax} — {reasoning}")
899
+
900
+ # Emit structured criteria via stdout markers so the orchestrator
901
+ # (_extract_criteria_from_stdout) and client can render a rubric
902
+ # breakdown. Schema per element:
903
+ # {"criteria": str, "score": float, "score_out_of": float, "description"?: str}
904
+ _print_criteria_markers(criteria)
851
905
  else:
852
906
  print(f"[C] Score: {normalized:.2f}")
853
907
 
@@ -6,6 +6,7 @@ from .verifier import (
6
6
  verifier,
7
7
  SyncVerifierFunction,
8
8
  )
9
+ from .local_executor import execute_verifier_local, LocalEnvironment
9
10
 
10
11
  __all__ = [
11
12
  "DatabaseSnapshot",
@@ -15,4 +16,6 @@ __all__ = [
15
16
  "TASK_FAILED_SCORE",
16
17
  "verifier",
17
18
  "SyncVerifierFunction",
19
+ "execute_verifier_local",
20
+ "LocalEnvironment",
18
21
  ]
@@ -0,0 +1,247 @@
1
+ """Local verifier execution against SQLite files.
2
+
3
+ Executes verifier function code directly against local SQLite database files,
4
+ without requiring authentication or a remote runner API server.
5
+ """
6
+
7
+ import inspect
8
+ import json
9
+ import re
10
+ import string
11
+ import traceback
12
+ from io import StringIO
13
+ from typing import Any, Dict, Optional
14
+
15
+ from .db import DatabaseSnapshot, IgnoreConfig, SnapshotDiff
16
+ from .code import TASK_SUCCESSFUL_SCORE, TASK_FAILED_SCORE
17
+
18
+
19
+ # ---------------------------------------------------------------------------
20
+ # Helper functions injected into verifier execution namespace
21
+ # ---------------------------------------------------------------------------
22
+
23
+ _TRANSLATOR = str.maketrans(string.punctuation, " " * len(string.punctuation))
24
+
25
+
26
+ def _normalize_text(value: str) -> str:
27
+ text = value.lower().translate(_TRANSLATOR)
28
+ return "".join(text.split())
29
+
30
+
31
+ def _stringify_content(content: Any) -> str:
32
+ if isinstance(content, (dict, list)):
33
+ return json.dumps(content, sort_keys=True)
34
+ return str(content)
35
+
36
+
37
+ def normalized_contains(target: str, blob: Any) -> bool:
38
+ """Check if target is contained in blob after normalising punctuation and case."""
39
+ normalized_target = _normalize_text(target)
40
+ normalized_blob = _normalize_text(_stringify_content(blob))
41
+ return normalized_target in normalized_blob
42
+
43
+
44
+ def normalized_string_comparison(target: str, blob: Any) -> bool:
45
+ """Check if target equals blob after normalising punctuation and case."""
46
+ normalized_target = _normalize_text(target)
47
+ normalized_blob = _normalize_text(_stringify_content(blob))
48
+ return normalized_target == normalized_blob
49
+
50
+
51
+ def extract_numbers(text: str) -> list:
52
+ """Extract all numbers from a string."""
53
+ cleaned_text = text.replace(",", "")
54
+ pattern = r"-?\d+\.?\d*"
55
+ matches = re.findall(pattern, cleaned_text)
56
+ return [float(num) for num in matches]
57
+
58
+
59
+ def contains_number(text: str, target_number) -> bool:
60
+ """Check if text contains the target number."""
61
+ numbers = extract_numbers(text)
62
+ try:
63
+ if isinstance(target_number, str):
64
+ target_number = target_number.replace(",", "")
65
+ target = float(target_number)
66
+ except (ValueError, AttributeError):
67
+ return False
68
+ return target in numbers
69
+
70
+
71
+ # ---------------------------------------------------------------------------
72
+ # Lightweight Environment mock for local verifier execution
73
+ # ---------------------------------------------------------------------------
74
+
75
+ class _LocalInstance:
76
+ """Mock instance that supports load() as a no-op."""
77
+
78
+ def load(self):
79
+ pass
80
+
81
+
82
+ class LocalEnvironment:
83
+ """Lightweight environment that wraps local SQLite files for verifier execution.
84
+
85
+ Provides the same interface verifier functions expect from ``env``:
86
+ ``env.db("seed")``, ``env.db("current")``, and ``env.instance.load()``.
87
+ """
88
+
89
+ def __init__(self, seed_db: str, current_db: str):
90
+ self._snapshots: Dict[str, DatabaseSnapshot] = {
91
+ "seed": DatabaseSnapshot(seed_db, name="seed"),
92
+ "current": DatabaseSnapshot(current_db, name="current"),
93
+ }
94
+ self.instance = _LocalInstance()
95
+
96
+ def db(self, name: str = "current") -> DatabaseSnapshot:
97
+ if name not in self._snapshots:
98
+ raise KeyError(
99
+ f"Unknown database '{name}'. Available: {list(self._snapshots.keys())}"
100
+ )
101
+ return self._snapshots[name]
102
+
103
+
104
+ # ---------------------------------------------------------------------------
105
+ # Core execution function
106
+ # ---------------------------------------------------------------------------
107
+
108
+ def execute_verifier_local(
109
+ verifier_func: str,
110
+ seed_db: str,
111
+ current_db: str,
112
+ final_answer: Optional[str] = None,
113
+ ) -> Dict[str, Any]:
114
+ """Execute a verifier function string locally against SQLite database files.
115
+
116
+ No authentication or remote server required. The function is executed in an
117
+ isolated namespace with the same helpers available to production verifiers.
118
+
119
+ Args:
120
+ verifier_func: Python source code containing the verifier function definition.
121
+ seed_db: Path to the seed (before) SQLite database file.
122
+ current_db: Path to the current (after) SQLite database file.
123
+ final_answer: Optional final answer string passed to the verifier.
124
+
125
+ Returns:
126
+ Dict with keys:
127
+ - ``success`` (bool): Whether execution completed without errors.
128
+ - ``result`` (Any): The return value of the verifier function (typically a score).
129
+ - ``error`` (str | None): Error message and traceback if execution failed.
130
+ - ``stdout`` (str): Captured stdout output from the verifier function.
131
+ """
132
+ import sys
133
+
134
+ # Capture stdout
135
+ captured_stdout = StringIO()
136
+
137
+ try:
138
+ # Build the local environment
139
+ env = LocalEnvironment(seed_db, current_db)
140
+
141
+ # Clean the verifier code – strip decorators and fleet imports
142
+ cleaned_code = re.sub(r"@verifier\([^)]*\)\s*\n", "", verifier_func)
143
+ cleaned_code = re.sub(
144
+ r"^from fleet\.verifiers.*import.*$\n?",
145
+ "",
146
+ cleaned_code,
147
+ flags=re.MULTILINE,
148
+ )
149
+ cleaned_code = re.sub(
150
+ r"^from fleet import verifier.*$\n?",
151
+ "",
152
+ cleaned_code,
153
+ flags=re.MULTILINE,
154
+ )
155
+ cleaned_code = re.sub(
156
+ r"^import fleet\.verifiers.*$\n?",
157
+ "",
158
+ cleaned_code,
159
+ flags=re.MULTILINE,
160
+ )
161
+ cleaned_code = re.sub(
162
+ r"^import fleet$\n?", "", cleaned_code, flags=re.MULTILINE
163
+ )
164
+
165
+ # Build execution namespace with all helpers available to verifiers
166
+ exec_globals: Dict[str, Any] = {
167
+ # Score constants
168
+ "TASK_SUCCESSFUL_SCORE": TASK_SUCCESSFUL_SCORE,
169
+ "TASK_FAILED_SCORE": TASK_FAILED_SCORE,
170
+ # Helper functions
171
+ "normalized_contains": normalized_contains,
172
+ "normalized_string_comparison": normalized_string_comparison,
173
+ "extract_numbers": extract_numbers,
174
+ "contains_number": contains_number,
175
+ # Database classes
176
+ "DatabaseSnapshot": DatabaseSnapshot,
177
+ "IgnoreConfig": IgnoreConfig,
178
+ "SnapshotDiff": SnapshotDiff,
179
+ # Environment type hint (not enforced at runtime)
180
+ "Environment": type(env),
181
+ # Standard library modules commonly used in verifiers
182
+ "json": json,
183
+ "re": re,
184
+ "string": string,
185
+ # Builtins
186
+ "__builtins__": __builtins__,
187
+ }
188
+
189
+ # Execute the verifier code to define the function(s)
190
+ local_namespace: Dict[str, Any] = {}
191
+ exec(cleaned_code, exec_globals, local_namespace)
192
+
193
+ # Merge so helper functions defined in verifier code are accessible
194
+ exec_globals.update(local_namespace)
195
+
196
+ # Find the verifier function (the one defined in user code)
197
+ func_obj = None
198
+ for name, obj in local_namespace.items():
199
+ if inspect.isfunction(obj) and obj.__code__.co_filename == "<string>":
200
+ func_obj = obj
201
+ break
202
+
203
+ if func_obj is None:
204
+ return {
205
+ "success": False,
206
+ "result": None,
207
+ "error": "No function found in verifier code",
208
+ "stdout": "",
209
+ }
210
+
211
+ # Redirect stdout to capture print() output from verifiers
212
+ old_stdout = sys.stdout
213
+ sys.stdout = captured_stdout
214
+
215
+ try:
216
+ # Execute the verifier – verifiers take (env, final_answer=None)
217
+ sig = inspect.signature(func_obj)
218
+ params = list(sig.parameters.values())
219
+
220
+ if len(params) >= 2:
221
+ result = func_obj(env, final_answer)
222
+ elif len(params) == 1:
223
+ result = func_obj(env)
224
+ else:
225
+ result = func_obj()
226
+ finally:
227
+ sys.stdout = old_stdout
228
+
229
+ return {
230
+ "success": True,
231
+ "result": result,
232
+ "error": None,
233
+ "stdout": captured_stdout.getvalue(),
234
+ }
235
+
236
+ except Exception as e:
237
+ # Restore stdout if it was redirected
238
+ if sys.stdout is not sys.__stdout__ and sys.stdout is captured_stdout:
239
+ sys.stdout = sys.__stdout__
240
+
241
+ error_msg = f"{type(e).__name__}: {e}\n{traceback.format_exc()}"
242
+ return {
243
+ "success": False,
244
+ "result": None,
245
+ "error": error_msg,
246
+ "stdout": captured_stdout.getvalue(),
247
+ }
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: fleet-python
3
- Version: 0.2.115
3
+ Version: 0.2.117
4
4
  Summary: Python SDK for Fleet environments
5
5
  Author-email: Fleet AI <nic@fleet.so>
6
6
  License: Apache-2.0
@@ -101,6 +101,7 @@ fleet/verifiers/bundler.py
101
101
  fleet/verifiers/code.py
102
102
  fleet/verifiers/db.py
103
103
  fleet/verifiers/decorator.py
104
+ fleet/verifiers/local_executor.py
104
105
  fleet/verifiers/parse.py
105
106
  fleet/verifiers/sql_differ.py
106
107
  fleet/verifiers/verifier.py
@@ -117,6 +118,7 @@ tests/test_app_method.py
117
118
  tests/test_expect_exactly.py
118
119
  tests/test_expect_only.py
119
120
  tests/test_instance_dispatch.py
121
+ tests/test_judge_criteria_markers.py
120
122
  tests/test_sqlite_resource_dual_mode.py
121
123
  tests/test_sqlite_shared_memory_behavior.py
122
124
  tests/test_verifier_from_string.py
@@ -5,7 +5,7 @@ build-backend = "setuptools.build_meta"
5
5
  [project]
6
6
  name = "fleet-python"
7
7
 
8
- version = "0.2.115"
8
+ version = "0.2.117"
9
9
  description = "Python SDK for Fleet environments"
10
10
  authors = [
11
11
  {name = "Fleet AI", email = "nic@fleet.so"},
@@ -0,0 +1,192 @@
1
+ """Tests for structured criteria stdout markers in fleet.judge.
2
+
3
+ Validates that _print_criteria_markers emits the correct
4
+ >>> CRITERIA >>> / <<< CRITERIA <<< markers that the orchestrator
5
+ (theseus PR #1967) and client (client PR #1737) expect.
6
+ """
7
+
8
+ import json
9
+ import re
10
+ from io import StringIO
11
+ from unittest.mock import patch
12
+
13
+ import pytest
14
+
15
+ from fleet.judge import (
16
+ _print_criteria_markers,
17
+ _print_judge_result,
18
+ _parse_grade_response,
19
+ )
20
+
21
+
22
+ # ---------------------------------------------------------------------------
23
+ # Helpers
24
+ # ---------------------------------------------------------------------------
25
+
26
+ _MARKER_RE = re.compile(
27
+ r">>> CRITERIA >>>\s*\n(.*?)\n<<< CRITERIA <<<",
28
+ re.DOTALL,
29
+ )
30
+
31
+
32
+ def _capture_print(fn, *args, **kwargs):
33
+ """Capture all print() output from a function call."""
34
+ buf = StringIO()
35
+ with patch("builtins.print", side_effect=lambda *a, **kw: buf.write(" ".join(str(x) for x in a) + "\n")):
36
+ fn(*args, **kwargs)
37
+ return buf.getvalue()
38
+
39
+
40
+ def _extract_criteria_from_stdout(stdout: str):
41
+ """Mirror the orchestrator's extraction logic (theseus PR #1967)."""
42
+ m = _MARKER_RE.search(stdout)
43
+ if not m:
44
+ return None
45
+ parsed = json.loads(m.group(1).strip())
46
+ if isinstance(parsed, list):
47
+ return parsed
48
+ return None
49
+
50
+
51
+ # ---------------------------------------------------------------------------
52
+ # _print_criteria_markers
53
+ # ---------------------------------------------------------------------------
54
+
55
+ class TestPrintCriteriaMarkers:
56
+ """Tests for _print_criteria_markers."""
57
+
58
+ def test_basic_criteria_output(self):
59
+ """Emits valid markers with normalised scores."""
60
+ criteria = [
61
+ {"name": "Accuracy", "score": 8, "max_score": 10, "reasoning": "Good job"},
62
+ {"name": "Style", "score": 5, "max_score": 5, "reasoning": "Perfect"},
63
+ ]
64
+ stdout = _capture_print(_print_criteria_markers, criteria)
65
+
66
+ parsed = _extract_criteria_from_stdout(stdout)
67
+ assert parsed is not None, f"Markers not found in stdout:\n{stdout}"
68
+ assert len(parsed) == 2
69
+
70
+ assert parsed[0]["criteria"] == "Accuracy"
71
+ assert parsed[0]["score"] == pytest.approx(0.8, abs=0.01)
72
+ assert parsed[0]["score_out_of"] == 1.0
73
+ assert parsed[0]["description"] == "Good job"
74
+
75
+ assert parsed[1]["criteria"] == "Style"
76
+ assert parsed[1]["score"] == pytest.approx(1.0, abs=0.01)
77
+ assert parsed[1]["score_out_of"] == 1.0
78
+
79
+ def test_zero_max_score_passthrough(self):
80
+ """When max_score is 0, raw score passes through."""
81
+ criteria = [
82
+ {"name": "Metric", "score": 0.75, "max_score": 0},
83
+ ]
84
+ stdout = _capture_print(_print_criteria_markers, criteria)
85
+ parsed = _extract_criteria_from_stdout(stdout)
86
+ assert parsed is not None
87
+ assert parsed[0]["score"] == pytest.approx(0.75, abs=0.01)
88
+
89
+ def test_empty_criteria_no_markers(self):
90
+ """Empty list should produce no markers."""
91
+ stdout = _capture_print(_print_criteria_markers, [])
92
+ assert ">>> CRITERIA >>>" not in stdout
93
+
94
+ def test_reasoning_maps_to_description(self):
95
+ """The 'reasoning' field maps to 'description' in the marker schema."""
96
+ criteria = [
97
+ {"name": "Test", "score": 3, "max_score": 5, "reasoning": "Some reasoning here"},
98
+ ]
99
+ stdout = _capture_print(_print_criteria_markers, criteria)
100
+ parsed = _extract_criteria_from_stdout(stdout)
101
+ assert parsed[0]["description"] == "Some reasoning here"
102
+
103
+ def test_missing_reasoning_no_description(self):
104
+ """When reasoning is empty, description key should be absent."""
105
+ criteria = [
106
+ {"name": "Test", "score": 3, "max_score": 5, "reasoning": ""},
107
+ ]
108
+ stdout = _capture_print(_print_criteria_markers, criteria)
109
+ parsed = _extract_criteria_from_stdout(stdout)
110
+ assert "description" not in parsed[0]
111
+
112
+ def test_output_parseable_by_orchestrator_regex(self):
113
+ """Ensure the output matches the exact regex the orchestrator uses."""
114
+ criteria = [
115
+ {"name": "A", "score": 1, "max_score": 2, "reasoning": "half"},
116
+ ]
117
+ stdout = _capture_print(_print_criteria_markers, criteria)
118
+
119
+ # Use the exact regex from theseus PR #1967
120
+ m = re.search(
121
+ r">>> CRITERIA >>>\s*\n(.*?)\n<<< CRITERIA <<<",
122
+ stdout,
123
+ re.DOTALL,
124
+ )
125
+ assert m is not None, "Output doesn't match orchestrator regex"
126
+ data = json.loads(m.group(1).strip())
127
+ assert isinstance(data, list)
128
+ assert data[0]["criteria"] == "A"
129
+
130
+
131
+ # ---------------------------------------------------------------------------
132
+ # _print_judge_result integration
133
+ # ---------------------------------------------------------------------------
134
+
135
+ class TestPrintJudgeResult:
136
+ """Tests for _print_judge_result emitting criteria markers."""
137
+
138
+ def test_criteria_markers_emitted(self):
139
+ """_print_judge_result emits criteria markers when criteria present."""
140
+ data = {
141
+ "model_used": "claude-sonnet",
142
+ "provider_used": "anthropic",
143
+ "total_score": 15,
144
+ "max_score": 20,
145
+ "normalized_score": 0.75,
146
+ "criteria": [
147
+ {"name": "Accuracy", "score": 8, "max_score": 10, "reasoning": "Good"},
148
+ {"name": "Style", "score": 7, "max_score": 10, "reasoning": "Decent"},
149
+ ],
150
+ }
151
+ stdout = _capture_print(_print_judge_result, data)
152
+ parsed = _extract_criteria_from_stdout(stdout)
153
+ assert parsed is not None
154
+ assert len(parsed) == 2
155
+
156
+ def test_no_criteria_no_markers(self):
157
+ """_print_judge_result doesn't emit markers when no criteria."""
158
+ data = {
159
+ "model_used": "claude-sonnet",
160
+ "provider_used": "anthropic",
161
+ "total_score": 0,
162
+ "max_score": 0,
163
+ "normalized_score": 0.5,
164
+ }
165
+ stdout = _capture_print(_print_judge_result, data)
166
+ assert ">>> CRITERIA >>>" not in stdout
167
+
168
+
169
+ # ---------------------------------------------------------------------------
170
+ # _parse_grade_response integration
171
+ # ---------------------------------------------------------------------------
172
+
173
+ class TestParseGradeResponse:
174
+ """Tests for _parse_grade_response emitting criteria markers."""
175
+
176
+ def test_full_flow_emits_markers(self):
177
+ """_parse_grade_response → _print_judge_result → criteria markers."""
178
+ data = {
179
+ "model_used": "claude-sonnet",
180
+ "provider_used": "anthropic",
181
+ "total_score": 9,
182
+ "max_score": 10,
183
+ "normalized_score": 0.9,
184
+ "criteria": [
185
+ {"name": "Completeness", "score": 9, "max_score": 10, "reasoning": "Almost perfect"},
186
+ ],
187
+ }
188
+ stdout = _capture_print(_parse_grade_response, data)
189
+ parsed = _extract_criteria_from_stdout(stdout)
190
+ assert parsed is not None
191
+ assert parsed[0]["criteria"] == "Completeness"
192
+ assert parsed[0]["score"] == pytest.approx(0.9, abs=0.01)
File without changes
File without changes
File without changes