synth-ai 0.2.4.dev8__py3-none-any.whl → 0.2.5__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of synth-ai might be problematic. Click here for more details.

Files changed (112) hide show
  1. synth_ai/__init__.py +1 -1
  2. synth_ai/cli/__init__.py +6 -0
  3. synth_ai/cli/demo.py +68 -9
  4. synth_ai/cli/rl_demo.py +137 -0
  5. synth_ai/cli/root.py +65 -0
  6. synth_ai/demos/core/__init__.py +1 -0
  7. synth_ai/demos/core/cli.py +685 -0
  8. synth_ai/demos/demo_task_apps/__init__.py +1 -0
  9. synth_ai/demos/demo_task_apps/core.py +374 -0
  10. synth_ai/demos/demo_task_apps/math/__init__.py +1 -0
  11. synth_ai/demos/demo_task_apps/math/app.py +37 -0
  12. synth_ai/demos/demo_task_apps/math/config.toml +44 -0
  13. synth_ai/demos/demo_task_apps/math/deploy_modal.py +60 -0
  14. synth_ai/demos/demo_task_apps/math/deploy_task_app.sh +22 -0
  15. synth_ai/environments/examples/bandit/__init__.py +33 -0
  16. synth_ai/environments/examples/bandit/engine.py +294 -0
  17. synth_ai/environments/examples/bandit/environment.py +194 -0
  18. synth_ai/environments/examples/bandit/taskset.py +200 -0
  19. synth_ai/environments/examples/crafter_classic/agent_demos/analyze_semantic_words_markdown.py +250 -0
  20. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_comprehensive_evaluation.py +59 -0
  21. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_evaluation_browser.py +152 -0
  22. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_evaluation_config.toml +24 -0
  23. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_evaluation_framework.py +1194 -0
  24. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/crafter_synth_config.toml +56 -0
  25. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/filter_config_modal.toml +32 -0
  26. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/filter_traces_sft_turso.py +724 -0
  27. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/kick_off_ft_modal.py +384 -0
  28. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/analyze_action_results.py +53 -0
  29. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/analyze_agent_actions.py +178 -0
  30. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/analyze_latest_run.py +222 -0
  31. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/analyze_lm_traces.py +183 -0
  32. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/analyze_no_rewards.py +210 -0
  33. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/analyze_trace_issue.py +206 -0
  34. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/check_db_schema.py +49 -0
  35. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/check_latest_results.py +64 -0
  36. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/debug_agent_responses.py +88 -0
  37. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_modal_ft/old/quick_trace_check.py +77 -0
  38. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/compare_experiments.py +324 -0
  39. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/filter_traces_sft_turso.py +580 -0
  40. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/kick_off_ft_oai.py +362 -0
  41. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/multi_model_config.toml +49 -0
  42. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/analyze_enhanced_hooks.py +332 -0
  43. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/analyze_hook_events.py +97 -0
  44. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/analyze_hook_results.py +217 -0
  45. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/check_hook_storage.py +87 -0
  46. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/check_seeds.py +88 -0
  47. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/compare_seed_performance.py +195 -0
  48. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/custom_eval_pipelines.py +400 -0
  49. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/plot_hook_frequency.py +195 -0
  50. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/old/seed_analysis_summary.py +56 -0
  51. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_openai_ft/run_rollouts_for_models_and_compare_v3.py +858 -0
  52. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_quick_evaluation.py +52 -0
  53. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_react_agent.py +874 -0
  54. synth_ai/environments/examples/crafter_classic/agent_demos/crafter_trace_evaluation.py +1412 -0
  55. synth_ai/environments/examples/crafter_classic/agent_demos/example_v3_usage.py +216 -0
  56. synth_ai/environments/examples/crafter_classic/agent_demos/old/compare_traces.py +296 -0
  57. synth_ai/environments/examples/crafter_classic/agent_demos/old/crafter_comprehensive_evaluation.py +58 -0
  58. synth_ai/environments/examples/crafter_classic/agent_demos/old/crafter_env_serialization.py +464 -0
  59. synth_ai/environments/examples/crafter_classic/agent_demos/old/crafter_evaluation_browser.py +152 -0
  60. synth_ai/environments/examples/crafter_classic/agent_demos/old/crafter_quick_evaluation.py +51 -0
  61. synth_ai/environments/examples/crafter_classic/agent_demos/old/crafter_trace_evaluation.py +1412 -0
  62. synth_ai/environments/examples/crafter_classic/agent_demos/old/debug_player_loss.py +112 -0
  63. synth_ai/environments/examples/crafter_classic/agent_demos/old/diagnose_service.py +203 -0
  64. synth_ai/environments/examples/crafter_classic/agent_demos/old/diagnose_slowness.py +305 -0
  65. synth_ai/environments/examples/crafter_classic/agent_demos/old/eval_by_difficulty.py +126 -0
  66. synth_ai/environments/examples/crafter_classic/agent_demos/old/eval_example.py +94 -0
  67. synth_ai/environments/examples/crafter_classic/agent_demos/old/explore_saved_states.py +142 -0
  68. synth_ai/environments/examples/crafter_classic/agent_demos/old/filter_traces_sft.py +26 -0
  69. synth_ai/environments/examples/crafter_classic/agent_demos/old/filter_traces_sft_OLD.py +984 -0
  70. synth_ai/environments/examples/crafter_classic/agent_demos/old/generate_ft_data_gemini.py +724 -0
  71. synth_ai/environments/examples/crafter_classic/agent_demos/old/generate_ft_data_modal.py +386 -0
  72. synth_ai/environments/examples/crafter_classic/agent_demos/old/generate_ft_metadata.py +205 -0
  73. synth_ai/environments/examples/crafter_classic/agent_demos/old/kick_off_ft_gemini.py +150 -0
  74. synth_ai/environments/examples/crafter_classic/agent_demos/old/kick_off_ft_modal.py +283 -0
  75. synth_ai/environments/examples/crafter_classic/agent_demos/old/prepare_vertex_ft.py +280 -0
  76. synth_ai/environments/examples/crafter_classic/agent_demos/old/profile_env_slowness.py +456 -0
  77. synth_ai/environments/examples/crafter_classic/agent_demos/old/replicate_issue.py +166 -0
  78. synth_ai/environments/examples/crafter_classic/agent_demos/old/run_and_eval.py +102 -0
  79. synth_ai/environments/examples/crafter_classic/agent_demos/old/run_comparison.py +128 -0
  80. synth_ai/environments/examples/crafter_classic/agent_demos/old/run_qwen_rollouts.py +655 -0
  81. synth_ai/environments/examples/crafter_classic/agent_demos/old/trace_eval_OLD.py +202 -0
  82. synth_ai/environments/examples/crafter_classic/agent_demos/old/validate_openai_format.py +166 -0
  83. synth_ai/environments/examples/crafter_classic/environment.py +41 -2
  84. synth_ai/environments/examples/crafter_custom/agent_demos/__init__.py +1 -0
  85. synth_ai/environments/examples/crafter_custom/agent_demos/trace_eval.py +202 -0
  86. synth_ai/environments/examples/crafter_custom/old/analyze_diamond_issue.py +159 -0
  87. synth_ai/environments/examples/crafter_custom/old/analyze_diamond_spawning.py +158 -0
  88. synth_ai/environments/examples/crafter_custom/old/compare_worlds.py +71 -0
  89. synth_ai/environments/examples/crafter_custom/old/dataset_stats.py +105 -0
  90. synth_ai/environments/examples/crafter_custom/old/diamond_spawning_summary.py +119 -0
  91. synth_ai/environments/examples/crafter_custom/old/example_dataset_usage.py +52 -0
  92. synth_ai/environments/examples/enron/units/keyword_stats.py +112 -0
  93. synth_ai/environments/examples/minigrid/agent_demos/minigrid_evaluation_framework.py +1188 -0
  94. synth_ai/environments/examples/minigrid/agent_demos/minigrid_quick_evaluation.py +48 -0
  95. synth_ai/environments/examples/minigrid/agent_demos/minigrid_react_agent.py +562 -0
  96. synth_ai/environments/examples/minigrid/agent_demos/minigrid_trace_evaluation.py +221 -0
  97. synth_ai/environments/examples/nethack/agent_demos/nethack_evaluation_framework.py +981 -0
  98. synth_ai/environments/examples/nethack/agent_demos/nethack_quick_evaluation.py +74 -0
  99. synth_ai/environments/examples/nethack/agent_demos/nethack_react_agent.py +831 -0
  100. synth_ai/environments/examples/red/agent_demos/__init__.py +1 -0
  101. synth_ai/environments/examples/red/units/__init__.py +1 -0
  102. synth_ai/environments/examples/sokoban/agent_demos/sokoban_full_eval.py +899 -0
  103. synth_ai/environments/examples/sokoban/units/astar_common.py +95 -0
  104. synth_ai/environments/service/app.py +8 -0
  105. synth_ai/install_sqld.sh +40 -0
  106. synth_ai-0.2.5.dist-info/METADATA +106 -0
  107. {synth_ai-0.2.4.dev8.dist-info → synth_ai-0.2.5.dist-info}/RECORD +111 -12
  108. {synth_ai-0.2.4.dev8.dist-info → synth_ai-0.2.5.dist-info}/entry_points.txt +1 -0
  109. synth_ai-0.2.4.dev8.dist-info/METADATA +0 -635
  110. {synth_ai-0.2.4.dev8.dist-info → synth_ai-0.2.5.dist-info}/WHEEL +0 -0
  111. {synth_ai-0.2.4.dev8.dist-info → synth_ai-0.2.5.dist-info}/licenses/LICENSE +0 -0
  112. {synth_ai-0.2.4.dev8.dist-info → synth_ai-0.2.5.dist-info}/top_level.txt +0 -0
@@ -0,0 +1,874 @@
1
+ import asyncio
2
+ import json
3
+ import uuid
4
+ from collections import deque
5
+ from pathlib import Path
6
+ from typing import Any, Deque, Dict, List, Optional, Set, Union
7
+
8
+ import pytest
9
+ import toml
10
+ from pydantic import BaseModel, Field
11
+ from synth_ai.environments.examples.crafter_classic.engine import (
12
+ CRAFTER_ACTION_MAP, # map of action name to int
13
+ )
14
+
15
+ # Crafter specific imports
16
+ from synth_ai.environments.examples.crafter_classic.environment import (
17
+ CrafterClassicEnvironment,
18
+ CrafterPrivateState,
19
+ CrafterPublicState,
20
+ )
21
+ from synth_ai.lm.core.main import LM
22
+ from synth_ai.lm.tools.base import BaseTool
23
+ from synth_sdk.tracing.abstractions import Dataset, RewardSignal, TrainingQuestion
24
+ from synth_sdk.tracing.decorators import trace_event_async
25
+ from synth_sdk.tracing.trackers import SynthTracker
26
+ from synth_sdk.tracing.utils import get_system_id
27
+
28
+ # Convert CRAFTER_ACTION_MAP to ACTION_STRING_TO_INT and INT_TO_ACTION_STRING
29
+ ACTION_STRING_TO_INT: Dict[str, int] = CRAFTER_ACTION_MAP
30
+ INT_TO_ACTION_STRING: Dict[int, str] = {v: k for k, v in CRAFTER_ACTION_MAP.items()}
31
+
32
+
33
+ import logging
34
+
35
+ from synth_ai.environments.environment.shared_engine import (
36
+ GetObservationCallable,
37
+ InternalObservation,
38
+ )
39
+ from synth_ai.environments.environment.tools import EnvToolCall
40
+ from synth_ai.environments.examples.crafter_classic.taskset import (
41
+ CrafterTaskInstance,
42
+ CrafterTaskInstanceMetadata,
43
+ )
44
+ from synth_ai.environments.tasks.core import Impetus, Intent
45
+
46
+ logging.disable(logging.CRITICAL)
47
+
48
+
49
+ # --- Helper to build crafter semantic mapping ---
50
+ def get_crafter_semantic_mapping():
51
+ """Build the crafter semantic ID to item name mapping."""
52
+ import itertools
53
+
54
+ import crafter
55
+
56
+ # Create a dummy env to get ID mappings (same as environment.py)
57
+ dummyenv = None
58
+ try:
59
+ dummyenv = crafter.Env()
60
+ max_id = (
61
+ max(
62
+ max(dummyenv._world._mat_ids.values()),
63
+ max(dummyenv._sem_view._obj_ids.values()),
64
+ )
65
+ + 1
66
+ )
67
+ id_to_item = ["void"] * max_id
68
+ for name, ind in itertools.chain(
69
+ dummyenv._world._mat_ids.items(), dummyenv._sem_view._obj_ids.items()
70
+ ):
71
+ if name is None:
72
+ clean = "none"
73
+ elif hasattr(name, "__name__"):
74
+ clean = name.__name__
75
+ else:
76
+ clean = str(name)
77
+ id_to_item[ind] = clean.lower()
78
+ player_idx = id_to_item.index("player")
79
+ return id_to_item, player_idx
80
+ finally:
81
+ if dummyenv:
82
+ try:
83
+ dummyenv.close()
84
+ except Exception:
85
+ pass
86
+ del dummyenv
87
+
88
+
89
+ # --- Helper function to format observation for LLM ---
90
+ def format_obs_for_llm_from_states(pub: CrafterPublicState, priv: CrafterPrivateState) -> str:
91
+ inventory_str = ", ".join(f"{k}:{v}" for k, v in pub.inventory.items() if v > 0)
92
+ if not inventory_str:
93
+ inventory_str = "empty"
94
+
95
+ achievements_str = ", ".join(k for k, v in pub.achievements_status.items() if v)
96
+ if not achievements_str:
97
+ achievements_str = "none"
98
+
99
+ # Add map view around player using the real crafter semantic mapping
100
+ map_view = ""
101
+ if pub.semantic_map is not None:
102
+ px, py = pub.player_position
103
+ view_size = 7 # 7x7 view around player
104
+ half_view = view_size // 2
105
+
106
+ # Get the real crafter semantic mapping
107
+ id_to_item, player_idx = get_crafter_semantic_mapping()
108
+
109
+ # Create a local view around the player using same logic as _plain_grid
110
+ map_view += f"\nLocal Map View ({view_size}x{view_size} around player):\n"
111
+ matrix = []
112
+ for dy in range(-half_view, half_view + 1):
113
+ row = []
114
+ for dx in range(-half_view, half_view + 1):
115
+ x, y = px + dx, py + dy
116
+ if pub.semantic_map is None or not (
117
+ 0 <= x < pub.semantic_map.shape[0] and 0 <= y < pub.semantic_map.shape[1]
118
+ ):
119
+ row.append("void")
120
+ else:
121
+ idx = pub.semantic_map[x, y]
122
+ if dx == 0 and dy == 0:
123
+ row.append("player") # Player position
124
+ else:
125
+ # Use the real crafter mapping
126
+ item_name = id_to_item[idx] if idx < len(id_to_item) else "unknown"
127
+ row.append(item_name)
128
+ matrix.append(row)
129
+
130
+ # Transpose the matrix like _plain_grid does
131
+ transposed = list(zip(*matrix))
132
+ # Convert each row to a space-separated string
133
+ for row in transposed:
134
+ map_view += " ".join(row) + "\n"
135
+
136
+ # Create a legend of items actually visible in the map
137
+ visible_items = set()
138
+ for row in transposed:
139
+ for item in row:
140
+ if item not in ["void", "player"]:
141
+ visible_items.add(item)
142
+
143
+ if visible_items:
144
+ map_view += f"\nVisible items: {', '.join(sorted(visible_items))}"
145
+ else:
146
+ map_view += "\nNo special items visible (mostly grass/empty)"
147
+
148
+ # Simplified observation, focusing on key elements
149
+ return (
150
+ f"Steps: {pub.num_steps_taken}/{pub.max_steps_episode}\n"
151
+ f"Health: {priv.player_internal_stats.get('health', 'N/A')}\n"
152
+ f"Inventory: {inventory_str}\n"
153
+ f"Unlocked Achievements: {achievements_str}\n"
154
+ f"Player Position: {pub.player_position}\n"
155
+ f"Last Reward: {priv.reward_last_step:.2f}\n"
156
+ f"Terminated: {priv.terminated} | Truncated: {priv.truncated}"
157
+ f"{map_view}"
158
+ )
159
+
160
+
161
+ # ---------------------------------- custom observation callable (Optional, can be simpler for Crafter) ------------------------------ #
162
+ # For now, let's assume the default observation from the environment is sufficient,
163
+ # or we will use the direct public/private states.
164
+ # If history is needed, we can adapt the Sokoban HistoryObservationCallable.
165
+ class CrafterHistoryObservationCallable(GetObservationCallable):
166
+ def __init__(self, max_history: int = 1): # Keep only current obs for simplicity now
167
+ self._hist_obs: Deque[str] = deque(maxlen=max_history)
168
+ self._hist_pub_state: Deque[CrafterPublicState] = deque(maxlen=max_history)
169
+ self._hist_priv_state: Deque[CrafterPrivateState] = deque(maxlen=max_history)
170
+
171
+ async def get_observation(
172
+ self, pub: CrafterPublicState, priv: CrafterPrivateState
173
+ ) -> InternalObservation:
174
+ if pub is None or priv is None:
175
+ return {
176
+ "error": "Missing public or private state in get_observation",
177
+ "history_formatted_obs": list(self._hist_obs),
178
+ } # type: ignore[return-value]
179
+
180
+ formatted_obs = format_obs_for_llm_from_states(pub, priv)
181
+ self._hist_obs.append(formatted_obs)
182
+ self._hist_pub_state.append(pub)
183
+ self._hist_priv_state.append(priv)
184
+
185
+ return {
186
+ "public": pub,
187
+ "private": priv,
188
+ "formatted_obs": formatted_obs, # Current formatted obs
189
+ "history_formatted_obs": list(self._hist_obs), # History of formatted obs
190
+ "history_public_states": list(self._hist_pub_state),
191
+ "history_private_states": list(self._hist_priv_state),
192
+ } # type: ignore[return-value]
193
+
194
+
195
+ # --- Pydantic Models for Tool Arguments ---
196
+ class CrafterInteractArgs(BaseModel):
197
+ actions_list: List[str] = Field(
198
+ description="A list of action names to execute in sequence in the Crafter environment (e.g., ['move_up', 'move_up', 'place_stone']). Can contain 1-10 actions."
199
+ )
200
+ reasoning: str = Field(description="A brief explanation of why these actions were chosen.")
201
+
202
+
203
+ # class TerminateArgs(BaseModel):
204
+ # reason: str = Field(
205
+ # description="A detailed reason for why the agent is terminating."
206
+ # )
207
+
208
+
209
+ # --- ReAct agent for Crafter -------------------------------------------------- #
210
+ class CrafterInteractTool(BaseTool):
211
+ """Tool for interacting with Crafter environment"""
212
+
213
+ name: str = "crafter_interact"
214
+ arguments: type[BaseModel] = CrafterInteractArgs
215
+ description: str = (
216
+ "Interacts with the Crafter environment by proposing a sequence of 1-10 actions to execute."
217
+ )
218
+
219
+
220
+ # class TerminateTool(BaseTool):
221
+ # """Tool for terminating agent execution"""
222
+ # name: str = "terminate"
223
+ # arguments: type[BaseModel] = TerminateArgs
224
+ # description: str = "Terminates the agent's execution if the task is considered complete or no useful progress can be made."
225
+
226
+
227
+ class CrafterMove(EnvToolCall): # Simple EnvToolCall wrapper
228
+ def __init__(self, action: int):
229
+ super().__init__(tool="interact", args={"action": action})
230
+
231
+
232
+ class ReActAgent:
233
+ def __init__(self, llm, max_turns: int = 50): # Increased max_turns for Crafter
234
+ self.llm, self.max_turns = llm, max_turns
235
+ self.history: List[Dict[str, Any]] = []
236
+ self.system_name: str = "crafter-react-ex" # Changed system name
237
+ self.system_id: Any = get_system_id(self.system_name)
238
+ self.system_instance_id: str = str(uuid.uuid4())
239
+ self.last_obs_dict: Optional[Dict[str, Any]] = (
240
+ None # To store raw observation for terminate guardrails
241
+ )
242
+ self.current_achievements: Set[str] = set() # To track unique achievements
243
+
244
+ self.tools = [
245
+ CrafterInteractTool(),
246
+ # TerminateTool(), # Commented out to prevent early quitting
247
+ ]
248
+
249
+ def _format_history_for_prompt(self) -> str:
250
+ prompt_history = []
251
+ for entry in self.history:
252
+ if entry["type"] == "obs":
253
+ prompt_history.append(f"OBSERVATION:\n{entry['content']}")
254
+ elif entry["type"] == "tool_call":
255
+ args_str = json.dumps(entry["tool_arguments"])
256
+ prompt_history.append(
257
+ f"THOUGHT:\nI will call the tool `{entry['tool_name']}` with arguments: {args_str}\nACTION: (Tool call executed)"
258
+ )
259
+ elif entry["type"] == "tool_response":
260
+ prompt_history.append(
261
+ "TOOL_RESPONSE:\n(Action executed, new observation will follow if not terminal)"
262
+ )
263
+ return "\n".join(prompt_history)
264
+
265
+ @trace_event_async(event_type="react_agent_decide")
266
+ async def decide(
267
+ self, obs_str: str, current_raw_obs: Dict[str, Any]
268
+ ) -> List[int]: # Return list of action integers
269
+ self.history.append({"type": "obs", "content": obs_str})
270
+ self.last_obs_dict = current_raw_obs # Store for terminate guardrail
271
+
272
+ # Update current achievements from the raw observation
273
+ if current_raw_obs and isinstance(current_raw_obs.get("public"), CrafterPublicState):
274
+ pub_state: CrafterPublicState = current_raw_obs["public"]
275
+ for ach, unlocked in pub_state.achievements_status.items():
276
+ if unlocked:
277
+ self.current_achievements.add(ach)
278
+
279
+ formatted_prompt_history = self._format_history_for_prompt()
280
+
281
+ # Updated prompt for Crafter
282
+ prompt = (
283
+ f"{formatted_prompt_history}\n\n"
284
+ "Based on the history above, particularly the last observation (health, inventory, achievements, position), "
285
+ "what is your reasoning and which `crafter_interact` tool should you call next? "
286
+ "Prioritize actions that lead to new achievements or ensure survival (e.g., find food if health is low)."
287
+ )
288
+
289
+ system_message = (
290
+ "You are an agent playing Crafter. Your goal is to survive and unlock as many achievements as possible. "
291
+ "Review the history of observations, thoughts, and actions. "
292
+ "Based on this history, particularly the last observation, decide on the best sequence of actions. "
293
+ "You MUST call the available tool: `crafter_interact`.\\n\\n"
294
+ "For `crafter_interact`, provide a list of 1-10 actions to execute in sequence. "
295
+ "Planning ahead with multiple actions is often more efficient than single actions. "
296
+ f"Available actions are: {', '.join(ACTION_STRING_TO_INT.keys())}.\\n"
297
+ "Always provide a `reasoning` field in your tool call."
298
+ )
299
+
300
+ # Trace the LLM interaction input so that full messages (system & user) are included in the trace
301
+ SynthTracker.track_lm(
302
+ messages=[
303
+ {"role": "system", "content": system_message},
304
+ {"role": "user", "content": prompt},
305
+ ],
306
+ model_name=self.llm.model_name,
307
+ model_params=None,
308
+ finetune=False,
309
+ )
310
+
311
+ response_obj = await self.llm.respond_async(
312
+ system_message=system_message, user_message=prompt, tools=self.tools
313
+ )
314
+
315
+ # Trace the assistant's reply/output so that it is captured alongside the inputs
316
+ SynthTracker.track_lm_output(
317
+ messages=[{"role": "assistant", "content": response_obj.raw_response}],
318
+ model_name=self.llm.model_name,
319
+ finetune=False,
320
+ )
321
+
322
+ tool_calls = response_obj.tool_calls
323
+
324
+ # Handle case where tool_calls is None or empty (noop to prevent crash)
325
+ if not tool_calls:
326
+ # print(f"[WARNING] No tool calls returned by {self.llm.model_name}, returning noop action")
327
+ self.history.append(
328
+ {
329
+ "type": "tool_call",
330
+ "tool_name": "noop",
331
+ "tool_arguments": {"reason": "no_tool_calls_returned"},
332
+ }
333
+ )
334
+ self.history.append(
335
+ {
336
+ "type": "tool_response",
337
+ "content": "Noop executed due to missing tool calls",
338
+ }
339
+ )
340
+ return [0] # Return 'noop' action (action index 0)
341
+
342
+ tool_call_data = tool_calls[0]
343
+
344
+ # Handle both dict and object formats
345
+ if isinstance(tool_call_data, dict):
346
+ tool_name = tool_call_data["function"]["name"]
347
+ tool_args_str = tool_call_data["function"]["arguments"]
348
+ else:
349
+ tool_name = tool_call_data.function.name
350
+ tool_args_str = tool_call_data.function.arguments
351
+
352
+ tool_arguments = json.loads(tool_args_str)
353
+
354
+ # Track the tool call details for richer debugging and training signals
355
+ SynthTracker.track_state(
356
+ variable_name="tool_call",
357
+ variable_value={"tool_name": tool_name, "arguments": tool_arguments},
358
+ origin="agent",
359
+ )
360
+
361
+ self.history.append(
362
+ {
363
+ "type": "tool_call",
364
+ "tool_name": tool_name,
365
+ "tool_arguments": tool_arguments,
366
+ }
367
+ )
368
+ self.history.append({"type": "tool_response", "content": "Tool executed"})
369
+
370
+ if tool_name == "crafter_interact":
371
+ actions_list = tool_arguments["actions_list"]
372
+
373
+ # Convert action names to integers
374
+ action_ints = []
375
+ for action_str in actions_list:
376
+ if action_str in ACTION_STRING_TO_INT:
377
+ action_ints.append(ACTION_STRING_TO_INT[action_str])
378
+ else:
379
+ print(f"[WARNING] Invalid action '{action_str}', using noop instead")
380
+ action_ints.append(0) # noop action
381
+
382
+ return action_ints
383
+
384
+ # elif tool_name == "terminate":
385
+ # reason = tool_arguments["reason"]
386
+ #
387
+ # # Add the human-readable termination reason to the history
388
+ # self.history.append({
389
+ # "type": "termination",
390
+ # "content": f"Agent terminated: {reason}",
391
+ # "reason": reason
392
+ # })
393
+ #
394
+ # return [-1] # Special termination indicator
395
+
396
+
397
+ # --- Test for a single agent run ---
398
+ @pytest.mark.asyncio
399
+ async def test_react_agent_crafter(tmp_path: Path):
400
+ # Create a simple Crafter task instance for testing
401
+ # For Crafter, the seed in metadata is important for reproducibility.
402
+ # initial_engine_snapshot can be None if the engine handles reset with seed.
403
+ task_metadata = CrafterTaskInstanceMetadata(
404
+ difficulty="easy",
405
+ seed=42,
406
+ # Other metadata fields can be default or placeholders if not critical for this test
407
+ num_trees_radius=0, # Placeholder, actual values depend on seed and world gen
408
+ num_cows_radius=0, # Placeholder
409
+ num_hostiles_radius=0, # Placeholder
410
+ )
411
+ inst = CrafterTaskInstance(
412
+ id=uuid.uuid4(),
413
+ impetus=Impetus(instructions="Survive and unlock achievements."),
414
+ intent=Intent(
415
+ rubric={"goal": "Unlock achievements and survive"},
416
+ gold_trajectories=None,
417
+ gold_state_diff={},
418
+ ),
419
+ metadata=task_metadata,
420
+ is_reproducible=True,
421
+ initial_engine_snapshot=None, # Engine will init with seed from metadata
422
+ )
423
+
424
+ hist_cb = CrafterHistoryObservationCallable(max_history=1)
425
+ env = CrafterClassicEnvironment(inst, custom_step_obs=hist_cb)
426
+ # env.engine.package_sokoban_env.render_mode = "raw" # Not applicable to Crafter
427
+
428
+ llm = LM(model_name="gpt-4.1-nano", formatting_model_name="gpt-4.1-nano", temperature=0.0)
429
+ agent = ReActAgent(llm, max_turns=30) # Increased for meaningful progress
430
+ print("[DEBUG] Created agent with max_turns=30")
431
+
432
+ async def run_episode():
433
+ obs_payload = await env.initialize()
434
+
435
+ if "error" in obs_payload:
436
+ print(f"Error during env.initialize: {obs_payload['error']}")
437
+ return False, 0
438
+
439
+ # Initial observation for the agent
440
+ # The CrafterHistoryObservationCallable returns a dict with 'public', 'private', 'formatted_obs'
441
+ current_formatted_obs = obs_payload["formatted_obs"]
442
+ raw_obs_for_agent_decision = (
443
+ obs_payload # Pass the whole payload which includes public and private states
444
+ )
445
+
446
+ for turn in range(agent.max_turns):
447
+ action_sequence = await agent.decide(current_formatted_obs, raw_obs_for_agent_decision)
448
+
449
+ if action_sequence == [-1]: # Agent decided to terminate
450
+ obs_payload_next = obs_payload # No new observation if terminated by agent
451
+ break
452
+
453
+ # Execute each action in the sequence
454
+ for act_idx in action_sequence:
455
+ step_result = await env.step([[CrafterMove(act_idx)]])
456
+ obs_payload_next = step_result
457
+
458
+ if "error" in obs_payload_next:
459
+ break
460
+
461
+ # Update observation for next action in sequence
462
+ current_formatted_obs = obs_payload_next["formatted_obs"]
463
+ raw_obs_for_agent_decision = obs_payload_next
464
+ obs_payload = obs_payload_next
465
+
466
+ # Check if environment terminated after this sub-action
467
+ if obs_payload_next["private"].terminated or obs_payload_next["private"].truncated:
468
+ priv_state = obs_payload_next["private"]
469
+ pub_state = obs_payload_next["public"]
470
+ player_health = priv_state.player_internal_stats.get("health", "N/A")
471
+ steps_taken = pub_state.num_steps_taken
472
+ max_steps = pub_state.max_steps_episode
473
+
474
+ break
475
+
476
+ if "error" in obs_payload_next:
477
+ break
478
+
479
+ # Update observations for the next agent decision
480
+ current_formatted_obs = obs_payload_next["formatted_obs"]
481
+ raw_obs_for_agent_decision = obs_payload_next
482
+
483
+ agent.history.append({"type": "tool_response", "content": "Action executed"})
484
+
485
+ obs_payload = obs_payload_next
486
+
487
+ if obs_payload_next["private"].terminated or obs_payload_next["private"].truncated:
488
+ break
489
+
490
+ # Ensure obs_payload_next is defined even if loop didn't run or agent terminated early
491
+ if "obs_payload_next" not in locals():
492
+ obs_payload_next = obs_payload
493
+
494
+ if "error" in obs_payload_next:
495
+ return False, len(agent.current_achievements)
496
+
497
+ # Success could be defined as surviving some steps or achieving something
498
+ # For this test, let's say it's successful if it ran and terminated/truncated by env
499
+ final_private_state: CrafterPrivateState = obs_payload_next["private"]
500
+ episode_successful = final_private_state.terminated or final_private_state.truncated
501
+ return episode_successful, len(agent.current_achievements)
502
+
503
+ episode_completed, num_achievements = await run_episode()
504
+
505
+ dataset = Dataset(
506
+ questions=[
507
+ TrainingQuestion(
508
+ id="crafter_ep_test",
509
+ intent="survive and achieve",
510
+ criteria="completed_episode_or_achieved_something",
511
+ )
512
+ ],
513
+ reward_signals=[
514
+ RewardSignal(
515
+ question_id="crafter_ep_test",
516
+ system_instance_id=agent.system_instance_id,
517
+ reward=1
518
+ if episode_completed or num_achievements > 0
519
+ else 0, # Reward if completed or got any achievement
520
+ )
521
+ ],
522
+ )
523
+ # upload(dataset=dataset) # Optional: uncomment to upload trace
524
+
525
+ assert episode_completed or num_achievements > 0, (
526
+ "Agent failed to complete the episode or unlock any achievement in the test."
527
+ )
528
+
529
+
530
+ async def eval_react_crafter(
531
+ model_name: str = "gpt-4.1-nano",
532
+ formatting_model_name: str = "gpt-4.1-nano",
533
+ modes: Optional[List[str]] = None,
534
+ n_instances_per_mode: int = 3,
535
+ ) -> List[Dict[str, Any]]:
536
+ """
537
+ Run ReAct agents on Crafter instances of different difficulties,
538
+ and returns a list of dictionaries containing aggregated results for each mode.
539
+ """
540
+ # Import the new evaluation framework
541
+ from synth_ai.environments.examples.crafter_classic.agent_demos.eval_framework import (
542
+ run_crafter_eval,
543
+ )
544
+
545
+ if modes is None:
546
+ modes = ["easy", "hard"]
547
+
548
+ print(f"šŸŽÆ Running Crafter evaluation with new standardized framework")
549
+ print(f" Model: {model_name}")
550
+ print(f" Modes: {modes}")
551
+ print(f" Trajectories per mode: {n_instances_per_mode}")
552
+
553
+ # Use the new comprehensive evaluation framework
554
+ report = await run_crafter_eval(
555
+ model_names=[model_name],
556
+ difficulties=modes,
557
+ num_trajectories=n_instances_per_mode,
558
+ max_turns=30,
559
+ )
560
+
561
+ # Convert to old format for backward compatibility
562
+ results_for_model = []
563
+ for agg_result in report["raw_aggregate_results"]:
564
+ results_for_model.append(
565
+ {
566
+ "Model": agg_result["model_name"],
567
+ "Difficulty": agg_result["difficulty"],
568
+ "Successful Runs": f"{int(agg_result['success_rate'] * agg_result['num_trajectories'])}/{agg_result['num_trajectories']}",
569
+ "Avg Unique Achievements": f"{agg_result['avg_achievements_per_trajectory']:.2f}",
570
+ }
571
+ )
572
+
573
+ return results_for_model
574
+
575
+
576
+ # Keep the old function for backward compatibility
577
+ async def eval_react_crafter_legacy(
578
+ model_name: str = "gpt-4.1-nano",
579
+ formatting_model_name: str = "gpt-4.1-nano",
580
+ modes: Optional[List[str]] = None,
581
+ n_instances_per_mode: int = 3,
582
+ ) -> List[Dict[str, Any]]:
583
+ """
584
+ LEGACY VERSION - Run ReAct agents on Crafter instances of different difficulties,
585
+ and returns a list of dictionaries containing aggregated results for each mode.
586
+ """
587
+
588
+ if modes is None:
589
+ modes = ["easy", "hard"]
590
+
591
+ current_model_name_for_eval = model_name
592
+
593
+ _temp_llm_for_names = LM(
594
+ model_name=current_model_name_for_eval,
595
+ formatting_model_name=formatting_model_name,
596
+ temperature=0.0,
597
+ )
598
+ _temp_agent_for_names = ReActAgent(_temp_llm_for_names)
599
+ actual_system_name = (
600
+ _temp_agent_for_names.system_name
601
+ ) # Still useful for logging within this func
602
+
603
+ # ------------------------------------------------------------------ helpers
604
+ async def run_episode_eval(
605
+ inst: CrafterTaskInstance, agent_max_turns: int
606
+ ) -> tuple[bool, int, list[str], int]: # Added achievements list and steps taken
607
+ """Run single episode and return (success, num_achievements, achievements_list, steps_taken)"""
608
+ llm = LM(
609
+ model_name=current_model_name_for_eval,
610
+ formatting_model_name=current_model_name_for_eval,
611
+ temperature=0.0,
612
+ )
613
+
614
+ hist_cb = CrafterHistoryObservationCallable(max_history=1)
615
+ env = CrafterClassicEnvironment(inst, custom_step_obs=hist_cb)
616
+ agent = ReActAgent(llm, max_turns=agent_max_turns)
617
+
618
+ obs_payload = await env.initialize()
619
+ if "error" in obs_payload:
620
+ return False, 0, [], 0
621
+
622
+ current_formatted_obs = obs_payload["formatted_obs"]
623
+ raw_obs_for_agent_decision = obs_payload
624
+
625
+ turn_count = 0
626
+ for turn_idx in range(agent.max_turns):
627
+ turn_count += 1
628
+ # Remove noisy progress output
629
+
630
+ action_sequence = await agent.decide(current_formatted_obs, raw_obs_for_agent_decision)
631
+
632
+ if action_sequence == [-1]: # agent terminated
633
+ break
634
+
635
+ # Execute each action in the sequence
636
+ for i, act_idx in enumerate(action_sequence):
637
+ obs_payload_next = await env.step([[CrafterMove(act_idx)]])
638
+
639
+ if "error" in obs_payload_next:
640
+ break # Break out of action sequence on error
641
+
642
+ # Update observation for next action in sequence
643
+ current_formatted_obs = obs_payload_next["formatted_obs"]
644
+ raw_obs_for_agent_decision = obs_payload_next
645
+
646
+ # Check if environment terminated after this sub-action
647
+ if obs_payload_next["private"].terminated or obs_payload_next["private"].truncated:
648
+ break
649
+
650
+ if "error" in obs_payload_next:
651
+ return (
652
+ False,
653
+ len(agent.current_achievements),
654
+ list(agent.current_achievements),
655
+ 0,
656
+ )
657
+
658
+ current_formatted_obs = obs_payload_next["formatted_obs"]
659
+ raw_obs_for_agent_decision = obs_payload_next
660
+ agent.history.append({"type": "tool_response", "content": "Action executed"})
661
+
662
+ obs_payload = obs_payload_next
663
+ if obs_payload["private"].terminated or obs_payload["private"].truncated:
664
+ break
665
+
666
+ final_private_state: CrafterPrivateState = obs_payload["private"]
667
+ final_public_state: CrafterPublicState = obs_payload["public"]
668
+
669
+ run_successful = (final_private_state.terminated or final_private_state.truncated) or len(
670
+ agent.current_achievements
671
+ ) >= 1
672
+
673
+ num_unique_achievements = len(agent.current_achievements)
674
+ achievements_list = list(agent.current_achievements)
675
+ steps_taken = final_public_state.num_steps_taken
676
+
677
+ return run_successful, num_unique_achievements, achievements_list, steps_taken
678
+
679
+ # ---------------------------------------------------------------- instance factory
680
+ async def make_crafter_instances(
681
+ difficulty: str, n_instances: int = 3, start_seed: int = 0
682
+ ) -> List[CrafterTaskInstance]:
683
+ instances = []
684
+ for i in range(n_instances):
685
+ current_seed = start_seed + i
686
+ metadata = CrafterTaskInstanceMetadata(
687
+ difficulty=difficulty,
688
+ seed=current_seed,
689
+ num_trees_radius=0,
690
+ num_cows_radius=0,
691
+ num_hostiles_radius=0,
692
+ )
693
+ instance = CrafterTaskInstance(
694
+ id=uuid.uuid4(),
695
+ impetus=Impetus(
696
+ instructions=f"Survive and unlock achievements in a {difficulty} environment."
697
+ ),
698
+ intent=Intent(rubric={}, gold_trajectories=None, gold_state_diff={}),
699
+ metadata=metadata,
700
+ is_reproducible=True,
701
+ initial_engine_snapshot=None,
702
+ )
703
+ instances.append(instance)
704
+ return instances
705
+
706
+ # ---------------------------------------------------------------- evaluation
707
+ configs = []
708
+ for mode in modes:
709
+ if mode == "easy":
710
+ configs.append(("easy", n_instances_per_mode, 15))
711
+ elif mode == "hard":
712
+ configs.append(("hard", n_instances_per_mode, 15))
713
+
714
+ results_for_model = [] # Stores dicts for each mode for the current model
715
+ base_seed_for_difficulty = {"easy": 1000, "hard": 2000}
716
+
717
+ print(
718
+ f"Starting Crafter ReAct Agent Evaluation for Model: {current_model_name_for_eval}, System: {actual_system_name}"
719
+ )
720
+
721
+ all_generated_task_data = []
722
+ print("\nGenerating task instances...")
723
+ all_tasks_for_eval: Dict[str, List[CrafterTaskInstance]] = {}
724
+ for label, num_agents, _ in configs:
725
+ insts = await make_crafter_instances(
726
+ label, n_instances=num_agents, start_seed=base_seed_for_difficulty[label]
727
+ )
728
+ all_tasks_for_eval[label] = insts
729
+ for inst in insts:
730
+ instance_dict = await inst.serialize()
731
+ all_generated_task_data.append(instance_dict)
732
+ print(f"Generated {len(insts)} instances for {label} difficulty.")
733
+
734
+ dataset_dir = Path(__file__).parent.parent / "dataset"
735
+ dataset_dir.mkdir(parents=True, exist_ok=True)
736
+ synthetic_mix_path = dataset_dir / "synthetic_mix.json"
737
+ with open(synthetic_mix_path, "w") as f:
738
+ json.dump(all_generated_task_data, f, indent=2)
739
+ print(
740
+ f"Saved all {len(all_generated_task_data)} generated task instances to {synthetic_mix_path}"
741
+ )
742
+
743
+ for label, num_agents, max_episode_turns in configs:
744
+ print(
745
+ f"\nRunning {num_agents} agents on {label} difficulty tasks (max_turns: {max_episode_turns}) for model {current_model_name_for_eval}..."
746
+ )
747
+ current_difficulty_instances = all_tasks_for_eval[label]
748
+
749
+ import time
750
+
751
+ start_time = time.time()
752
+ results_per_episode = await asyncio.gather(
753
+ *(run_episode_eval(inst, max_episode_turns) for inst in current_difficulty_instances)
754
+ )
755
+ end_time = time.time()
756
+ print(
757
+ f"Completed {len(current_difficulty_instances)} episodes in {end_time - start_time:.1f}s"
758
+ )
759
+
760
+ # Process detailed results
761
+ successful_episodes = 0
762
+ total_achievements = 0
763
+ detailed_results = []
764
+
765
+ for i, (success, num_achievements, achievements_list, steps_taken) in enumerate(
766
+ results_per_episode
767
+ ):
768
+ episode_result = {
769
+ "episode_id": i + 1,
770
+ "instance_id": current_difficulty_instances[i].id,
771
+ "success": success,
772
+ "achievements_count": num_achievements,
773
+ "achievements": achievements_list,
774
+ "steps_taken": steps_taken,
775
+ "turns_used": "unknown", # Could track this if needed
776
+ }
777
+ detailed_results.append(episode_result)
778
+
779
+ if success:
780
+ successful_episodes += 1
781
+ total_achievements += num_achievements
782
+
783
+ avg_achievements = total_achievements / len(results_per_episode)
784
+
785
+ # Print detailed trajectory information
786
+ print(f"\nšŸ“Š Detailed Results for {model_name} on {label}:")
787
+ print("-" * 80)
788
+ for result in detailed_results:
789
+ status = "āœ… SUCCESS" if result["success"] else "āŒ FAILED"
790
+ achievements_str = (
791
+ ", ".join(result["achievements"]) if result["achievements"] else "None"
792
+ )
793
+ print(
794
+ f"Episode {result['episode_id']}: {status} | "
795
+ f"Steps: {result['steps_taken']} | "
796
+ f"Achievements ({result['achievements_count']}): {achievements_str}"
797
+ )
798
+ print("-" * 80)
799
+
800
+ print(
801
+ f"Completed {label} for model {model_name}: {successful_episodes}/{len(results_per_episode)} successful, Avg. Achievements: {avg_achievements:.2f}"
802
+ )
803
+
804
+ results_for_model.append(
805
+ {
806
+ "Model": model_name,
807
+ "Difficulty": label,
808
+ "Successful Runs": f"{successful_episodes}/{len(results_per_episode)}",
809
+ "Avg Unique Achievements": f"{avg_achievements:.2f}",
810
+ }
811
+ )
812
+
813
+ return results_for_model
814
+
815
+
816
+ async def run_model_comparison_from_config():
817
+ """Run model comparison using parameters from eval_config.toml"""
818
+ # Load configuration
819
+ config_path = Path(__file__).parent / "eval_config.toml"
820
+ if not config_path.exists():
821
+ raise FileNotFoundError(f"Configuration file not found: {config_path}")
822
+
823
+ config = toml.load(config_path)
824
+ eval_config = config["evaluation"]
825
+
826
+ models = eval_config["models"]
827
+ difficulties = eval_config["difficulties"]
828
+ max_turns = eval_config["max_turns"]
829
+ n_trajectories = eval_config["trajectories_per_condition"]
830
+
831
+ # Update global max_turns from config
832
+ global agent_max_turns
833
+ agent_max_turns = max_turns
834
+
835
+ print("šŸŽÆ Crafter Multi-Action Model Comparison")
836
+ print("=" * 50)
837
+ print(f"Models: {', '.join(models)}")
838
+ print(f"Difficulties: {', '.join(difficulties)}")
839
+ print(f"Max turns: {max_turns}")
840
+ print(f"Trajectories per condition: {n_trajectories}")
841
+ print("=" * 50)
842
+
843
+ all_results = []
844
+
845
+ for model_name in models:
846
+ print(f"\nšŸ¤– Running {model_name}...")
847
+
848
+ # Update the global variable for the model
849
+ global current_model_name_for_eval
850
+ current_model_name_for_eval = model_name
851
+
852
+ model_results = await eval_react_crafter_legacy(
853
+ model_name=model_name,
854
+ formatting_model_name=model_name,
855
+ modes=difficulties,
856
+ n_instances_per_mode=n_trajectories,
857
+ )
858
+
859
+ all_results.extend(model_results)
860
+ print(f"āœ… {model_name} completed")
861
+
862
+ print("\n" + "=" * 60)
863
+ print("šŸ† FINAL COMPARISON RESULTS")
864
+ print("=" * 60)
865
+
866
+ from tabulate import tabulate
867
+
868
+ print(tabulate(all_results, headers="keys", tablefmt="github"))
869
+
870
+ return all_results
871
+
872
+
873
+ if __name__ == "__main__":
874
+ asyncio.run(run_model_comparison_from_config())