amd-gaia 0.14.3__py3-none-any.whl → 0.15.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (181) hide show
  1. {amd_gaia-0.14.3.dist-info → amd_gaia-0.15.1.dist-info}/METADATA +223 -223
  2. amd_gaia-0.15.1.dist-info/RECORD +178 -0
  3. {amd_gaia-0.14.3.dist-info → amd_gaia-0.15.1.dist-info}/entry_points.txt +1 -0
  4. {amd_gaia-0.14.3.dist-info → amd_gaia-0.15.1.dist-info}/licenses/LICENSE.md +20 -20
  5. gaia/__init__.py +29 -29
  6. gaia/agents/__init__.py +19 -19
  7. gaia/agents/base/__init__.py +9 -9
  8. gaia/agents/base/agent.py +2177 -2177
  9. gaia/agents/base/api_agent.py +120 -120
  10. gaia/agents/base/console.py +1841 -1841
  11. gaia/agents/base/errors.py +237 -237
  12. gaia/agents/base/mcp_agent.py +86 -86
  13. gaia/agents/base/tools.py +83 -83
  14. gaia/agents/blender/agent.py +556 -556
  15. gaia/agents/blender/agent_simple.py +133 -135
  16. gaia/agents/blender/app.py +211 -211
  17. gaia/agents/blender/app_simple.py +41 -41
  18. gaia/agents/blender/core/__init__.py +16 -16
  19. gaia/agents/blender/core/materials.py +506 -506
  20. gaia/agents/blender/core/objects.py +316 -316
  21. gaia/agents/blender/core/rendering.py +225 -225
  22. gaia/agents/blender/core/scene.py +220 -220
  23. gaia/agents/blender/core/view.py +146 -146
  24. gaia/agents/chat/__init__.py +9 -9
  25. gaia/agents/chat/agent.py +835 -835
  26. gaia/agents/chat/app.py +1058 -1058
  27. gaia/agents/chat/session.py +508 -508
  28. gaia/agents/chat/tools/__init__.py +15 -15
  29. gaia/agents/chat/tools/file_tools.py +96 -96
  30. gaia/agents/chat/tools/rag_tools.py +1729 -1729
  31. gaia/agents/chat/tools/shell_tools.py +436 -436
  32. gaia/agents/code/__init__.py +7 -7
  33. gaia/agents/code/agent.py +549 -549
  34. gaia/agents/code/cli.py +377 -0
  35. gaia/agents/code/models.py +135 -135
  36. gaia/agents/code/orchestration/__init__.py +24 -24
  37. gaia/agents/code/orchestration/checklist_executor.py +1763 -1763
  38. gaia/agents/code/orchestration/checklist_generator.py +713 -713
  39. gaia/agents/code/orchestration/factories/__init__.py +9 -9
  40. gaia/agents/code/orchestration/factories/base.py +63 -63
  41. gaia/agents/code/orchestration/factories/nextjs_factory.py +118 -118
  42. gaia/agents/code/orchestration/factories/python_factory.py +106 -106
  43. gaia/agents/code/orchestration/orchestrator.py +841 -841
  44. gaia/agents/code/orchestration/project_analyzer.py +391 -391
  45. gaia/agents/code/orchestration/steps/__init__.py +67 -67
  46. gaia/agents/code/orchestration/steps/base.py +188 -188
  47. gaia/agents/code/orchestration/steps/error_handler.py +314 -314
  48. gaia/agents/code/orchestration/steps/nextjs.py +828 -828
  49. gaia/agents/code/orchestration/steps/python.py +307 -307
  50. gaia/agents/code/orchestration/template_catalog.py +469 -469
  51. gaia/agents/code/orchestration/workflows/__init__.py +14 -14
  52. gaia/agents/code/orchestration/workflows/base.py +80 -80
  53. gaia/agents/code/orchestration/workflows/nextjs.py +186 -186
  54. gaia/agents/code/orchestration/workflows/python.py +94 -94
  55. gaia/agents/code/prompts/__init__.py +11 -11
  56. gaia/agents/code/prompts/base_prompt.py +77 -77
  57. gaia/agents/code/prompts/code_patterns.py +2036 -2036
  58. gaia/agents/code/prompts/nextjs_prompt.py +40 -40
  59. gaia/agents/code/prompts/python_prompt.py +109 -109
  60. gaia/agents/code/schema_inference.py +365 -365
  61. gaia/agents/code/system_prompt.py +41 -41
  62. gaia/agents/code/tools/__init__.py +42 -42
  63. gaia/agents/code/tools/cli_tools.py +1138 -1138
  64. gaia/agents/code/tools/code_formatting.py +319 -319
  65. gaia/agents/code/tools/code_tools.py +769 -769
  66. gaia/agents/code/tools/error_fixing.py +1347 -1347
  67. gaia/agents/code/tools/external_tools.py +180 -180
  68. gaia/agents/code/tools/file_io.py +845 -845
  69. gaia/agents/code/tools/prisma_tools.py +190 -190
  70. gaia/agents/code/tools/project_management.py +1016 -1016
  71. gaia/agents/code/tools/testing.py +321 -321
  72. gaia/agents/code/tools/typescript_tools.py +122 -122
  73. gaia/agents/code/tools/validation_parsing.py +461 -461
  74. gaia/agents/code/tools/validation_tools.py +806 -806
  75. gaia/agents/code/tools/web_dev_tools.py +1758 -1758
  76. gaia/agents/code/validators/__init__.py +16 -16
  77. gaia/agents/code/validators/antipattern_checker.py +241 -241
  78. gaia/agents/code/validators/ast_analyzer.py +197 -197
  79. gaia/agents/code/validators/requirements_validator.py +145 -145
  80. gaia/agents/code/validators/syntax_validator.py +171 -171
  81. gaia/agents/docker/__init__.py +7 -7
  82. gaia/agents/docker/agent.py +642 -642
  83. gaia/agents/emr/__init__.py +8 -8
  84. gaia/agents/emr/agent.py +1506 -1506
  85. gaia/agents/emr/cli.py +1322 -1322
  86. gaia/agents/emr/constants.py +475 -475
  87. gaia/agents/emr/dashboard/__init__.py +4 -4
  88. gaia/agents/emr/dashboard/server.py +1974 -1974
  89. gaia/agents/jira/__init__.py +11 -11
  90. gaia/agents/jira/agent.py +894 -894
  91. gaia/agents/jira/jql_templates.py +299 -299
  92. gaia/agents/routing/__init__.py +7 -7
  93. gaia/agents/routing/agent.py +567 -570
  94. gaia/agents/routing/system_prompt.py +75 -75
  95. gaia/agents/summarize/__init__.py +11 -0
  96. gaia/agents/summarize/agent.py +885 -0
  97. gaia/agents/summarize/prompts.py +129 -0
  98. gaia/api/__init__.py +23 -23
  99. gaia/api/agent_registry.py +238 -238
  100. gaia/api/app.py +305 -305
  101. gaia/api/openai_server.py +575 -575
  102. gaia/api/schemas.py +186 -186
  103. gaia/api/sse_handler.py +373 -373
  104. gaia/apps/__init__.py +4 -4
  105. gaia/apps/llm/__init__.py +6 -6
  106. gaia/apps/llm/app.py +173 -169
  107. gaia/apps/summarize/app.py +116 -633
  108. gaia/apps/summarize/html_viewer.py +133 -133
  109. gaia/apps/summarize/pdf_formatter.py +284 -284
  110. gaia/audio/__init__.py +2 -2
  111. gaia/audio/audio_client.py +439 -439
  112. gaia/audio/audio_recorder.py +269 -269
  113. gaia/audio/kokoro_tts.py +599 -599
  114. gaia/audio/whisper_asr.py +432 -432
  115. gaia/chat/__init__.py +16 -16
  116. gaia/chat/app.py +430 -430
  117. gaia/chat/prompts.py +522 -522
  118. gaia/chat/sdk.py +1228 -1225
  119. gaia/cli.py +5481 -5621
  120. gaia/database/__init__.py +10 -10
  121. gaia/database/agent.py +176 -176
  122. gaia/database/mixin.py +290 -290
  123. gaia/database/testing.py +64 -64
  124. gaia/eval/batch_experiment.py +2332 -2332
  125. gaia/eval/claude.py +542 -542
  126. gaia/eval/config.py +37 -37
  127. gaia/eval/email_generator.py +512 -512
  128. gaia/eval/eval.py +3179 -3179
  129. gaia/eval/groundtruth.py +1130 -1130
  130. gaia/eval/transcript_generator.py +582 -582
  131. gaia/eval/webapp/README.md +167 -167
  132. gaia/eval/webapp/package-lock.json +875 -875
  133. gaia/eval/webapp/package.json +20 -20
  134. gaia/eval/webapp/public/app.js +3402 -3402
  135. gaia/eval/webapp/public/index.html +87 -87
  136. gaia/eval/webapp/public/styles.css +3661 -3661
  137. gaia/eval/webapp/server.js +415 -415
  138. gaia/eval/webapp/test-setup.js +72 -72
  139. gaia/llm/__init__.py +9 -2
  140. gaia/llm/base_client.py +60 -0
  141. gaia/llm/exceptions.py +12 -0
  142. gaia/llm/factory.py +70 -0
  143. gaia/llm/lemonade_client.py +3236 -3221
  144. gaia/llm/lemonade_manager.py +294 -294
  145. gaia/llm/providers/__init__.py +9 -0
  146. gaia/llm/providers/claude.py +108 -0
  147. gaia/llm/providers/lemonade.py +120 -0
  148. gaia/llm/providers/openai_provider.py +79 -0
  149. gaia/llm/vlm_client.py +382 -382
  150. gaia/logger.py +189 -189
  151. gaia/mcp/agent_mcp_server.py +245 -245
  152. gaia/mcp/blender_mcp_client.py +138 -138
  153. gaia/mcp/blender_mcp_server.py +648 -648
  154. gaia/mcp/context7_cache.py +332 -332
  155. gaia/mcp/external_services.py +518 -518
  156. gaia/mcp/mcp_bridge.py +811 -550
  157. gaia/mcp/servers/__init__.py +6 -6
  158. gaia/mcp/servers/docker_mcp.py +83 -83
  159. gaia/perf_analysis.py +361 -0
  160. gaia/rag/__init__.py +10 -10
  161. gaia/rag/app.py +293 -293
  162. gaia/rag/demo.py +304 -304
  163. gaia/rag/pdf_utils.py +235 -235
  164. gaia/rag/sdk.py +2194 -2194
  165. gaia/security.py +163 -163
  166. gaia/talk/app.py +289 -289
  167. gaia/talk/sdk.py +538 -538
  168. gaia/testing/__init__.py +87 -87
  169. gaia/testing/assertions.py +330 -330
  170. gaia/testing/fixtures.py +333 -333
  171. gaia/testing/mocks.py +493 -493
  172. gaia/util.py +46 -46
  173. gaia/utils/__init__.py +33 -33
  174. gaia/utils/file_watcher.py +675 -675
  175. gaia/utils/parsing.py +223 -223
  176. gaia/version.py +100 -100
  177. amd_gaia-0.14.3.dist-info/RECORD +0 -168
  178. gaia/agents/code/app.py +0 -266
  179. gaia/llm/llm_client.py +0 -729
  180. {amd_gaia-0.14.3.dist-info → amd_gaia-0.15.1.dist-info}/WHEEL +0 -0
  181. {amd_gaia-0.14.3.dist-info → amd_gaia-0.15.1.dist-info}/top_level.txt +0 -0
@@ -1,332 +1,332 @@
1
- # Copyright(C) 2025-2026 Advanced Micro Devices, Inc. All rights reserved.
2
- # SPDX-License-Identifier: MIT
3
- """Persistent cache and rate protection for Context7 API calls."""
4
-
5
- import hashlib
6
- import json
7
- from dataclasses import asdict, dataclass
8
- from datetime import datetime, timedelta
9
- from pathlib import Path
10
- from typing import Dict, Optional
11
-
12
- from gaia.logger import get_logger
13
-
14
- logger = get_logger(__name__)
15
-
16
-
17
- class Context7Cache:
18
- """File-based persistent cache for Context7 results.
19
-
20
- Caches library ID mappings and documentation across sessions to reduce API calls.
21
- """
22
-
23
- # TTL values
24
- TTL_LIBRARY_ID = timedelta(days=7)
25
- TTL_DOCUMENTATION = timedelta(hours=24)
26
- TTL_FAILED = timedelta(hours=1)
27
-
28
- def __init__(self, cache_dir: Optional[Path] = None):
29
- """Initialize Context7 cache.
30
-
31
- Args:
32
- cache_dir: Optional custom cache directory (defaults to ~/.gaia/cache/context7)
33
- """
34
- self.cache_dir = cache_dir or Path.home() / ".gaia" / "cache" / "context7"
35
- self.cache_dir.mkdir(parents=True, exist_ok=True)
36
- self.docs_dir = self.cache_dir / "documentation"
37
- self.docs_dir.mkdir(exist_ok=True)
38
-
39
- self.library_ids_file = self.cache_dir / "library_ids.json"
40
- self.rate_state_file = self.cache_dir / "rate_state.json"
41
-
42
- def get_library_id(self, library_name: str) -> Optional[str]:
43
- """Get cached library ID if valid.
44
-
45
- Args:
46
- library_name: Library name to lookup (e.g., "nextjs")
47
-
48
- Returns:
49
- Cached library ID or None if not found/expired
50
- """
51
- cache = self._load_json(self.library_ids_file)
52
- key = library_name.lower()
53
-
54
- if key in cache:
55
- entry = cache[key]
56
- if self._is_valid(entry, self.TTL_LIBRARY_ID):
57
- return entry["value"]
58
-
59
- return None
60
-
61
- def set_library_id(self, library_name: str, library_id: Optional[str]):
62
- """Cache library ID resolution.
63
-
64
- Args:
65
- library_name: Library name (e.g., "nextjs")
66
- library_id: Resolved Context7 ID (e.g., "/vercel/next.js") or None if failed
67
- """
68
- cache = self._load_json(self.library_ids_file)
69
- cache[library_name.lower()] = {
70
- "value": library_id,
71
- "timestamp": datetime.now().isoformat(),
72
- }
73
- self._save_json(self.library_ids_file, cache)
74
-
75
- def get_documentation(self, library: str, query: str) -> Optional[str]:
76
- """Get cached documentation if valid.
77
-
78
- Args:
79
- library: Library identifier
80
- query: Documentation query
81
-
82
- Returns:
83
- Cached documentation content or None if not found/expired
84
- """
85
- cache_file = self._doc_cache_file(library, query)
86
- if cache_file.exists():
87
- entry = self._load_json(cache_file)
88
- if self._is_valid(entry, self.TTL_DOCUMENTATION):
89
- return entry.get("content")
90
-
91
- return None
92
-
93
- def set_documentation(self, library: str, query: str, content: str):
94
- """Cache documentation result.
95
-
96
- Args:
97
- library: Library identifier
98
- query: Documentation query
99
- content: Documentation content to cache
100
- """
101
- cache_file = self._doc_cache_file(library, query)
102
- self._save_json(
103
- cache_file, {"content": content, "timestamp": datetime.now().isoformat()}
104
- )
105
-
106
- def _doc_cache_file(self, library: str, query: str) -> Path:
107
- """Generate cache filename for documentation.
108
-
109
- Args:
110
- library: Library identifier
111
- query: Documentation query
112
-
113
- Returns:
114
- Path to cache file
115
- """
116
- key = f"{library}:{query}"
117
- hash_key = hashlib.md5(key.encode()).hexdigest()[:12]
118
- safe_lib = library.replace("/", "_").replace(".", "_")
119
- return self.docs_dir / f"{safe_lib}_{hash_key}.json"
120
-
121
- def _is_valid(self, entry: Dict, ttl: timedelta) -> bool:
122
- """Check if cache entry is still valid.
123
-
124
- Args:
125
- entry: Cache entry dict
126
- ttl: Time-to-live duration
127
-
128
- Returns:
129
- True if entry is valid and not expired
130
- """
131
- if not entry or "timestamp" not in entry:
132
- return False
133
-
134
- try:
135
- timestamp = datetime.fromisoformat(entry["timestamp"])
136
- return datetime.now() - timestamp < ttl
137
- except (ValueError, TypeError):
138
- return False
139
-
140
- def _load_json(self, path: Path) -> Dict:
141
- """Load JSON file or return empty dict.
142
-
143
- Args:
144
- path: Path to JSON file
145
-
146
- Returns:
147
- Loaded data or empty dict
148
- """
149
- if path.exists():
150
- try:
151
- return json.loads(path.read_text(encoding="utf-8"))
152
- except (json.JSONDecodeError, OSError):
153
- logger.warning(f"Failed to load cache file: {path}")
154
- return {}
155
- return {}
156
-
157
- def _save_json(self, path: Path, data: Dict):
158
- """Save data to JSON file.
159
-
160
- Args:
161
- path: Path to JSON file
162
- data: Data to save
163
- """
164
- try:
165
- path.write_text(json.dumps(data, indent=2), encoding="utf-8")
166
- except OSError as e:
167
- logger.error(f"Failed to save cache file: {path}: {e}")
168
-
169
- def clear(self):
170
- """Clear all cached data."""
171
- if self.library_ids_file.exists():
172
- self.library_ids_file.unlink()
173
-
174
- for f in self.docs_dir.glob("*.json"):
175
- f.unlink()
176
-
177
- logger.info("Context7 cache cleared")
178
-
179
-
180
- @dataclass
181
- class RateState:
182
- """Rate limiter state persisted to disk."""
183
-
184
- tokens: float = 30.0 # Available tokens
185
- last_update: str = "" # ISO timestamp
186
- consecutive_failures: int = 0
187
- circuit_open_until: str = "" # ISO timestamp if open
188
-
189
- # Constants
190
- MAX_TOKENS: float = 30.0
191
- REFILL_RATE: float = 0.5 # tokens per minute (30/hour)
192
- CIRCUIT_OPEN_DURATION: int = 300 # 5 minutes
193
-
194
-
195
- class Context7RateLimiter:
196
- """Token bucket rate limiter with circuit breaker.
197
-
198
- Protects against Context7 rate limiting by:
199
- - Limiting requests to 30/hour (0.5/minute refill)
200
- - Opening circuit breaker after 5 consecutive failures
201
- - Reopening circuit after 5 minute cooldown
202
- """
203
-
204
- def __init__(self, state_file: Optional[Path] = None):
205
- """Initialize rate limiter.
206
-
207
- Args:
208
- state_file: Optional custom state file path
209
- """
210
- self.state_file = state_file or (
211
- Path.home() / ".gaia" / "cache" / "context7" / "rate_state.json"
212
- )
213
- self.state_file.parent.mkdir(parents=True, exist_ok=True)
214
- self.state = self._load_state()
215
-
216
- def can_make_request(self) -> tuple[bool, str]:
217
- """Check if we can make a request.
218
-
219
- Returns:
220
- Tuple of (can_proceed, reason)
221
- """
222
- self._refill_tokens()
223
-
224
- # Check circuit breaker
225
- if self.state.circuit_open_until:
226
- try:
227
- open_until = datetime.fromisoformat(self.state.circuit_open_until)
228
- if datetime.now() < open_until:
229
- remaining = (open_until - datetime.now()).seconds
230
- return False, f"Circuit breaker open. Retry in {remaining}s"
231
- else:
232
- # Circuit recovered
233
- self.state.circuit_open_until = ""
234
- self.state.consecutive_failures = 0
235
- logger.info("Circuit breaker closed - recovered from failures")
236
- except (ValueError, TypeError):
237
- # Invalid timestamp, clear it
238
- self.state.circuit_open_until = ""
239
-
240
- # Check tokens
241
- if self.state.tokens < 1.0:
242
- return False, "Rate limit reached. Try again in a few minutes."
243
-
244
- return True, "OK"
245
-
246
- def consume_token(self):
247
- """Consume a token for a request."""
248
- self._refill_tokens()
249
- self.state.tokens = max(0, self.state.tokens - 1)
250
- self._save_state()
251
-
252
- def record_success(self):
253
- """Record successful request."""
254
- self.state.consecutive_failures = 0
255
- self._save_state()
256
-
257
- def record_failure(self, is_rate_limit: bool = False):
258
- """Record failed request.
259
-
260
- Args:
261
- is_rate_limit: True if failure was due to rate limiting (HTTP 429)
262
- """
263
- self.state.consecutive_failures += 1
264
-
265
- # Open circuit on rate limit or too many failures
266
- if is_rate_limit or self.state.consecutive_failures >= 5:
267
- open_until = datetime.now() + timedelta(
268
- seconds=self.state.CIRCUIT_OPEN_DURATION
269
- )
270
- self.state.circuit_open_until = open_until.isoformat()
271
- logger.warning(
272
- f"Circuit breaker opened until {open_until} "
273
- f"(failures: {self.state.consecutive_failures})"
274
- )
275
-
276
- self._save_state()
277
-
278
- def _refill_tokens(self):
279
- """Refill tokens based on time elapsed."""
280
- now = datetime.now()
281
-
282
- if self.state.last_update:
283
- try:
284
- last = datetime.fromisoformat(self.state.last_update)
285
- elapsed_minutes = (now - last).total_seconds() / 60
286
- refill = elapsed_minutes * self.state.REFILL_RATE
287
- self.state.tokens = min(
288
- self.state.MAX_TOKENS, self.state.tokens + refill
289
- )
290
- except (ValueError, TypeError):
291
- # Invalid timestamp, reset
292
- self.state.tokens = self.state.MAX_TOKENS
293
-
294
- self.state.last_update = now.isoformat()
295
-
296
- def _load_state(self) -> RateState:
297
- """Load state from disk.
298
-
299
- Returns:
300
- Loaded rate state or new state if file doesn't exist
301
- """
302
- if self.state_file.exists():
303
- try:
304
- data = json.loads(self.state_file.read_text(encoding="utf-8"))
305
- return RateState(**data)
306
- except (json.JSONDecodeError, TypeError, OSError) as e:
307
- logger.warning(f"Failed to load rate state: {e}, creating new state")
308
-
309
- return RateState(last_update=datetime.now().isoformat())
310
-
311
- def _save_state(self):
312
- """Save state to disk."""
313
- try:
314
- self.state_file.write_text(
315
- json.dumps(asdict(self.state), indent=2), encoding="utf-8"
316
- )
317
- except OSError as e:
318
- logger.error(f"Failed to save rate state: {e}")
319
-
320
- def get_status(self) -> dict:
321
- """Get current rate limiter status.
322
-
323
- Returns:
324
- Dict with status information
325
- """
326
- self._refill_tokens()
327
- return {
328
- "tokens_available": round(self.state.tokens, 1),
329
- "max_tokens": self.state.MAX_TOKENS,
330
- "circuit_open": bool(self.state.circuit_open_until),
331
- "consecutive_failures": self.state.consecutive_failures,
332
- }
1
+ # Copyright(C) 2025-2026 Advanced Micro Devices, Inc. All rights reserved.
2
+ # SPDX-License-Identifier: MIT
3
+ """Persistent cache and rate protection for Context7 API calls."""
4
+
5
+ import hashlib
6
+ import json
7
+ from dataclasses import asdict, dataclass
8
+ from datetime import datetime, timedelta
9
+ from pathlib import Path
10
+ from typing import Dict, Optional
11
+
12
+ from gaia.logger import get_logger
13
+
14
+ logger = get_logger(__name__)
15
+
16
+
17
+ class Context7Cache:
18
+ """File-based persistent cache for Context7 results.
19
+
20
+ Caches library ID mappings and documentation across sessions to reduce API calls.
21
+ """
22
+
23
+ # TTL values
24
+ TTL_LIBRARY_ID = timedelta(days=7)
25
+ TTL_DOCUMENTATION = timedelta(hours=24)
26
+ TTL_FAILED = timedelta(hours=1)
27
+
28
+ def __init__(self, cache_dir: Optional[Path] = None):
29
+ """Initialize Context7 cache.
30
+
31
+ Args:
32
+ cache_dir: Optional custom cache directory (defaults to ~/.gaia/cache/context7)
33
+ """
34
+ self.cache_dir = cache_dir or Path.home() / ".gaia" / "cache" / "context7"
35
+ self.cache_dir.mkdir(parents=True, exist_ok=True)
36
+ self.docs_dir = self.cache_dir / "documentation"
37
+ self.docs_dir.mkdir(exist_ok=True)
38
+
39
+ self.library_ids_file = self.cache_dir / "library_ids.json"
40
+ self.rate_state_file = self.cache_dir / "rate_state.json"
41
+
42
+ def get_library_id(self, library_name: str) -> Optional[str]:
43
+ """Get cached library ID if valid.
44
+
45
+ Args:
46
+ library_name: Library name to lookup (e.g., "nextjs")
47
+
48
+ Returns:
49
+ Cached library ID or None if not found/expired
50
+ """
51
+ cache = self._load_json(self.library_ids_file)
52
+ key = library_name.lower()
53
+
54
+ if key in cache:
55
+ entry = cache[key]
56
+ if self._is_valid(entry, self.TTL_LIBRARY_ID):
57
+ return entry["value"]
58
+
59
+ return None
60
+
61
+ def set_library_id(self, library_name: str, library_id: Optional[str]):
62
+ """Cache library ID resolution.
63
+
64
+ Args:
65
+ library_name: Library name (e.g., "nextjs")
66
+ library_id: Resolved Context7 ID (e.g., "/vercel/next.js") or None if failed
67
+ """
68
+ cache = self._load_json(self.library_ids_file)
69
+ cache[library_name.lower()] = {
70
+ "value": library_id,
71
+ "timestamp": datetime.now().isoformat(),
72
+ }
73
+ self._save_json(self.library_ids_file, cache)
74
+
75
+ def get_documentation(self, library: str, query: str) -> Optional[str]:
76
+ """Get cached documentation if valid.
77
+
78
+ Args:
79
+ library: Library identifier
80
+ query: Documentation query
81
+
82
+ Returns:
83
+ Cached documentation content or None if not found/expired
84
+ """
85
+ cache_file = self._doc_cache_file(library, query)
86
+ if cache_file.exists():
87
+ entry = self._load_json(cache_file)
88
+ if self._is_valid(entry, self.TTL_DOCUMENTATION):
89
+ return entry.get("content")
90
+
91
+ return None
92
+
93
+ def set_documentation(self, library: str, query: str, content: str):
94
+ """Cache documentation result.
95
+
96
+ Args:
97
+ library: Library identifier
98
+ query: Documentation query
99
+ content: Documentation content to cache
100
+ """
101
+ cache_file = self._doc_cache_file(library, query)
102
+ self._save_json(
103
+ cache_file, {"content": content, "timestamp": datetime.now().isoformat()}
104
+ )
105
+
106
+ def _doc_cache_file(self, library: str, query: str) -> Path:
107
+ """Generate cache filename for documentation.
108
+
109
+ Args:
110
+ library: Library identifier
111
+ query: Documentation query
112
+
113
+ Returns:
114
+ Path to cache file
115
+ """
116
+ key = f"{library}:{query}"
117
+ hash_key = hashlib.md5(key.encode()).hexdigest()[:12]
118
+ safe_lib = library.replace("/", "_").replace(".", "_")
119
+ return self.docs_dir / f"{safe_lib}_{hash_key}.json"
120
+
121
+ def _is_valid(self, entry: Dict, ttl: timedelta) -> bool:
122
+ """Check if cache entry is still valid.
123
+
124
+ Args:
125
+ entry: Cache entry dict
126
+ ttl: Time-to-live duration
127
+
128
+ Returns:
129
+ True if entry is valid and not expired
130
+ """
131
+ if not entry or "timestamp" not in entry:
132
+ return False
133
+
134
+ try:
135
+ timestamp = datetime.fromisoformat(entry["timestamp"])
136
+ return datetime.now() - timestamp < ttl
137
+ except (ValueError, TypeError):
138
+ return False
139
+
140
+ def _load_json(self, path: Path) -> Dict:
141
+ """Load JSON file or return empty dict.
142
+
143
+ Args:
144
+ path: Path to JSON file
145
+
146
+ Returns:
147
+ Loaded data or empty dict
148
+ """
149
+ if path.exists():
150
+ try:
151
+ return json.loads(path.read_text(encoding="utf-8"))
152
+ except (json.JSONDecodeError, OSError):
153
+ logger.warning(f"Failed to load cache file: {path}")
154
+ return {}
155
+ return {}
156
+
157
+ def _save_json(self, path: Path, data: Dict):
158
+ """Save data to JSON file.
159
+
160
+ Args:
161
+ path: Path to JSON file
162
+ data: Data to save
163
+ """
164
+ try:
165
+ path.write_text(json.dumps(data, indent=2), encoding="utf-8")
166
+ except OSError as e:
167
+ logger.error(f"Failed to save cache file: {path}: {e}")
168
+
169
+ def clear(self):
170
+ """Clear all cached data."""
171
+ if self.library_ids_file.exists():
172
+ self.library_ids_file.unlink()
173
+
174
+ for f in self.docs_dir.glob("*.json"):
175
+ f.unlink()
176
+
177
+ logger.info("Context7 cache cleared")
178
+
179
+
180
+ @dataclass
181
+ class RateState:
182
+ """Rate limiter state persisted to disk."""
183
+
184
+ tokens: float = 30.0 # Available tokens
185
+ last_update: str = "" # ISO timestamp
186
+ consecutive_failures: int = 0
187
+ circuit_open_until: str = "" # ISO timestamp if open
188
+
189
+ # Constants
190
+ MAX_TOKENS: float = 30.0
191
+ REFILL_RATE: float = 0.5 # tokens per minute (30/hour)
192
+ CIRCUIT_OPEN_DURATION: int = 300 # 5 minutes
193
+
194
+
195
+ class Context7RateLimiter:
196
+ """Token bucket rate limiter with circuit breaker.
197
+
198
+ Protects against Context7 rate limiting by:
199
+ - Limiting requests to 30/hour (0.5/minute refill)
200
+ - Opening circuit breaker after 5 consecutive failures
201
+ - Reopening circuit after 5 minute cooldown
202
+ """
203
+
204
+ def __init__(self, state_file: Optional[Path] = None):
205
+ """Initialize rate limiter.
206
+
207
+ Args:
208
+ state_file: Optional custom state file path
209
+ """
210
+ self.state_file = state_file or (
211
+ Path.home() / ".gaia" / "cache" / "context7" / "rate_state.json"
212
+ )
213
+ self.state_file.parent.mkdir(parents=True, exist_ok=True)
214
+ self.state = self._load_state()
215
+
216
+ def can_make_request(self) -> tuple[bool, str]:
217
+ """Check if we can make a request.
218
+
219
+ Returns:
220
+ Tuple of (can_proceed, reason)
221
+ """
222
+ self._refill_tokens()
223
+
224
+ # Check circuit breaker
225
+ if self.state.circuit_open_until:
226
+ try:
227
+ open_until = datetime.fromisoformat(self.state.circuit_open_until)
228
+ if datetime.now() < open_until:
229
+ remaining = (open_until - datetime.now()).seconds
230
+ return False, f"Circuit breaker open. Retry in {remaining}s"
231
+ else:
232
+ # Circuit recovered
233
+ self.state.circuit_open_until = ""
234
+ self.state.consecutive_failures = 0
235
+ logger.info("Circuit breaker closed - recovered from failures")
236
+ except (ValueError, TypeError):
237
+ # Invalid timestamp, clear it
238
+ self.state.circuit_open_until = ""
239
+
240
+ # Check tokens
241
+ if self.state.tokens < 1.0:
242
+ return False, "Rate limit reached. Try again in a few minutes."
243
+
244
+ return True, "OK"
245
+
246
+ def consume_token(self):
247
+ """Consume a token for a request."""
248
+ self._refill_tokens()
249
+ self.state.tokens = max(0, self.state.tokens - 1)
250
+ self._save_state()
251
+
252
+ def record_success(self):
253
+ """Record successful request."""
254
+ self.state.consecutive_failures = 0
255
+ self._save_state()
256
+
257
+ def record_failure(self, is_rate_limit: bool = False):
258
+ """Record failed request.
259
+
260
+ Args:
261
+ is_rate_limit: True if failure was due to rate limiting (HTTP 429)
262
+ """
263
+ self.state.consecutive_failures += 1
264
+
265
+ # Open circuit on rate limit or too many failures
266
+ if is_rate_limit or self.state.consecutive_failures >= 5:
267
+ open_until = datetime.now() + timedelta(
268
+ seconds=self.state.CIRCUIT_OPEN_DURATION
269
+ )
270
+ self.state.circuit_open_until = open_until.isoformat()
271
+ logger.warning(
272
+ f"Circuit breaker opened until {open_until} "
273
+ f"(failures: {self.state.consecutive_failures})"
274
+ )
275
+
276
+ self._save_state()
277
+
278
+ def _refill_tokens(self):
279
+ """Refill tokens based on time elapsed."""
280
+ now = datetime.now()
281
+
282
+ if self.state.last_update:
283
+ try:
284
+ last = datetime.fromisoformat(self.state.last_update)
285
+ elapsed_minutes = (now - last).total_seconds() / 60
286
+ refill = elapsed_minutes * self.state.REFILL_RATE
287
+ self.state.tokens = min(
288
+ self.state.MAX_TOKENS, self.state.tokens + refill
289
+ )
290
+ except (ValueError, TypeError):
291
+ # Invalid timestamp, reset
292
+ self.state.tokens = self.state.MAX_TOKENS
293
+
294
+ self.state.last_update = now.isoformat()
295
+
296
+ def _load_state(self) -> RateState:
297
+ """Load state from disk.
298
+
299
+ Returns:
300
+ Loaded rate state or new state if file doesn't exist
301
+ """
302
+ if self.state_file.exists():
303
+ try:
304
+ data = json.loads(self.state_file.read_text(encoding="utf-8"))
305
+ return RateState(**data)
306
+ except (json.JSONDecodeError, TypeError, OSError) as e:
307
+ logger.warning(f"Failed to load rate state: {e}, creating new state")
308
+
309
+ return RateState(last_update=datetime.now().isoformat())
310
+
311
+ def _save_state(self):
312
+ """Save state to disk."""
313
+ try:
314
+ self.state_file.write_text(
315
+ json.dumps(asdict(self.state), indent=2), encoding="utf-8"
316
+ )
317
+ except OSError as e:
318
+ logger.error(f"Failed to save rate state: {e}")
319
+
320
+ def get_status(self) -> dict:
321
+ """Get current rate limiter status.
322
+
323
+ Returns:
324
+ Dict with status information
325
+ """
326
+ self._refill_tokens()
327
+ return {
328
+ "tokens_available": round(self.state.tokens, 1),
329
+ "max_tokens": self.state.MAX_TOKENS,
330
+ "circuit_open": bool(self.state.circuit_open_until),
331
+ "consecutive_failures": self.state.consecutive_failures,
332
+ }