@pentatonic-ai/ai-agent-sdk 0.7.7 → 0.7.8
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@pentatonic-ai/ai-agent-sdk",
|
|
3
|
-
"version": "0.7.
|
|
3
|
+
"version": "0.7.8",
|
|
4
4
|
"description": "TES SDK — LLM observability and lifecycle tracking via Pentatonic Thing Event System. Track token usage, tool calls, and conversations. Manage things through event-sourced lifecycle stages with AI enrichment and vector search.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "./dist/index.cjs",
|
|
@@ -438,12 +438,25 @@ async def health():
|
|
|
438
438
|
return out
|
|
439
439
|
|
|
440
440
|
|
|
441
|
+
def _arena_scoped_rid(arena: str, content: str) -> str:
|
|
442
|
+
"""Stable, arena-scoped record id.
|
|
443
|
+
|
|
444
|
+
Same content in different arenas hashes to *different* ids — so two
|
|
445
|
+
tenants storing the byte-identical chunk get distinct primary keys
|
|
446
|
+
in the underlying stores (L4 sqlite-vec, L5 Milvus, L6 Milvus).
|
|
447
|
+
Without this, arena=tenant-b's write would overwrite arena=tenant-a's
|
|
448
|
+
row of the same content, silently breaking multi-tenant isolation
|
|
449
|
+
for any duplicated phrase.
|
|
450
|
+
"""
|
|
451
|
+
return hashlib.sha1(f"{arena}:{content}".encode()).hexdigest()[:32]
|
|
452
|
+
|
|
453
|
+
|
|
441
454
|
@app.post("/store")
|
|
442
455
|
async def store(req: StoreRequest):
|
|
443
456
|
"""Single-record ingest. Same wire format as pentatonic-memory v0.5."""
|
|
444
|
-
rid = (req.metadata or {}).get("id") or hashlib.sha1(req.content.encode()).hexdigest()[:32]
|
|
445
|
-
record = {"id": rid, "content": req.content, "metadata": req.metadata or {}}
|
|
446
457
|
arena = (req.metadata or {}).get("arena", "general")
|
|
458
|
+
rid = (req.metadata or {}).get("id") or _arena_scoped_rid(arena, req.content)
|
|
459
|
+
record = {"id": rid, "content": req.content, "metadata": req.metadata or {}}
|
|
447
460
|
|
|
448
461
|
# Stash the full metadata under every key shape any layer could echo back.
|
|
449
462
|
# L5/L6 use derivatives of rid; L2-internal returns paths shaped like
|
|
@@ -483,16 +496,16 @@ async def store_batch(req: StoreBatchRequest):
|
|
|
483
496
|
return {"inserted": 0, "ids": []}
|
|
484
497
|
|
|
485
498
|
# Normalise each record to {id, content, metadata}.
|
|
499
|
+
arena = req.arena or "general"
|
|
486
500
|
normalised = []
|
|
487
501
|
for r in req.records:
|
|
488
502
|
content = r.get("content") or r.get("text") or ""
|
|
489
503
|
if not content:
|
|
490
504
|
continue
|
|
491
|
-
rid = r.get("id") or
|
|
505
|
+
rid = r.get("id") or _arena_scoped_rid(arena, content)
|
|
492
506
|
normalised.append({"id": rid, "content": content, "metadata": r.get("metadata") or {}})
|
|
493
507
|
|
|
494
508
|
# Stash metadata for every record so /search can re-attach it.
|
|
495
|
-
arena = req.arena or "general"
|
|
496
509
|
for r in normalised:
|
|
497
510
|
_stash_all_keys(r["id"], r.get("metadata") or {}, arena)
|
|
498
511
|
|
|
@@ -970,6 +970,77 @@ def serve(port: int = DEFAULT_PORT):
|
|
|
970
970
|
milvus.load_collection(COLLECTION_NAME)
|
|
971
971
|
return {"status": "rebuilt"}
|
|
972
972
|
|
|
973
|
+
@api.post("/repair-fts")
|
|
974
|
+
def api_repair_fts():
|
|
975
|
+
"""Backfill the SQLite `chunks` content table from Milvus.
|
|
976
|
+
|
|
977
|
+
Pre-v0.7.6 the /index-batch path wrote straight to the FTS5
|
|
978
|
+
virtual table and never populated `chunks`. The result: BM25
|
|
979
|
+
search (which JOINs chunks ON rowid) returned zero hits even
|
|
980
|
+
though Milvus had the data, and /stats fts_chunks reported 0.
|
|
981
|
+
|
|
982
|
+
v0.7.6 fixed new writes; this endpoint cleans up old rows by
|
|
983
|
+
walking Milvus and INSERT-OR-REPLACE'ing every row into chunks.
|
|
984
|
+
The trigger on chunks then mirrors them into chunks_fts. Idempotent.
|
|
985
|
+
Runs in-process so the L6 service's already-open Milvus handle
|
|
986
|
+
is reused — no file-lock conflict (Milvus Lite locks the .db).
|
|
987
|
+
"""
|
|
988
|
+
milvus = get_milvus()
|
|
989
|
+
# Milvus Lite caps query() at 16384 per call; page through.
|
|
990
|
+
page_size = 16384
|
|
991
|
+
offset = 0
|
|
992
|
+
rows: list = []
|
|
993
|
+
while True:
|
|
994
|
+
page = milvus.query(
|
|
995
|
+
COLLECTION_NAME,
|
|
996
|
+
filter="id != ''",
|
|
997
|
+
output_fields=[
|
|
998
|
+
"id", "text", "source_file", "arena", "doc_type",
|
|
999
|
+
"heading", "chunk_index", "content_hash",
|
|
1000
|
+
"entities_json", "indexed_at",
|
|
1001
|
+
],
|
|
1002
|
+
limit=page_size,
|
|
1003
|
+
offset=offset,
|
|
1004
|
+
)
|
|
1005
|
+
if not page:
|
|
1006
|
+
break
|
|
1007
|
+
rows.extend(page)
|
|
1008
|
+
if len(page) < page_size:
|
|
1009
|
+
break
|
|
1010
|
+
offset += page_size
|
|
1011
|
+
fts_conn = get_fts_db()
|
|
1012
|
+
repaired = 0
|
|
1013
|
+
for r in rows:
|
|
1014
|
+
try:
|
|
1015
|
+
fts_conn.execute(
|
|
1016
|
+
"INSERT OR REPLACE INTO chunks "
|
|
1017
|
+
"(id, text, source_file, arena, doc_type, heading, "
|
|
1018
|
+
" chunk_index, content_hash, entities_json, indexed_at) "
|
|
1019
|
+
"VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?)",
|
|
1020
|
+
(
|
|
1021
|
+
r.get("id"),
|
|
1022
|
+
r.get("text", "") or "",
|
|
1023
|
+
r.get("source_file", "") or "",
|
|
1024
|
+
r.get("arena", "") or "general",
|
|
1025
|
+
r.get("doc_type", "") or "general",
|
|
1026
|
+
r.get("heading", "") or "",
|
|
1027
|
+
int(r.get("chunk_index", 0) or 0),
|
|
1028
|
+
r.get("content_hash", "") or "",
|
|
1029
|
+
r.get("entities_json", "") or "[]",
|
|
1030
|
+
r.get("indexed_at", "") or "",
|
|
1031
|
+
),
|
|
1032
|
+
)
|
|
1033
|
+
repaired += 1
|
|
1034
|
+
except Exception as exc:
|
|
1035
|
+
log.warning("repair-fts: skipping row %s: %s", r.get("id"), exc)
|
|
1036
|
+
fts_conn.commit()
|
|
1037
|
+
fts_conn.close()
|
|
1038
|
+
return {
|
|
1039
|
+
"status": "ok",
|
|
1040
|
+
"milvus_rows": len(rows),
|
|
1041
|
+
"repaired": repaired,
|
|
1042
|
+
}
|
|
1043
|
+
|
|
973
1044
|
log.info(f"L6 Document Store — http://127.0.0.1:{port}")
|
|
974
1045
|
uvicorn.run(api, host=os.environ.get("HOST","127.0.0.1"), port=port, log_level="info")
|
|
975
1046
|
|
|
@@ -125,6 +125,37 @@ print("yes" if ok and data else "no")')
|
|
|
125
125
|
[ "$all_match" = "yes" ] && ok "metadata_filter scopes to probe + arena" \
|
|
126
126
|
|| fail "metadata_filter let other rows through"
|
|
127
127
|
|
|
128
|
+
# ---------------------------------------------------------------------------
|
|
129
|
+
# Same content across two arenas — proves the arena-aware id derivation.
|
|
130
|
+
# Pre-v0.7.8, identical content collapsed to one row in L4/L5/L6 because
|
|
131
|
+
# the id was sha1(content); the second tenant's write overwrote the first.
|
|
132
|
+
# ---------------------------------------------------------------------------
|
|
133
|
+
|
|
134
|
+
echo ""
|
|
135
|
+
echo "=== same-content-across-arenas ==="
|
|
136
|
+
post '{"content":"shared phrase about Eclipse","metadata":{"arena":"e2e-tenant-x","probe":"e2e-arena"}}' >/dev/null
|
|
137
|
+
post '{"content":"shared phrase about Eclipse","metadata":{"arena":"e2e-tenant-y","probe":"e2e-arena"}}' >/dev/null
|
|
138
|
+
sleep 3
|
|
139
|
+
|
|
140
|
+
SX=$(curl -sf -X POST "$BASE/search" -H "Content-Type: application/json" \
|
|
141
|
+
-d '{"query":"Eclipse","limit":10,"arena":"e2e-tenant-x"}')
|
|
142
|
+
SY=$(curl -sf -X POST "$BASE/search" -H "Content-Type: application/json" \
|
|
143
|
+
-d '{"query":"Eclipse","limit":10,"arena":"e2e-tenant-y"}')
|
|
144
|
+
|
|
145
|
+
x_has_phrase=$(echo "$SX" | python3 -c '
|
|
146
|
+
import json,sys
|
|
147
|
+
data=json.load(sys.stdin).get("results",[])
|
|
148
|
+
print("yes" if any("Eclipse" in r.get("content","") for r in data) else "no")')
|
|
149
|
+
y_has_phrase=$(echo "$SY" | python3 -c '
|
|
150
|
+
import json,sys
|
|
151
|
+
data=json.load(sys.stdin).get("results",[])
|
|
152
|
+
print("yes" if any("Eclipse" in r.get("content","") for r in data) else "no")')
|
|
153
|
+
|
|
154
|
+
[ "$x_has_phrase" = "yes" ] && ok "tenant-x: shared phrase preserved" \
|
|
155
|
+
|| fail "tenant-x lost the shared phrase (id collision?)"
|
|
156
|
+
[ "$y_has_phrase" = "yes" ] && ok "tenant-y: shared phrase preserved" \
|
|
157
|
+
|| fail "tenant-y lost the shared phrase (id collision?)"
|
|
158
|
+
|
|
128
159
|
# ---------------------------------------------------------------------------
|
|
129
160
|
# /forget — by metadata_contains. Cleans up so reruns are idempotent.
|
|
130
161
|
# ---------------------------------------------------------------------------
|