risicare 0.2.1 → 0.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +131 -45
- package/dist/frameworks/instructor.cjs +45 -17
- package/dist/frameworks/instructor.cjs.map +1 -1
- package/dist/frameworks/instructor.js +47 -17
- package/dist/frameworks/instructor.js.map +1 -1
- package/dist/frameworks/langchain.cjs +73 -6
- package/dist/frameworks/langchain.cjs.map +1 -1
- package/dist/frameworks/langchain.d.cts +20 -4
- package/dist/frameworks/langchain.d.ts +20 -4
- package/dist/frameworks/langchain.js +75 -6
- package/dist/frameworks/langchain.js.map +1 -1
- package/dist/frameworks/langgraph.cjs +73 -6
- package/dist/frameworks/langgraph.cjs.map +1 -1
- package/dist/frameworks/langgraph.js +75 -6
- package/dist/frameworks/langgraph.js.map +1 -1
- package/dist/frameworks/llamaindex.cjs +41 -14
- package/dist/frameworks/llamaindex.cjs.map +1 -1
- package/dist/frameworks/llamaindex.js +43 -14
- package/dist/frameworks/llamaindex.js.map +1 -1
- package/dist/index.cjs +1494 -67
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +436 -1
- package/dist/index.d.ts +436 -1
- package/dist/index.js +1515 -67
- package/dist/index.js.map +1 -1
- package/dist/providers/anthropic/index.cjs +74 -24
- package/dist/providers/anthropic/index.cjs.map +1 -1
- package/dist/providers/anthropic/index.js +76 -24
- package/dist/providers/anthropic/index.js.map +1 -1
- package/dist/providers/bedrock/index.cjs +81 -24
- package/dist/providers/bedrock/index.cjs.map +1 -1
- package/dist/providers/bedrock/index.js +83 -24
- package/dist/providers/bedrock/index.js.map +1 -1
- package/dist/providers/cerebras/index.cjs +78 -25
- package/dist/providers/cerebras/index.cjs.map +1 -1
- package/dist/providers/cerebras/index.js +80 -25
- package/dist/providers/cerebras/index.js.map +1 -1
- package/dist/providers/cohere/index.cjs +95 -25
- package/dist/providers/cohere/index.cjs.map +1 -1
- package/dist/providers/cohere/index.js +97 -25
- package/dist/providers/cohere/index.js.map +1 -1
- package/dist/providers/google/index.cjs +77 -25
- package/dist/providers/google/index.cjs.map +1 -1
- package/dist/providers/google/index.js +79 -25
- package/dist/providers/google/index.js.map +1 -1
- package/dist/providers/groq/index.cjs +80 -25
- package/dist/providers/groq/index.cjs.map +1 -1
- package/dist/providers/groq/index.js +82 -25
- package/dist/providers/groq/index.js.map +1 -1
- package/dist/providers/huggingface/index.cjs +80 -25
- package/dist/providers/huggingface/index.cjs.map +1 -1
- package/dist/providers/huggingface/index.js +82 -25
- package/dist/providers/huggingface/index.js.map +1 -1
- package/dist/providers/mistral/index.cjs +72 -24
- package/dist/providers/mistral/index.cjs.map +1 -1
- package/dist/providers/mistral/index.js +74 -24
- package/dist/providers/mistral/index.js.map +1 -1
- package/dist/providers/ollama/index.cjs +83 -25
- package/dist/providers/ollama/index.cjs.map +1 -1
- package/dist/providers/ollama/index.js +85 -25
- package/dist/providers/ollama/index.js.map +1 -1
- package/dist/providers/openai/index.cjs +1429 -28
- package/dist/providers/openai/index.cjs.map +1 -1
- package/dist/providers/openai/index.js +1447 -28
- package/dist/providers/openai/index.js.map +1 -1
- package/dist/providers/together/index.cjs +80 -25
- package/dist/providers/together/index.cjs.map +1 -1
- package/dist/providers/together/index.js +82 -25
- package/dist/providers/together/index.js.map +1 -1
- package/dist/providers/vercel-ai/index.cjs +45 -17
- package/dist/providers/vercel-ai/index.cjs.map +1 -1
- package/dist/providers/vercel-ai/index.js +47 -17
- package/dist/providers/vercel-ai/index.js.map +1 -1
- package/package.json +17 -7
package/README.md
CHANGED
|
@@ -1,57 +1,147 @@
|
|
|
1
|
-
#
|
|
1
|
+
# risicare
|
|
2
2
|
|
|
3
|
-
|
|
3
|
+
AI agent observability and self-healing for Node.js and TypeScript.
|
|
4
4
|
|
|
5
5
|
[](https://www.npmjs.com/package/risicare)
|
|
6
|
-
[](https://www.npmjs.com/package/risicare)
|
|
7
|
+
[](https://www.npmjs.com/package/risicare)
|
|
8
|
+
[](https://opensource.org/licenses/MIT)
|
|
7
9
|
|
|
8
|
-
|
|
10
|
+
Monitor your AI agents in production. Trace every LLM call, detect errors automatically, and get AI-generated fixes — with 3 lines of setup.
|
|
11
|
+
|
|
12
|
+
## Quickstart
|
|
9
13
|
|
|
10
14
|
```bash
|
|
11
15
|
npm install risicare
|
|
12
16
|
```
|
|
13
17
|
|
|
14
18
|
```typescript
|
|
15
|
-
import { init } from 'risicare';
|
|
19
|
+
import { init, agent, shutdown } from 'risicare';
|
|
20
|
+
import { patchOpenAI } from 'risicare/openai';
|
|
21
|
+
import OpenAI from 'openai';
|
|
16
22
|
|
|
23
|
+
// 1. Initialize
|
|
17
24
|
init({
|
|
18
25
|
apiKey: 'rsk-...',
|
|
19
26
|
endpoint: 'https://app.risicare.ai',
|
|
20
27
|
});
|
|
21
28
|
|
|
22
|
-
//
|
|
29
|
+
// 2. Patch your LLM client
|
|
30
|
+
const openai = patchOpenAI(new OpenAI());
|
|
31
|
+
|
|
32
|
+
// 3. Wrap your agent — all LLM calls inside are traced automatically
|
|
33
|
+
const myAgent = agent({ name: 'research-agent' }, async (query: string) => {
|
|
34
|
+
const response = await openai.chat.completions.create({
|
|
35
|
+
model: 'gpt-4',
|
|
36
|
+
messages: [{ role: 'user', content: query }],
|
|
37
|
+
});
|
|
38
|
+
return response.choices[0].message.content;
|
|
39
|
+
});
|
|
40
|
+
|
|
41
|
+
// Run it — traces appear in your dashboard instantly
|
|
42
|
+
const result = await myAgent('What is quantum computing?');
|
|
43
|
+
await shutdown();
|
|
23
44
|
```
|
|
24
45
|
|
|
25
|
-
|
|
46
|
+
**That's it.** Your agent's LLM calls, latency, token usage, and costs now appear in the [Risicare dashboard](https://app.risicare.ai).
|
|
47
|
+
|
|
48
|
+
## Features
|
|
49
|
+
|
|
50
|
+
- **12 LLM providers** — OpenAI, Anthropic, Google, Mistral, Groq, Cohere, Together, Ollama, HuggingFace, Cerebras, Bedrock, Vercel AI
|
|
51
|
+
- **4 framework integrations** — LangChain, LangGraph, Instructor, LlamaIndex
|
|
52
|
+
- **Self-healing** — Automatic error diagnosis and AI-generated fix suggestions
|
|
53
|
+
- **Evaluation scores** — Rate agent quality with `score()` and 13 built-in scorers
|
|
54
|
+
- **Streaming support** — `tracedStream()` for async iterator tracing
|
|
55
|
+
- **Context propagation** — Automatic across `async/await`, `Promise`, `setTimeout`, `EventEmitter`
|
|
56
|
+
- **Zero runtime dependencies** — No bloat in your node_modules
|
|
57
|
+
- **Dual CJS/ESM** — Works with `require()` and `import`
|
|
58
|
+
- **Full TypeScript** — Strict types and IntelliSense out of the box
|
|
59
|
+
- **Non-blocking** — Async batch export with circuit breaker and retry
|
|
60
|
+
- **Zero overhead when disabled** — Frozen NOOP_SPAN singleton, no allocations
|
|
26
61
|
|
|
27
|
-
|
|
62
|
+
## LLM Providers
|
|
28
63
|
|
|
29
64
|
```typescript
|
|
30
65
|
import { patchOpenAI } from 'risicare/openai';
|
|
31
|
-
import
|
|
66
|
+
import { patchAnthropic } from 'risicare/anthropic';
|
|
67
|
+
import { patchGoogle } from 'risicare/google';
|
|
68
|
+
// ... and 9 more
|
|
32
69
|
|
|
33
70
|
const openai = patchOpenAI(new OpenAI());
|
|
34
|
-
//
|
|
71
|
+
// Every call is now traced — model, tokens, latency, cost
|
|
35
72
|
```
|
|
36
73
|
|
|
37
|
-
|
|
74
|
+
All 12 providers:
|
|
38
75
|
|
|
39
|
-
|
|
40
|
-
|----------|--------|
|
|
41
|
-
| OpenAI | `risicare/openai` |
|
|
42
|
-
| Anthropic | `risicare/anthropic` |
|
|
43
|
-
| Vercel AI SDK | `risicare/vercel-ai` |
|
|
76
|
+
`openai` · `anthropic` · `google` · `mistral` · `groq` · `cohere` · `together` · `ollama` · `huggingface` · `cerebras` · `bedrock` · `vercel-ai`
|
|
44
77
|
|
|
45
|
-
##
|
|
78
|
+
## Framework Integrations
|
|
79
|
+
|
|
80
|
+
```typescript
|
|
81
|
+
// LangChain
|
|
82
|
+
import { RisicareCallbackHandler } from 'risicare/langchain';
|
|
83
|
+
const handler = new RisicareCallbackHandler();
|
|
84
|
+
await chain.invoke(input, { callbacks: [handler] });
|
|
85
|
+
|
|
86
|
+
// LangGraph
|
|
87
|
+
import { instrumentLangGraph } from 'risicare/langgraph';
|
|
88
|
+
const tracedGraph = instrumentLangGraph(compiledGraph);
|
|
89
|
+
|
|
90
|
+
// Instructor
|
|
91
|
+
import { patchInstructor } from 'risicare/instructor';
|
|
92
|
+
const client = patchInstructor(instructor);
|
|
93
|
+
|
|
94
|
+
// LlamaIndex
|
|
95
|
+
import { RisicareLlamaIndexHandler } from 'risicare/llamaindex';
|
|
96
|
+
```
|
|
46
97
|
|
|
47
|
-
|
|
98
|
+
## Core API
|
|
48
99
|
|
|
49
100
|
```typescript
|
|
50
|
-
import {
|
|
101
|
+
import {
|
|
102
|
+
init, shutdown, // Lifecycle
|
|
103
|
+
agent, session, // Identity & grouping
|
|
104
|
+
traceThink, traceDecide, traceAct, // Decision phases
|
|
105
|
+
reportError, score, // Self-healing & evaluation
|
|
106
|
+
tracedStream, // Streaming
|
|
107
|
+
} from 'risicare';
|
|
108
|
+
|
|
109
|
+
init({ apiKey, endpoint }) // Initialize SDK
|
|
110
|
+
agent({ name }, fn) // Wrap function with agent identity
|
|
111
|
+
session({ sessionId, userId }, fn) // Group traces into user sessions
|
|
112
|
+
traceThink('analyze', async () => {...}) // Tag reasoning phase
|
|
113
|
+
traceDecide('choose', async () => {...}) // Tag decision phase
|
|
114
|
+
traceAct('execute', async () => {...}) // Tag action phase
|
|
115
|
+
reportError(error) // Report caught errors for diagnosis
|
|
116
|
+
score(traceId, 'quality', 0.92) // Record evaluation score [0.0-1.0]
|
|
117
|
+
tracedStream(asyncIterable, 'stream') // Trace async iterators
|
|
118
|
+
await shutdown() // Flush pending spans and close
|
|
119
|
+
```
|
|
120
|
+
|
|
121
|
+
## Self-Healing
|
|
122
|
+
|
|
123
|
+
When your agent fails, Risicare automatically:
|
|
124
|
+
|
|
125
|
+
1. **Classifies** the error (154 codes across TOOL, MEMORY, REASONING, OUTPUT, etc.)
|
|
126
|
+
2. **Diagnoses** the root cause using AI analysis
|
|
127
|
+
3. **Generates** a fix you can review and apply
|
|
128
|
+
|
|
129
|
+
```typescript
|
|
130
|
+
try {
|
|
131
|
+
await myAgent(input);
|
|
132
|
+
} catch (error) {
|
|
133
|
+
reportError(error); // Triggers automatic diagnosis pipeline
|
|
134
|
+
}
|
|
135
|
+
```
|
|
136
|
+
|
|
137
|
+
## Decision Phases
|
|
138
|
+
|
|
139
|
+
Structure your traces to see how your agent thinks, decides, and acts:
|
|
51
140
|
|
|
141
|
+
```typescript
|
|
52
142
|
const myAgent = agent({ name: 'planner', role: 'coordinator' }, async (input) => {
|
|
53
143
|
const analysis = await traceThink('analyze', async () => {
|
|
54
|
-
return await openai.chat.completions.create({ ... });
|
|
144
|
+
return await openai.chat.completions.create({ /* ... */ });
|
|
55
145
|
});
|
|
56
146
|
|
|
57
147
|
const decision = await traceDecide('choose-action', async () => {
|
|
@@ -62,38 +152,34 @@ const myAgent = agent({ name: 'planner', role: 'coordinator' }, async (input) =>
|
|
|
62
152
|
return executeAction(decision);
|
|
63
153
|
});
|
|
64
154
|
});
|
|
65
|
-
|
|
66
|
-
// Wrap in a session for user-level tracking
|
|
67
|
-
const result = await session({ userId: 'user-123' }, () => myAgent(input));
|
|
68
155
|
```
|
|
69
156
|
|
|
70
|
-
##
|
|
71
|
-
|
|
72
|
-
- **Zero runtime dependencies** — no bloat in your node_modules
|
|
73
|
-
- **Dual CJS/ESM** — works with CommonJS `require()` and ES module `import`
|
|
74
|
-
- **Full TypeScript** — strict types, generics, and IntelliSense out of the box
|
|
75
|
-
- **Non-blocking** — async batch export with circuit breaker and retry
|
|
76
|
-
- **Context propagation** — automatic across `async/await`, `Promise`, `setTimeout`, and `EventEmitter`
|
|
77
|
-
- **Zero overhead when disabled** — frozen NOOP_SPAN singleton, no allocations
|
|
157
|
+
## Sessions
|
|
78
158
|
|
|
79
|
-
|
|
159
|
+
Group traces from the same user conversation:
|
|
80
160
|
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
| **Tier 4** | `traceThink / traceDecide / traceAct` | Decision phase visibility |
|
|
88
|
-
| **Tier 5** | `traceMessage / traceDelegate` | Multi-agent communication |
|
|
161
|
+
```typescript
|
|
162
|
+
const result = await session(
|
|
163
|
+
{ sessionId: 'sess-abc123', userId: 'user-456' },
|
|
164
|
+
() => myAgent(userMessage)
|
|
165
|
+
);
|
|
166
|
+
```
|
|
89
167
|
|
|
90
168
|
## Requirements
|
|
91
169
|
|
|
92
|
-
- Node.js
|
|
93
|
-
- TypeScript
|
|
170
|
+
- Node.js 18+
|
|
171
|
+
- TypeScript 5.0+ (optional, types included)
|
|
94
172
|
|
|
95
|
-
##
|
|
173
|
+
## Documentation
|
|
96
174
|
|
|
97
|
-
- [
|
|
98
|
-
- [
|
|
175
|
+
- [Full docs](https://risicare.ai/docs)
|
|
176
|
+
- [Quickstart guide](https://risicare.ai/docs/start/quickstart)
|
|
177
|
+
- [JS SDK reference](https://risicare.ai/docs/reference/js-sdk)
|
|
178
|
+
- [LLM providers](https://risicare.ai/docs/instrument/providers)
|
|
179
|
+
- [Self-healing](https://risicare.ai/docs/heal)
|
|
180
|
+
- [Evaluations](https://risicare.ai/docs/observe/evaluations)
|
|
99
181
|
- [Python SDK](https://pypi.org/project/risicare/)
|
|
182
|
+
|
|
183
|
+
## License
|
|
184
|
+
|
|
185
|
+
[MIT](LICENSE)
|
|
@@ -3,6 +3,9 @@ var __defProp = Object.defineProperty;
|
|
|
3
3
|
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
4
4
|
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
5
5
|
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
6
|
+
var __esm = (fn, res) => function __init() {
|
|
7
|
+
return fn && (res = (0, fn[__getOwnPropNames(fn)[0]])(fn = 0)), res;
|
|
8
|
+
};
|
|
6
9
|
var __export = (target, all) => {
|
|
7
10
|
for (var name in all)
|
|
8
11
|
__defProp(target, name, { get: all[name], enumerable: true });
|
|
@@ -17,6 +20,37 @@ var __copyProps = (to, from, except, desc) => {
|
|
|
17
20
|
};
|
|
18
21
|
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
19
22
|
|
|
23
|
+
// src/globals.ts
|
|
24
|
+
function getTracer() {
|
|
25
|
+
return G[PREFIX + "tracer"];
|
|
26
|
+
}
|
|
27
|
+
function getDebug() {
|
|
28
|
+
return G[PREFIX + "debug"] ?? false;
|
|
29
|
+
}
|
|
30
|
+
var import_node_async_hooks, G, PREFIX;
|
|
31
|
+
var init_globals = __esm({
|
|
32
|
+
"src/globals.ts"() {
|
|
33
|
+
"use strict";
|
|
34
|
+
import_node_async_hooks = require("async_hooks");
|
|
35
|
+
G = globalThis;
|
|
36
|
+
PREFIX = "__risicare_";
|
|
37
|
+
}
|
|
38
|
+
});
|
|
39
|
+
|
|
40
|
+
// src/utils/log.ts
|
|
41
|
+
function debug(msg) {
|
|
42
|
+
if (getDebug()) {
|
|
43
|
+
process.stderr.write(`[risicare] ${msg}
|
|
44
|
+
`);
|
|
45
|
+
}
|
|
46
|
+
}
|
|
47
|
+
var init_log = __esm({
|
|
48
|
+
"src/utils/log.ts"() {
|
|
49
|
+
"use strict";
|
|
50
|
+
init_globals();
|
|
51
|
+
}
|
|
52
|
+
});
|
|
53
|
+
|
|
20
54
|
// src/frameworks/instructor.ts
|
|
21
55
|
var instructor_exports = {};
|
|
22
56
|
__export(instructor_exports, {
|
|
@@ -97,31 +131,25 @@ var NOOP_SPAN = Object.freeze({
|
|
|
97
131
|
}
|
|
98
132
|
});
|
|
99
133
|
|
|
100
|
-
// src/
|
|
101
|
-
|
|
102
|
-
var G = globalThis;
|
|
103
|
-
var PREFIX = "__risicare_";
|
|
104
|
-
function getTracer() {
|
|
105
|
-
return G[PREFIX + "tracer"];
|
|
106
|
-
}
|
|
107
|
-
function getDebug() {
|
|
108
|
-
return G[PREFIX + "debug"] ?? false;
|
|
109
|
-
}
|
|
134
|
+
// src/context/storage.ts
|
|
135
|
+
init_globals();
|
|
110
136
|
|
|
111
|
-
// src/
|
|
112
|
-
|
|
113
|
-
|
|
114
|
-
|
|
115
|
-
|
|
116
|
-
}
|
|
117
|
-
}
|
|
137
|
+
// src/exporters/batch.ts
|
|
138
|
+
init_log();
|
|
139
|
+
|
|
140
|
+
// src/exporters/http.ts
|
|
141
|
+
init_log();
|
|
118
142
|
|
|
119
143
|
// src/client.ts
|
|
144
|
+
init_log();
|
|
145
|
+
init_globals();
|
|
120
146
|
function getTracer2() {
|
|
121
147
|
return getTracer();
|
|
122
148
|
}
|
|
149
|
+
var _ERROR_DEDUP_TTL_MS = 5 * 60 * 1e3;
|
|
123
150
|
|
|
124
151
|
// src/frameworks/instructor.ts
|
|
152
|
+
init_log();
|
|
125
153
|
function patchInstructor(client) {
|
|
126
154
|
return new Proxy(client, {
|
|
127
155
|
get(target, prop, receiver) {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/frameworks/instructor.ts","../../src/ids.ts","../../src/noop.ts","../../src/globals.ts","../../src/utils/log.ts","../../src/client.ts"],"sourcesContent":["/**\n * Instructor integration via Proxy wrapping.\n *\n * Instructor wraps an LLM client's create() method to extract structured\n * data from completions using a response model (Zod schema, etc.).\n * This integration intercepts the instructor client's create() to add a span\n * around the structured extraction call.\n *\n * Usage:\n * import { patchInstructor } from 'risicare/frameworks/instructor';\n * import Instructor from '@instructor-ai/instructor';\n * const client = patchInstructor(Instructor.from(openai));\n *\n * Does NOT suppress provider instrumentation -- Instructor delegates to the\n * underlying LLM client, which should produce its own provider span.\n */\n\nimport { getTracer } from '../client.js';\nimport { SpanKind } from '../types.js';\nimport { debug } from '../utils/log.js';\n\n/**\n * Wrap an Instructor client with Risicare tracing.\n *\n * Returns a Proxy that intercepts `create` calls to create a span\n * capturing the structured extraction metadata (response model, retries).\n *\n * @param client - An Instructor client with a create() method\n * @returns A proxied client with tracing enabled\n */\nexport function patchInstructor<T extends object>(client: T): T {\n return new Proxy(client, {\n get(target, prop, receiver) {\n const value = Reflect.get(target, prop, receiver);\n\n if (prop === 'create' && typeof value === 'function') {\n return function patchedCreate(\n this: unknown,\n ...args: unknown[]\n ): unknown {\n const tracer = getTracer();\n if (!tracer?.enabled) {\n debug(\n 'Tracer not initialized — call init() before using patchInstructor()',\n );\n return value.apply(this, args);\n }\n\n const params = (args[0] ?? {}) as Record<string, unknown>;\n const responseModel =\n params.response_model ?? params.responseModel;\n\n // Extract a readable name from the response model\n let modelName = 'unknown';\n if (responseModel && typeof responseModel === 'function') {\n modelName =\n (responseModel as { name?: string }).name || 'unknown';\n } else if (responseModel && typeof responseModel === 'object') {\n modelName =\n ((responseModel as Record<string, unknown>).name as string) ??\n ((responseModel as Record<string, unknown>).description as string) ??\n 'schema';\n } else if (typeof responseModel === 'string') {\n modelName = responseModel;\n }\n\n return tracer.startSpan(\n {\n name: 'instructor.create',\n kind: SpanKind.LLM_CALL,\n attributes: {\n framework: 'instructor',\n 'framework.span_kind': 'extraction',\n 'instructor.response_model': modelName,\n ...(params.max_retries != null\n ? { 'instructor.max_retries': params.max_retries }\n : {}),\n ...(params.model != null\n ? { 'gen_ai.request.model': params.model }\n : {}),\n },\n },\n (_span) => {\n const result = value.apply(this, args);\n\n // Handle async results (Promises)\n if (\n result &&\n typeof (result as Promise<unknown>).then === 'function'\n ) {\n return (result as Promise<unknown>).then((r) => r);\n }\n return result;\n },\n );\n };\n }\n\n return value;\n },\n });\n}\n","/**\n * ID generation for traces and spans.\n *\n * Trace IDs: 32 lowercase hex characters (16 random bytes)\n * Span IDs: 16 lowercase hex characters (8 random bytes)\n *\n * Uses crypto.randomBytes for cryptographically secure randomness.\n */\n\nimport { randomBytes } from 'node:crypto';\n\nconst HEX_REGEX_32 = /^[0-9a-f]{32}$/;\nconst HEX_REGEX_16 = /^[0-9a-f]{16}$/;\n\nexport function generateTraceId(): string {\n return randomBytes(16).toString('hex');\n}\n\nexport function generateSpanId(): string {\n return randomBytes(8).toString('hex');\n}\n\nexport function generateAgentId(prefix?: string): string {\n const suffix = randomBytes(8).toString('hex');\n return prefix ? `${prefix}-${suffix}` : suffix;\n}\n\nexport function validateTraceId(id: string): boolean {\n return HEX_REGEX_32.test(id);\n}\n\nexport function validateSpanId(id: string): boolean {\n return HEX_REGEX_16.test(id);\n}\n","/**\n * No-op implementations for the disabled path.\n *\n * When tracing is disabled, all operations return these no-op objects\n * to maintain zero overhead. No allocations, no side effects.\n */\n\nimport { SpanKind, SpanStatus, type SpanPayload } from './types.js';\n\n/**\n * A frozen no-op span that silently ignores all operations.\n * Used when SDK is disabled to avoid overhead.\n */\nexport const NOOP_SPAN = Object.freeze({\n traceId: '00000000000000000000000000000000',\n spanId: '0000000000000000',\n parentSpanId: undefined,\n name: 'noop',\n kind: SpanKind.INTERNAL,\n startTime: '',\n startHrtime: 0,\n endTime: undefined,\n status: SpanStatus.UNSET,\n statusMessage: undefined,\n attributes: Object.freeze({}) as Record<string, unknown>,\n events: Object.freeze([]) as readonly [],\n links: Object.freeze([]) as readonly [],\n sessionId: undefined,\n agentId: undefined,\n agentName: undefined,\n semanticPhase: undefined,\n llmProvider: undefined,\n llmModel: undefined,\n llmPromptTokens: undefined,\n llmCompletionTokens: undefined,\n llmTotalTokens: undefined,\n llmCostUsd: undefined,\n toolName: undefined,\n toolSuccess: undefined,\n isEnded: true,\n durationMs: 0,\n\n setAttribute() { return this; },\n setAttributes() { return this; },\n setStatus() { return this; },\n addEvent() { return this; },\n addLink() { return this; },\n recordException() { return this; },\n setLlmFields() { return this; },\n setToolFields() { return this; },\n end() {},\n toPayload(): SpanPayload {\n return {\n traceId: this.traceId,\n spanId: this.spanId,\n name: this.name,\n kind: this.kind,\n startTime: this.startTime,\n status: this.status,\n attributes: {},\n events: [],\n links: [],\n };\n },\n});\n\nexport type NoopSpan = typeof NOOP_SPAN;\n","/**\n * Shared state via globalThis — ensures all entry point bundles share\n * the same singleton instances.\n *\n * Problem: tsup with `splitting: false` gives each entry point (index,\n * openai, anthropic, vercel-ai) its own copy of module-level variables.\n * This means `init()` from 'risicare' sets a tracer that 'risicare/openai'\n * can't see — breaking all provider instrumentation silently.\n *\n * Solution: Store all mutable singletons on globalThis with a namespaced\n * prefix. Every bundle reads/writes the same global slots.\n *\n * This pattern is used by React, OpenTelemetry, and other SDKs that must\n * share state across independently bundled entry points.\n *\n * @internal\n */\n\nimport { AsyncLocalStorage } from 'node:async_hooks';\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nconst G = globalThis as any;\nconst PREFIX = '__risicare_';\n\n// ─── Client & Tracer ────────────────────────────────────────────────────────\n\nexport function getClient(): unknown {\n return G[PREFIX + 'client'];\n}\n\nexport function setClient(client: unknown): void {\n G[PREFIX + 'client'] = client;\n}\n\nexport function getTracer(): unknown {\n return G[PREFIX + 'tracer'];\n}\n\nexport function setTracer(tracer: unknown): void {\n G[PREFIX + 'tracer'] = tracer;\n}\n\n// ─── Context Storage ────────────────────────────────────────────────────────\n\nexport function getContextStorage(): AsyncLocalStorage<unknown> {\n if (!G[PREFIX + 'ctx']) {\n G[PREFIX + 'ctx'] = new AsyncLocalStorage();\n }\n return G[PREFIX + 'ctx'];\n}\n\n// ─── Span Registry ──────────────────────────────────────────────────────────\n\nexport function getRegistry(): Map<string, unknown> {\n if (!G[PREFIX + 'registry']) {\n G[PREFIX + 'registry'] = new Map();\n }\n return G[PREFIX + 'registry'];\n}\n\nexport function getOpCount(): number {\n return G[PREFIX + 'opcount'] ?? 0;\n}\n\nexport function setOpCount(n: number): void {\n G[PREFIX + 'opcount'] = n;\n}\n\n// ─── Debug Flag ─────────────────────────────────────────────────────────────\n\nexport function getDebug(): boolean {\n return G[PREFIX + 'debug'] ?? false;\n}\n\nexport function setDebugFlag(enabled: boolean): void {\n G[PREFIX + 'debug'] = enabled;\n}\n","/**\n * Internal logger for the Risicare SDK.\n *\n * Centralizes all diagnostic output so that:\n * - Debug messages are gated by a single flag (zero-cost when disabled)\n * - Warnings always fire (operational alerts like queue full)\n * - All output goes to stderr with a consistent [risicare] prefix\n * - A future custom logger callback can be added in one place\n */\n\nimport { getDebug, setDebugFlag } from '../globals.js';\n\n/**\n * Enable or disable debug logging. Called once during init().\n * @internal\n */\nexport function setDebug(enabled: boolean): void {\n setDebugFlag(enabled);\n}\n\n/**\n * Log a debug message. Only outputs when debug mode is enabled.\n * @internal\n */\nexport function debug(msg: string): void {\n if (getDebug()) {\n process.stderr.write(`[risicare] ${msg}\\n`);\n }\n}\n\n/**\n * Log a warning. Always outputs regardless of debug mode.\n * Use sparingly — only for operational issues the user should see.\n * @internal\n */\nexport function warn(msg: string): void {\n process.stderr.write(`[risicare] WARNING: ${msg}\\n`);\n}\n","/**\n * RisicareClient — singleton client managing SDK lifecycle.\n *\n * Handles initialization, shutdown, and the connection between\n * the Tracer and the export pipeline (batch processor + HTTP exporter).\n *\n * Usage:\n * import { init, shutdown } from 'risicare';\n * init({ apiKey: 'rsk-...' }); // API key determines project\n * // ... instrument code ...\n * await shutdown(); // flush remaining spans\n */\n\nimport { type RisicareConfig, resolveConfig } from './config.js';\nimport { Tracer } from './tracer.js';\nimport { BatchSpanProcessor } from './exporters/batch.js';\nimport { HttpExporter } from './exporters/http.js';\nimport { ConsoleExporter } from './exporters/console.js';\nimport { SpanKind, SpanStatus } from './types.js';\nimport type { SpanExporter } from './exporters/base.js';\nimport { setDebug, debug } from './utils/log.js';\nimport {\n getClient as getGlobalClient,\n setClient as setGlobalClient,\n getTracer as getGlobalTracer,\n setTracer as setGlobalTracer,\n} from './globals.js';\n\n// ─── Client Class ───────────────────────────────────────────────────────────\n\nclass RisicareClient {\n readonly config: ReturnType<typeof resolveConfig>;\n readonly processor: BatchSpanProcessor;\n readonly tracer: Tracer;\n private _shutdownPromise: Promise<void> | undefined;\n private _shutdownHandlers: { signal: string; handler: () => void }[] = [];\n\n constructor(config?: Partial<RisicareConfig>) {\n this.config = resolveConfig(config);\n\n // API key format validation\n if (this.config.apiKey && !this.config.apiKey.startsWith('rsk-')) {\n debug('Warning: API key should start with \"rsk-\". Got: ' + this.config.apiKey.slice(0, 4) + '...');\n }\n\n // Build exporter chain\n let exporter: SpanExporter;\n if (this.config.debug && !this.config.apiKey) {\n exporter = new ConsoleExporter();\n } else if (this.config.apiKey) {\n exporter = new HttpExporter({\n endpoint: this.config.endpoint,\n apiKey: this.config.apiKey,\n projectId: this.config.projectId || undefined,\n environment: this.config.environment || undefined,\n compress: this.config.compress,\n });\n } else {\n // No API key and not debug — use console as fallback\n exporter = new ConsoleExporter();\n }\n\n this.processor = new BatchSpanProcessor({\n exporters: [exporter],\n batchSize: this.config.batchSize,\n batchTimeoutMs: this.config.batchTimeoutMs,\n maxQueueSize: this.config.maxQueueSize,\n debug: this.config.debug,\n });\n\n this.tracer = new Tracer({\n onSpanEnd: (span) => this.processor.onSpanEnd(span),\n sampleRate: this.config.sampleRate,\n enabled: this.config.enabled,\n traceContent: this.config.traceContent,\n });\n\n // Start the batch processor (enables span queuing and periodic flushing)\n this.processor.start();\n\n // Register shutdown hooks\n this._registerShutdownHooks();\n\n // Enable internal debug logging if configured\n setDebug(this.config.debug);\n debug(`Initialized: enabled=${this.config.enabled}, endpoint=${this.config.endpoint}`);\n }\n\n get enabled(): boolean {\n return this.tracer.enabled;\n }\n\n set enabled(value: boolean) {\n this.tracer.enabled = value;\n }\n\n // Audit #6: Promise-based shutdown dedup (fixes TOCTOU race condition)\n async shutdown(): Promise<void> {\n if (this._shutdownPromise) return this._shutdownPromise;\n this._shutdownPromise = this._doShutdown();\n return this._shutdownPromise;\n }\n\n private async _doShutdown(): Promise<void> {\n debug('Shutting down...');\n\n // Audit #3: Remove process listeners to prevent leak\n for (const { signal, handler } of this._shutdownHandlers) {\n process.removeListener(signal, handler);\n }\n this._shutdownHandlers = [];\n\n await this.processor.shutdown();\n }\n\n async flush(): Promise<void> {\n await this.processor.flush();\n }\n\n private _registerShutdownHooks(): void {\n const onShutdown = () => {\n // Audit #3: Add 5s timeout to prevent hanging on signal\n const timeout = setTimeout(() => process.exit(1), 5000);\n timeout.unref();\n this.shutdown().catch(() => {}).finally(() => clearTimeout(timeout));\n };\n\n const signals = ['beforeExit', 'SIGTERM', 'SIGINT'];\n for (const signal of signals) {\n process.once(signal, onShutdown);\n this._shutdownHandlers.push({ signal, handler: onShutdown });\n }\n }\n}\n\n// ─── Public API ─────────────────────────────────────────────────────────────\n\n/**\n * Initialize the Risicare SDK. Call once at application startup.\n *\n * @example\n * import { init } from 'risicare';\n * init({ apiKey: 'rsk-...', serviceName: 'my-agent', environment: 'production' });\n */\nexport function init(config?: Partial<RisicareConfig>): void {\n if (getGlobalClient()) {\n debug('Already initialized. Call shutdown() first to re-initialize.');\n return;\n }\n\n const client = new RisicareClient(config);\n setGlobalClient(client);\n setGlobalTracer(client.tracer);\n}\n\n/**\n * Gracefully shut down the SDK. Flushes pending spans before resolving.\n */\nexport async function shutdown(): Promise<void> {\n const client = getGlobalClient() as RisicareClient | undefined;\n if (!client) return;\n await client.shutdown();\n setGlobalClient(undefined);\n setGlobalTracer(undefined);\n}\n\n/**\n * Flush all pending spans without shutting down.\n */\nexport async function flush(): Promise<void> {\n const client = getGlobalClient() as RisicareClient | undefined;\n if (!client) return;\n await client.flush();\n}\n\n/**\n * Enable tracing at runtime.\n */\nexport function enable(): void {\n const client = getGlobalClient() as RisicareClient | undefined;\n if (client) client.enabled = true;\n}\n\n/**\n * Disable tracing at runtime. Spans will not be created or exported.\n */\nexport function disable(): void {\n const client = getGlobalClient() as RisicareClient | undefined;\n if (client) client.enabled = false;\n}\n\n/**\n * Check whether tracing is currently enabled.\n */\nexport function isEnabled(): boolean {\n const client = getGlobalClient() as RisicareClient | undefined;\n return client?.enabled ?? false;\n}\n\n/**\n * Get the global tracer instance. Returns undefined if not initialized.\n */\nexport function getTracer(): Tracer | undefined {\n return getGlobalTracer() as Tracer | undefined;\n}\n\n/**\n * Get the global tracer, or throw if not initialized.\n * @internal Used by decorators and providers that require an active tracer.\n */\nexport function requireTracer(): Tracer {\n const tracer = getGlobalTracer() as Tracer | undefined;\n if (!tracer) {\n throw new Error(\n 'Risicare SDK not initialized. Call init() before using tracing features.',\n );\n }\n return tracer;\n}\n\n/**\n * Check whether content tracing (prompt/completion capture) is enabled.\n */\nexport function getTraceContent(): boolean {\n const tracer = getGlobalTracer() as Tracer | undefined;\n return tracer?.traceContent ?? true;\n}\n\n/**\n * Get SDK metrics: exported spans, dropped spans, failed exports, queue stats.\n * Returns zero-valued metrics if SDK is not initialized.\n */\nexport function getMetrics() {\n const client = getGlobalClient() as RisicareClient | undefined;\n return client?.processor.getMetrics() ?? {\n exportedSpans: 0,\n droppedSpans: 0,\n failedExports: 0,\n queueSize: 0,\n queueCapacity: 0,\n queueUtilization: 0,\n };\n}\n\n// ─── reportError ──────────────────────────────────────────────────────────\n\n/**\n * Report a caught exception to the self-healing pipeline.\n *\n * Creates an error span that triggers diagnosis and fix generation.\n * This function never throws and is non-blocking.\n *\n * @param error - The caught exception (Error object or string)\n * @param options - Optional attributes and context overrides\n */\nexport function reportError(\n error: unknown,\n options?: { name?: string; attributes?: Record<string, unknown> },\n): void {\n try {\n const tracer = getTracer();\n if (!tracer) return;\n\n const err = error instanceof Error ? error : new Error(String(error));\n const spanName = options?.name ?? `error:${err.constructor.name}`;\n\n tracer.startSpan({ name: spanName, kind: SpanKind.INTERNAL }, (span) => {\n span.setStatus(SpanStatus.ERROR, err.message);\n span.setAttribute('error', true);\n span.setAttribute('error.type', err.constructor.name);\n span.setAttribute('error.message', err.message.slice(0, 2000));\n if (err.stack) span.setAttribute('error.stack', err.stack.slice(0, 4000));\n span.setAttribute('risicare.reported_error', true);\n if (options?.attributes) {\n for (const [k, v] of Object.entries(options.attributes)) {\n span.setAttribute(k, v);\n }\n }\n });\n } catch {\n // Never crash the host application\n debug('reportError failed');\n }\n}\n\n// ─── score ─────────────────────────────────────────────────────────────────\n\n/**\n * Record a custom evaluation score on a trace.\n *\n * Sends the score to the server in a fire-and-forget fashion.\n * This function never throws and is non-blocking.\n *\n * @param traceId - The trace to score\n * @param name - Score name (e.g., \"accuracy\", \"user_satisfaction\")\n * @param value - Score value between 0.0 and 1.0 inclusive\n * @param options - Optional span_id and comment\n */\nexport function score(\n traceId: string,\n name: string,\n value: number,\n options?: { spanId?: string; comment?: string },\n): void {\n try {\n if (typeof value !== 'number' || value < 0.0 || value > 1.0) {\n debug(`score: value must be in [0.0, 1.0], got ${value}. Score not sent.`);\n return;\n }\n if (!traceId || !name) {\n debug('score: traceId and name are required');\n return;\n }\n\n const client = getGlobalClient() as RisicareClient | undefined;\n if (!client?.enabled || !client.config.apiKey) return;\n\n const endpoint = client.config.endpoint.replace(/\\/$/, '');\n const url = `${endpoint}/api/v1/scores`;\n const body = JSON.stringify({\n trace_id: traceId,\n name,\n score: value,\n source: 'sdk',\n ...(options?.spanId && { span_id: options.spanId }),\n ...(options?.comment && { comment: options.comment }),\n });\n\n // Fire-and-forget — never blocks caller\n fetch(url, {\n method: 'POST',\n headers: {\n 'Content-Type': 'application/json',\n 'Authorization': `Bearer ${client.config.apiKey}`,\n },\n body,\n }).catch((err) => debug(`score: send failed: ${err}`));\n } catch {\n // Never crash the host application\n debug('score failed');\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACSA,yBAA4B;;;ACIrB,IAAM,YAAY,OAAO,OAAO;AAAA,EACrC,SAAS;AAAA,EACT,QAAQ;AAAA,EACR,cAAc;AAAA,EACd,MAAM;AAAA,EACN;AAAA,EACA,WAAW;AAAA,EACX,aAAa;AAAA,EACb,SAAS;AAAA,EACT;AAAA,EACA,eAAe;AAAA,EACf,YAAY,OAAO,OAAO,CAAC,CAAC;AAAA,EAC5B,QAAQ,OAAO,OAAO,CAAC,CAAC;AAAA,EACxB,OAAO,OAAO,OAAO,CAAC,CAAC;AAAA,EACvB,WAAW;AAAA,EACX,SAAS;AAAA,EACT,WAAW;AAAA,EACX,eAAe;AAAA,EACf,aAAa;AAAA,EACb,UAAU;AAAA,EACV,iBAAiB;AAAA,EACjB,qBAAqB;AAAA,EACrB,gBAAgB;AAAA,EAChB,YAAY;AAAA,EACZ,UAAU;AAAA,EACV,aAAa;AAAA,EACb,SAAS;AAAA,EACT,YAAY;AAAA,EAEZ,eAAe;AAAE,WAAO;AAAA,EAAM;AAAA,EAC9B,gBAAgB;AAAE,WAAO;AAAA,EAAM;AAAA,EAC/B,YAAY;AAAE,WAAO;AAAA,EAAM;AAAA,EAC3B,WAAW;AAAE,WAAO;AAAA,EAAM;AAAA,EAC1B,UAAU;AAAE,WAAO;AAAA,EAAM;AAAA,EACzB,kBAAkB;AAAE,WAAO;AAAA,EAAM;AAAA,EACjC,eAAe;AAAE,WAAO;AAAA,EAAM;AAAA,EAC9B,gBAAgB;AAAE,WAAO;AAAA,EAAM;AAAA,EAC/B,MAAM;AAAA,EAAC;AAAA,EACP,YAAyB;AACvB,WAAO;AAAA,MACL,SAAS,KAAK;AAAA,MACd,QAAQ,KAAK;AAAA,MACb,MAAM,KAAK;AAAA,MACX,MAAM,KAAK;AAAA,MACX,WAAW,KAAK;AAAA,MAChB,QAAQ,KAAK;AAAA,MACb,YAAY,CAAC;AAAA,MACb,QAAQ,CAAC;AAAA,MACT,OAAO,CAAC;AAAA,IACV;AAAA,EACF;AACF,CAAC;;;AC9CD,8BAAkC;AAGlC,IAAM,IAAI;AACV,IAAM,SAAS;AAYR,SAAS,YAAqB;AACnC,SAAO,EAAE,SAAS,QAAQ;AAC5B;AAkCO,SAAS,WAAoB;AAClC,SAAO,EAAE,SAAS,OAAO,KAAK;AAChC;;;AChDO,SAAS,MAAM,KAAmB;AACvC,MAAI,SAAS,GAAG;AACd,YAAQ,OAAO,MAAM,cAAc,GAAG;AAAA,CAAI;AAAA,EAC5C;AACF;;;AC8KO,SAASA,aAAgC;AAC9C,SAAO,UAAgB;AACzB;;;AL9KO,SAAS,gBAAkC,QAAc;AAC9D,SAAO,IAAI,MAAM,QAAQ;AAAA,IACvB,IAAI,QAAQ,MAAM,UAAU;AAC1B,YAAM,QAAQ,QAAQ,IAAI,QAAQ,MAAM,QAAQ;AAEhD,UAAI,SAAS,YAAY,OAAO,UAAU,YAAY;AACpD,eAAO,SAAS,iBAEX,MACM;AACT,gBAAM,SAASC,WAAU;AACzB,cAAI,CAAC,QAAQ,SAAS;AACpB;AAAA,cACE;AAAA,YACF;AACA,mBAAO,MAAM,MAAM,MAAM,IAAI;AAAA,UAC/B;AAEA,gBAAM,SAAU,KAAK,CAAC,KAAK,CAAC;AAC5B,gBAAM,gBACJ,OAAO,kBAAkB,OAAO;AAGlC,cAAI,YAAY;AAChB,cAAI,iBAAiB,OAAO,kBAAkB,YAAY;AACxD,wBACG,cAAoC,QAAQ;AAAA,UACjD,WAAW,iBAAiB,OAAO,kBAAkB,UAAU;AAC7D,wBACI,cAA0C,QAC1C,cAA0C,eAC5C;AAAA,UACJ,WAAW,OAAO,kBAAkB,UAAU;AAC5C,wBAAY;AAAA,UACd;AAEA,iBAAO,OAAO;AAAA,YACZ;AAAA,cACE,MAAM;AAAA,cACN;AAAA,cACA,YAAY;AAAA,gBACV,WAAW;AAAA,gBACX,uBAAuB;AAAA,gBACvB,6BAA6B;AAAA,gBAC7B,GAAI,OAAO,eAAe,OACtB,EAAE,0BAA0B,OAAO,YAAY,IAC/C,CAAC;AAAA,gBACL,GAAI,OAAO,SAAS,OAChB,EAAE,wBAAwB,OAAO,MAAM,IACvC,CAAC;AAAA,cACP;AAAA,YACF;AAAA,YACA,CAAC,UAAU;AACT,oBAAM,SAAS,MAAM,MAAM,MAAM,IAAI;AAGrC,kBACE,UACA,OAAQ,OAA4B,SAAS,YAC7C;AACA,uBAAQ,OAA4B,KAAK,CAAC,MAAM,CAAC;AAAA,cACnD;AACA,qBAAO;AAAA,YACT;AAAA,UACF;AAAA,QACF;AAAA,MACF;AAEA,aAAO;AAAA,IACT;AAAA,EACF,CAAC;AACH;","names":["getTracer","getTracer"]}
|
|
1
|
+
{"version":3,"sources":["../../src/globals.ts","../../src/utils/log.ts","../../src/frameworks/instructor.ts","../../src/ids.ts","../../src/noop.ts","../../src/context/storage.ts","../../src/exporters/batch.ts","../../src/exporters/http.ts","../../src/client.ts"],"sourcesContent":["/**\n * Shared state via globalThis — ensures all entry point bundles share\n * the same singleton instances.\n *\n * Problem: tsup with `splitting: false` gives each entry point (index,\n * openai, anthropic, vercel-ai) its own copy of module-level variables.\n * This means `init()` from 'risicare' sets a tracer that 'risicare/openai'\n * can't see — breaking all provider instrumentation silently.\n *\n * Solution: Store all mutable singletons on globalThis with a namespaced\n * prefix. Every bundle reads/writes the same global slots.\n *\n * This pattern is used by React, OpenTelemetry, and other SDKs that must\n * share state across independently bundled entry points.\n *\n * @internal\n */\n\nimport { AsyncLocalStorage } from 'node:async_hooks';\n\n// eslint-disable-next-line @typescript-eslint/no-explicit-any\nconst G = globalThis as any;\nconst PREFIX = '__risicare_';\n\n// ─── Client & Tracer ────────────────────────────────────────────────────────\n\nexport function getClient(): unknown {\n return G[PREFIX + 'client'];\n}\n\nexport function setClient(client: unknown): void {\n G[PREFIX + 'client'] = client;\n}\n\nexport function getTracer(): unknown {\n return G[PREFIX + 'tracer'];\n}\n\nexport function setTracer(tracer: unknown): void {\n G[PREFIX + 'tracer'] = tracer;\n}\n\n// ─── Context Storage ────────────────────────────────────────────────────────\n\nexport function getContextStorage(): AsyncLocalStorage<unknown> {\n if (!G[PREFIX + 'ctx']) {\n G[PREFIX + 'ctx'] = new AsyncLocalStorage();\n }\n return G[PREFIX + 'ctx'];\n}\n\n// ─── Span Registry ──────────────────────────────────────────────────────────\n\nexport function getRegistry(): Map<string, unknown> {\n if (!G[PREFIX + 'registry']) {\n G[PREFIX + 'registry'] = new Map();\n }\n return G[PREFIX + 'registry'];\n}\n\nexport function getOpCount(): number {\n return G[PREFIX + 'opcount'] ?? 0;\n}\n\nexport function setOpCount(n: number): void {\n G[PREFIX + 'opcount'] = n;\n}\n\n// ─── Debug Flag ─────────────────────────────────────────────────────────────\n\nexport function getDebug(): boolean {\n return G[PREFIX + 'debug'] ?? false;\n}\n\nexport function setDebugFlag(enabled: boolean): void {\n G[PREFIX + 'debug'] = enabled;\n}\n\n// ─── Fix Runtime ────────────────────────────────────────────────────────\n\nexport function getGlobalFixRuntime(): unknown {\n return G[PREFIX + 'fix_runtime'];\n}\n\nexport function setGlobalFixRuntime(runtime: unknown): void {\n G[PREFIX + 'fix_runtime'] = runtime;\n}\n","/**\n * Internal logger for the Risicare SDK.\n *\n * Centralizes all diagnostic output so that:\n * - Debug messages are gated by a single flag (zero-cost when disabled)\n * - Warnings always fire (operational alerts like queue full)\n * - All output goes to stderr with a consistent [risicare] prefix\n * - A future custom logger callback can be added in one place\n */\n\nimport { getDebug, setDebugFlag } from '../globals.js';\n\n/**\n * Enable or disable debug logging. Called once during init().\n * @internal\n */\nexport function setDebug(enabled: boolean): void {\n setDebugFlag(enabled);\n}\n\n/**\n * Log a debug message. Only outputs when debug mode is enabled.\n * @internal\n */\nexport function debug(msg: string): void {\n if (getDebug()) {\n process.stderr.write(`[risicare] ${msg}\\n`);\n }\n}\n\n/**\n * Log a warning. Always outputs regardless of debug mode.\n * Use sparingly — only for operational issues the user should see.\n * @internal\n */\nexport function warn(msg: string): void {\n process.stderr.write(`[risicare] WARNING: ${msg}\\n`);\n}\n","/**\n * Instructor integration via Proxy wrapping.\n *\n * Instructor wraps an LLM client's create() method to extract structured\n * data from completions using a response model (Zod schema, etc.).\n * This integration intercepts the instructor client's create() to add a span\n * around the structured extraction call.\n *\n * Usage:\n * import { patchInstructor } from 'risicare/frameworks/instructor';\n * import Instructor from '@instructor-ai/instructor';\n * const client = patchInstructor(Instructor.from(openai));\n *\n * Does NOT suppress provider instrumentation -- Instructor delegates to the\n * underlying LLM client, which should produce its own provider span.\n */\n\nimport { getTracer } from '../client.js';\nimport { SpanKind } from '../types.js';\nimport { debug } from '../utils/log.js';\n\n/**\n * Wrap an Instructor client with Risicare tracing.\n *\n * Returns a Proxy that intercepts `create` calls to create a span\n * capturing the structured extraction metadata (response model, retries).\n *\n * @param client - An Instructor client with a create() method\n * @returns A proxied client with tracing enabled\n */\nexport function patchInstructor<T extends object>(client: T): T {\n return new Proxy(client, {\n get(target, prop, receiver) {\n const value = Reflect.get(target, prop, receiver);\n\n if (prop === 'create' && typeof value === 'function') {\n return function patchedCreate(\n this: unknown,\n ...args: unknown[]\n ): unknown {\n const tracer = getTracer();\n if (!tracer?.enabled) {\n debug(\n 'Tracer not initialized — call init() before using patchInstructor()',\n );\n return value.apply(this, args);\n }\n\n const params = (args[0] ?? {}) as Record<string, unknown>;\n const responseModel =\n params.response_model ?? params.responseModel;\n\n // Extract a readable name from the response model\n let modelName = 'unknown';\n if (responseModel && typeof responseModel === 'function') {\n modelName =\n (responseModel as { name?: string }).name || 'unknown';\n } else if (responseModel && typeof responseModel === 'object') {\n modelName =\n ((responseModel as Record<string, unknown>).name as string) ??\n ((responseModel as Record<string, unknown>).description as string) ??\n 'schema';\n } else if (typeof responseModel === 'string') {\n modelName = responseModel;\n }\n\n return tracer.startSpan(\n {\n name: 'instructor.create',\n kind: SpanKind.LLM_CALL,\n attributes: {\n framework: 'instructor',\n 'framework.span_kind': 'extraction',\n 'instructor.response_model': modelName,\n ...(params.max_retries != null\n ? { 'instructor.max_retries': params.max_retries }\n : {}),\n ...(params.model != null\n ? { 'gen_ai.request.model': params.model }\n : {}),\n },\n },\n (_span) => {\n const result = value.apply(this, args);\n\n // Handle async results (Promises)\n if (\n result &&\n typeof (result as Promise<unknown>).then === 'function'\n ) {\n return (result as Promise<unknown>).then((r) => r);\n }\n return result;\n },\n );\n };\n }\n\n return value;\n },\n });\n}\n","/**\n * ID generation for traces and spans.\n *\n * Trace IDs: 32 lowercase hex characters (16 random bytes)\n * Span IDs: 16 lowercase hex characters (8 random bytes)\n *\n * Uses crypto.randomBytes for cryptographically secure randomness.\n */\n\nimport { randomBytes } from 'node:crypto';\n\nconst HEX_REGEX_32 = /^[0-9a-f]{32}$/;\nconst HEX_REGEX_16 = /^[0-9a-f]{16}$/;\n\nexport function generateTraceId(): string {\n return randomBytes(16).toString('hex');\n}\n\nexport function generateSpanId(): string {\n return randomBytes(8).toString('hex');\n}\n\nexport function generateAgentId(prefix?: string): string {\n const suffix = randomBytes(8).toString('hex');\n return prefix ? `${prefix}-${suffix}` : suffix;\n}\n\nexport function validateTraceId(id: string): boolean {\n return HEX_REGEX_32.test(id);\n}\n\nexport function validateSpanId(id: string): boolean {\n return HEX_REGEX_16.test(id);\n}\n","/**\n * No-op implementations for the disabled path.\n *\n * When tracing is disabled, all operations return these no-op objects\n * to maintain zero overhead. No allocations, no side effects.\n */\n\nimport { SpanKind, SpanStatus, type SpanPayload } from './types.js';\n\n/**\n * A frozen no-op span that silently ignores all operations.\n * Used when SDK is disabled to avoid overhead.\n */\nexport const NOOP_SPAN = Object.freeze({\n traceId: '00000000000000000000000000000000',\n spanId: '0000000000000000',\n parentSpanId: undefined,\n name: 'noop',\n kind: SpanKind.INTERNAL,\n startTime: '',\n startHrtime: 0,\n endTime: undefined,\n status: SpanStatus.UNSET,\n statusMessage: undefined,\n attributes: Object.freeze({}) as Record<string, unknown>,\n events: Object.freeze([]) as readonly [],\n links: Object.freeze([]) as readonly [],\n sessionId: undefined,\n agentId: undefined,\n agentName: undefined,\n semanticPhase: undefined,\n llmProvider: undefined,\n llmModel: undefined,\n llmPromptTokens: undefined,\n llmCompletionTokens: undefined,\n llmTotalTokens: undefined,\n llmCostUsd: undefined,\n toolName: undefined,\n toolSuccess: undefined,\n isEnded: true,\n durationMs: 0,\n\n setAttribute() { return this; },\n setAttributes() { return this; },\n setStatus() { return this; },\n addEvent() { return this; },\n addLink() { return this; },\n recordException() { return this; },\n setLlmFields() { return this; },\n setToolFields() { return this; },\n end() {},\n toPayload(): SpanPayload {\n return {\n traceId: this.traceId,\n spanId: this.spanId,\n name: this.name,\n kind: this.kind,\n startTime: this.startTime,\n status: this.status,\n attributes: {},\n events: [],\n links: [],\n };\n },\n});\n\nexport type NoopSpan = typeof NOOP_SPAN;\n","/**\n * AsyncLocalStorage-based context propagation.\n *\n * Uses a single AsyncLocalStorage instance with a composite state object.\n * This is simpler and more performant than multiple separate stores.\n *\n * Node.js AsyncLocalStorage automatically propagates through:\n * - Promise / async-await\n * - setTimeout / setImmediate\n * - EventEmitter callbacks\n * - process.nextTick\n * - async generators (unlike Python's contextvars!)\n */\n\nimport { AsyncLocalStorage } from 'node:async_hooks';\nimport type { Span } from '../span.js';\nimport type { SemanticPhase } from '../types.js';\nimport { getContextStorage } from '../globals.js';\n\n// ─── Context Types ───────────────────────────────────────────────────────────\n\nexport interface SessionContext {\n sessionId: string;\n userId?: string;\n metadata?: Record<string, unknown>;\n parentSessionId?: string;\n turnNumber?: number;\n}\n\nexport interface AgentContext {\n agentId: string;\n agentName?: string;\n agentRole?: string;\n agentType?: string;\n parentAgentId?: string;\n version?: number;\n metadata?: Record<string, unknown>;\n}\n\nexport interface ContextState {\n session?: SessionContext;\n agent?: AgentContext;\n span?: Span;\n phase?: SemanticPhase;\n /** When true, provider instrumentors skip span creation (framework is handling it). */\n _suppressProviderInstrumentation?: boolean;\n /**\n * Pre-allocated trace ID from getTraceContext() when no span exists yet.\n * Ensures that getTraceContext().traceId matches the next span created\n * in this context — preventing the \"trace ID mismatch\" bug.\n */\n _rootTraceId?: string;\n}\n\n// ─── Storage Accessor ────────────────────────────────────────────────────────\n\nfunction storage(): AsyncLocalStorage<ContextState> {\n return getContextStorage() as AsyncLocalStorage<ContextState>;\n}\n\n// ─── Core Operations ─────────────────────────────────────────────────────────\n\n/**\n * Get the current context state, or empty object if outside any context.\n */\nexport function getContext(): ContextState {\n return storage().getStore() ?? {};\n}\n\n/**\n * Run a callback within a new context scope.\n * The new scope inherits from the parent, with overrides applied.\n */\nexport function runWithContext<T>(overrides: Partial<ContextState>, fn: () => T): T {\n const parent = getContext();\n const merged: ContextState = { ...parent, ...overrides };\n return storage().run(merged, fn);\n}\n\n/**\n * Run an async callback within a new context scope.\n */\nexport function runWithContextAsync<T>(overrides: Partial<ContextState>, fn: () => Promise<T>): Promise<T> {\n const parent = getContext();\n const merged: ContextState = { ...parent, ...overrides };\n return storage().run(merged, fn);\n}\n\n// ─── Context Accessors ───────────────────────────────────────────────────────\n\nexport function getCurrentSession(): SessionContext | undefined {\n return getContext().session;\n}\n\nexport function getCurrentAgent(): AgentContext | undefined {\n return getContext().agent;\n}\n\nexport function getCurrentSpan(): Span | undefined {\n return getContext().span;\n}\n\nexport function getCurrentPhase(): SemanticPhase | undefined {\n return getContext().phase;\n}\n\nexport function getCurrentSessionId(): string | undefined {\n return getContext().session?.sessionId;\n}\n\nexport function getCurrentAgentId(): string | undefined {\n return getContext().agent?.agentId;\n}\n\nexport function getCurrentTraceId(): string | undefined {\n return getContext().span?.traceId;\n}\n\nexport function getCurrentSpanId(): string | undefined {\n return getContext().span?.spanId;\n}\n\nexport function getCurrentParentSpanId(): string | undefined {\n return getContext().span?.parentSpanId;\n}\n\n/**\n * Get all current context as a plain object (for debugging/serialization).\n */\nexport function getCurrentContext(): Record<string, unknown> {\n const ctx = getContext();\n return {\n session: ctx.session ? {\n sessionId: ctx.session.sessionId,\n userId: ctx.session.userId,\n ...(ctx.session.parentSessionId !== undefined ? { parentSessionId: ctx.session.parentSessionId } : {}),\n ...(ctx.session.turnNumber !== undefined ? { turnNumber: ctx.session.turnNumber } : {}),\n ...(ctx.session.metadata !== undefined ? { metadata: ctx.session.metadata } : {}),\n } : null,\n agent: ctx.agent ? {\n agentId: ctx.agent.agentId,\n agentName: ctx.agent.agentName,\n agentRole: ctx.agent.agentRole,\n agentType: ctx.agent.agentType,\n ...(ctx.agent.parentAgentId !== undefined ? { parentAgentId: ctx.agent.parentAgentId } : {}),\n ...(ctx.agent.version !== undefined ? { version: ctx.agent.version } : {}),\n ...(ctx.agent.metadata !== undefined ? { metadata: ctx.agent.metadata } : {}),\n } : null,\n span: ctx.span ? { spanId: ctx.span.spanId, traceId: ctx.span.traceId } : null,\n phase: ctx.phase ?? null,\n };\n}\n","/**\n * Batch span processor.\n *\n * Collects spans and exports them in batches based on:\n * - Batch size threshold (default: 100 spans)\n * - Time interval (default: 1000ms)\n *\n * Node.js is single-threaded — no locks needed.\n * Timer is unref()'d so it doesn't prevent process exit.\n */\n\nimport type { Span } from '../span.js';\nimport { ExportResult, type SpanExporter } from './base.js';\nimport { debug, warn } from '../utils/log.js';\n\nexport interface BatchProcessorOptions {\n exporters: SpanExporter[];\n batchSize?: number;\n batchTimeoutMs?: number;\n maxQueueSize?: number;\n debug?: boolean;\n}\n\nexport class BatchSpanProcessor {\n private readonly _exporters: SpanExporter[];\n private readonly _batchSize: number;\n private readonly _batchTimeoutMs: number;\n private readonly _maxQueueSize: number;\n private readonly _debug: boolean;\n\n private _queue: Span[] = [];\n private _timer: ReturnType<typeof setInterval> | null = null;\n private _started = false;\n private _flushing = false;\n private _beforeExitHandler: (() => void) | null = null;\n\n // Retry tracking for failed batches (Audit #5)\n private _retryCounts = new Map<string, number>();\n private static readonly MAX_RETRIES = 3;\n\n // Metrics\n droppedSpans = 0;\n exportedSpans = 0;\n failedExports = 0;\n\n constructor(options: BatchProcessorOptions) {\n this._exporters = [...options.exporters];\n this._batchSize = options.batchSize ?? 100;\n this._batchTimeoutMs = options.batchTimeoutMs ?? 1000;\n this._maxQueueSize = options.maxQueueSize ?? 10000;\n this._debug = options.debug ?? false;\n }\n\n start(): void {\n if (this._started) return;\n this._started = true;\n\n // Periodic flush timer — unref so it doesn't keep process alive\n this._timer = setInterval(() => {\n void this._exportBatch();\n }, this._batchTimeoutMs);\n this._timer.unref();\n\n // Flush on process exit (Audit #3: use .once, not .on; await, not void)\n this._beforeExitHandler = () => {\n void this.shutdown();\n };\n process.once('beforeExit', this._beforeExitHandler);\n }\n\n async shutdown(timeoutMs = 5000): Promise<void> {\n if (!this._started) return;\n this._started = false;\n\n if (this._timer) {\n clearInterval(this._timer);\n this._timer = null;\n }\n\n // Remove process listener to prevent leak (Audit #3)\n if (this._beforeExitHandler) {\n process.removeListener('beforeExit', this._beforeExitHandler);\n this._beforeExitHandler = null;\n }\n\n // Final flush with timeout\n const flushPromise = this._exportBatch();\n const timeoutPromise = new Promise<void>((resolve) => setTimeout(resolve, timeoutMs));\n await Promise.race([flushPromise, timeoutPromise]);\n\n // Shutdown exporters\n for (const exporter of this._exporters) {\n try {\n await exporter.shutdown();\n } catch (e) {\n debug(`Error shutting down ${exporter.name}: ${e}`);\n }\n }\n\n // Clean up retry tracking\n this._retryCounts.clear();\n\n debug(\n `BatchSpanProcessor shutdown. Exported: ${this.exportedSpans}, ` +\n `Dropped: ${this.droppedSpans}, Failed: ${this.failedExports}`\n );\n }\n\n onSpanEnd(span: Span): void {\n if (!this._started) return;\n\n // Check queue capacity\n if (this._queue.length >= this._maxQueueSize) {\n this.droppedSpans++;\n if (this.droppedSpans === 1 || this.droppedSpans % 1000 === 0) {\n warn(`Span queue full (${this._maxQueueSize}). ${this.droppedSpans} spans dropped so far.`);\n }\n return;\n }\n\n this._queue.push(span);\n\n // Trigger immediate flush if batch size reached\n if (this._queue.length >= this._batchSize) {\n void this._exportBatch();\n }\n }\n\n async flush(timeoutMs = 5000): Promise<boolean> {\n if (!this._started) return true;\n\n const start = Date.now();\n\n // Wait for in-flight exports to complete AND queue to drain\n while (this._queue.length > 0 || this._flushing) {\n if (!this._flushing && this._queue.length > 0) {\n await this._exportBatch();\n } else {\n // Wait a tick for the in-flight export to finish\n await new Promise((r) => setTimeout(r, 1));\n }\n if (Date.now() - start > timeoutMs) return false;\n }\n\n return true;\n }\n\n getMetrics() {\n return {\n exportedSpans: this.exportedSpans,\n droppedSpans: this.droppedSpans,\n failedExports: this.failedExports,\n queueSize: this._queue.length,\n queueCapacity: this._maxQueueSize,\n queueUtilization: this._queue.length / this._maxQueueSize,\n };\n }\n\n private async _exportBatch(): Promise<void> {\n if (this._flushing || this._queue.length === 0) return;\n this._flushing = true;\n\n try {\n // Take up to batchSize spans\n const batch = this._queue.splice(0, this._batchSize);\n if (batch.length === 0) return;\n\n debug(`Exporting batch of ${batch.length} spans`);\n\n // Export to all exporters — count once, not per exporter\n let batchExported = false;\n for (const exporter of this._exporters) {\n try {\n const result = await exporter.export(batch);\n if (result === ExportResult.SUCCESS) {\n if (!batchExported) {\n this.exportedSpans += batch.length;\n batchExported = true;\n // Clear retry counts for successfully exported spans\n for (const span of batch) {\n this._retryCounts.delete(span.spanId);\n }\n }\n } else {\n this.failedExports++;\n }\n } catch (e) {\n this.failedExports++;\n debug(`Export to ${exporter.name} failed: ${e}`);\n }\n }\n\n // Audit #5: Re-queue failed batches with retry limit\n if (!batchExported) {\n const retryable = batch.filter(span => {\n const count = (this._retryCounts.get(span.spanId) ?? 0) + 1;\n if (count > BatchSpanProcessor.MAX_RETRIES) {\n this._retryCounts.delete(span.spanId);\n this.droppedSpans++;\n return false;\n }\n this._retryCounts.set(span.spanId, count);\n return true;\n });\n if (retryable.length > 0) {\n this._queue.unshift(...retryable);\n debug(`Re-queued ${retryable.length} spans for retry (${batch.length - retryable.length} dropped after max retries)`);\n }\n }\n } finally {\n this._flushing = false;\n }\n }\n}\n","/**\n * HTTP exporter for sending spans to the Risicare gateway.\n *\n * Uses native fetch (Node.js 18+), with:\n * - Automatic retry with exponential backoff\n * - Circuit breaker with half-open probe state\n * - Gzip compression for large payloads\n * - Bearer token authentication\n */\n\nimport type { Span } from '../span.js';\nimport type { IngestRequest } from '../types.js';\nimport { ExportResult, type SpanExporter } from './base.js';\nimport { debug, warn } from '../utils/log.js';\n\nconst SDK_VERSION = '0.1.1';\n\nexport interface HttpExporterOptions {\n endpoint: string;\n apiKey?: string;\n projectId?: string;\n environment?: string;\n timeoutMs?: number;\n maxRetries?: number;\n compress?: boolean;\n}\n\nexport class HttpExporter implements SpanExporter {\n readonly name = 'http';\n\n private readonly _endpoint: string;\n private readonly _apiKey: string | undefined;\n private readonly _projectId: string | undefined;\n private readonly _environment: string | undefined;\n private readonly _timeoutMs: number;\n private readonly _maxRetries: number;\n private readonly _compress: boolean;\n\n // Circuit breaker\n private _consecutiveFailures = 0;\n private _circuitOpenUntil = 0;\n private readonly _circuitBreakerThreshold = 3;\n private readonly _circuitBreakerCooldownMs = 30_000;\n\n constructor(options: HttpExporterOptions) {\n this._endpoint = options.endpoint.replace(/\\/+$/, '');\n this._apiKey = options.apiKey;\n this._projectId = options.projectId;\n this._environment = options.environment;\n this._timeoutMs = options.timeoutMs ?? 5000;\n this._maxRetries = options.maxRetries ?? 3;\n this._compress = options.compress ?? false;\n }\n\n async export(spans: Span[]): Promise<ExportResult> {\n if (spans.length === 0) return ExportResult.SUCCESS;\n\n // Circuit breaker: skip if open\n const now = Date.now();\n let isHalfOpen = false;\n if (this._consecutiveFailures >= this._circuitBreakerThreshold) {\n if (now < this._circuitOpenUntil) {\n return ExportResult.FAILURE;\n }\n // Cooldown expired — probe with single request (Audit #7: half-open state)\n isHalfOpen = true;\n }\n\n const body: IngestRequest = {\n spans: spans.map((s) => s.toPayload()),\n };\n if (this._projectId) body.projectId = this._projectId;\n if (this._environment) body.environment = this._environment;\n\n // Half-open: single probe request. Normal: full retry loop.\n const maxAttempts = isHalfOpen ? 1 : this._maxRetries;\n\n // Retry loop with exponential backoff\n for (let attempt = 0; attempt < maxAttempts; attempt++) {\n const result = await this._sendRequest(body);\n\n if (result === ExportResult.SUCCESS) {\n this._consecutiveFailures = 0;\n return result;\n }\n\n // Backoff: 100ms, 200ms, 400ms\n if (attempt < maxAttempts - 1) {\n await sleep(100 * Math.pow(2, attempt));\n }\n }\n\n // All retries failed — update circuit breaker\n this._consecutiveFailures++;\n if (this._consecutiveFailures >= this._circuitBreakerThreshold) {\n this._circuitOpenUntil = Date.now() + this._circuitBreakerCooldownMs;\n warn(\n `HTTP exporter circuit breaker opened after ${this._consecutiveFailures} failures. ` +\n `Cooldown: ${this._circuitBreakerCooldownMs / 1000}s`\n );\n }\n\n return ExportResult.FAILURE;\n }\n\n private async _sendRequest(body: IngestRequest): Promise<ExportResult> {\n const url = `${this._endpoint}/v1/spans`;\n const headers: Record<string, string> = {\n 'Content-Type': 'application/json',\n 'User-Agent': `risicare-js/${SDK_VERSION} node/${process.version}`,\n };\n if (this._apiKey) {\n headers['Authorization'] = `Bearer ${this._apiKey}`;\n }\n\n let payload: string | Uint8Array = JSON.stringify(body);\n\n if (this._compress && payload.length > 1024) {\n try {\n const { gzipSync } = await import('node:zlib');\n payload = gzipSync(Buffer.from(payload));\n headers['Content-Encoding'] = 'gzip';\n } catch (e) {\n // Audit #16: log compression failures instead of silently swallowing\n debug(`Gzip compression failed, sending uncompressed: ${e}`);\n }\n }\n\n try {\n const controller = new AbortController();\n const timeoutId = setTimeout(() => controller.abort(), this._timeoutMs);\n\n const response = await fetch(url, {\n method: 'POST',\n headers,\n body: payload,\n signal: controller.signal,\n });\n\n clearTimeout(timeoutId);\n\n if (response.status < 300) {\n return ExportResult.SUCCESS;\n } else if (response.status === 408 || response.status === 504) {\n debug(`HTTP export timeout: ${response.status}`);\n return ExportResult.TIMEOUT;\n }\n debug(`HTTP export failed: ${response.status}`);\n return ExportResult.FAILURE;\n } catch (e) {\n debug(`HTTP export error: ${e}`);\n return ExportResult.FAILURE;\n }\n }\n\n shutdown(): void {\n // Native fetch doesn't need cleanup\n }\n}\n\nfunction sleep(ms: number): Promise<void> {\n return new Promise((resolve) => setTimeout(resolve, ms));\n}\n","/**\n * RisicareClient — singleton client managing SDK lifecycle.\n *\n * Handles initialization, shutdown, and the connection between\n * the Tracer and the export pipeline (batch processor + HTTP exporter).\n *\n * Usage:\n * import { init, shutdown } from 'risicare';\n * init({ apiKey: 'rsk-...' }); // API key determines project\n * // ... instrument code ...\n * await shutdown(); // flush remaining spans\n */\n\nimport { type RisicareConfig, resolveConfig } from './config.js';\nimport { Tracer } from './tracer.js';\nimport { BatchSpanProcessor } from './exporters/batch.js';\nimport { HttpExporter } from './exporters/http.js';\nimport { ConsoleExporter } from './exporters/console.js';\nimport { SpanKind, SpanStatus } from './types.js';\nimport type { SpanExporter } from './exporters/base.js';\nimport { setDebug, debug } from './utils/log.js';\nimport {\n getClient as getGlobalClient,\n setClient as setGlobalClient,\n getTracer as getGlobalTracer,\n setTracer as setGlobalTracer,\n} from './globals.js';\n\n// ─── Client Class ───────────────────────────────────────────────────────────\n\nclass RisicareClient {\n readonly config: ReturnType<typeof resolveConfig>;\n readonly processor: BatchSpanProcessor;\n readonly tracer: Tracer;\n private _shutdownPromise: Promise<void> | undefined;\n private _shutdownHandlers: { signal: string; handler: () => void }[] = [];\n\n constructor(config?: Partial<RisicareConfig>) {\n this.config = resolveConfig(config);\n\n // API key format validation\n if (this.config.apiKey && !this.config.apiKey.startsWith('rsk-')) {\n debug('Warning: API key should start with \"rsk-\". Got: ' + this.config.apiKey.slice(0, 4) + '...');\n }\n\n // Build exporter chain\n let exporter: SpanExporter;\n if (this.config.debug && !this.config.apiKey) {\n exporter = new ConsoleExporter();\n } else if (this.config.apiKey) {\n exporter = new HttpExporter({\n endpoint: this.config.endpoint,\n apiKey: this.config.apiKey,\n projectId: this.config.projectId || undefined,\n environment: this.config.environment || undefined,\n compress: this.config.compress,\n });\n } else {\n // No API key and not debug — use console as fallback\n exporter = new ConsoleExporter();\n }\n\n this.processor = new BatchSpanProcessor({\n exporters: [exporter],\n batchSize: this.config.batchSize,\n batchTimeoutMs: this.config.batchTimeoutMs,\n maxQueueSize: this.config.maxQueueSize,\n debug: this.config.debug,\n });\n\n this.tracer = new Tracer({\n onSpanEnd: (span) => this.processor.onSpanEnd(span),\n sampleRate: this.config.sampleRate,\n enabled: this.config.enabled,\n traceContent: this.config.traceContent,\n });\n\n // Start the batch processor (enables span queuing and periodic flushing)\n this.processor.start();\n\n // Start FixRuntime for self-healing fix application.\n // Loads active fixes from /api/v1/fixes/active, caches with TTL,\n // background-refreshes every 60s. Provider patches call interceptCall()\n // to apply fixes before LLM calls.\n if (this.config.apiKey && this.config.enabled) {\n try {\n const { initFixRuntime } = require('./runtime/runtime.js') as typeof import('./runtime/runtime.js');\n initFixRuntime({\n apiEndpoint: this.config.endpoint,\n apiKey: this.config.apiKey,\n enabled: true,\n debug: this.config.debug,\n });\n } catch {\n // FixRuntime failure is non-fatal — tracing still works\n }\n }\n\n // Register shutdown hooks\n this._registerShutdownHooks();\n\n // Enable internal debug logging if configured\n setDebug(this.config.debug);\n debug(`Initialized: enabled=${this.config.enabled}, endpoint=${this.config.endpoint}`);\n }\n\n get enabled(): boolean {\n return this.tracer.enabled;\n }\n\n set enabled(value: boolean) {\n this.tracer.enabled = value;\n }\n\n // Audit #6: Promise-based shutdown dedup (fixes TOCTOU race condition)\n async shutdown(): Promise<void> {\n if (this._shutdownPromise) return this._shutdownPromise;\n this._shutdownPromise = this._doShutdown();\n return this._shutdownPromise;\n }\n\n private async _doShutdown(): Promise<void> {\n debug('Shutting down...');\n\n // Audit #3: Remove process listeners to prevent leak\n for (const { signal, handler } of this._shutdownHandlers) {\n process.removeListener(signal, handler);\n }\n this._shutdownHandlers = [];\n\n await this.processor.shutdown();\n }\n\n async flush(): Promise<void> {\n await this.processor.flush();\n }\n\n private _registerShutdownHooks(): void {\n const onShutdown = () => {\n // Audit #3: Add 5s timeout to prevent hanging on signal\n const timeout = setTimeout(() => process.exit(1), 5000);\n timeout.unref();\n this.shutdown().catch(() => {}).finally(() => clearTimeout(timeout));\n };\n\n const signals = ['beforeExit', 'SIGTERM', 'SIGINT'];\n for (const signal of signals) {\n process.once(signal, onShutdown);\n this._shutdownHandlers.push({ signal, handler: onShutdown });\n }\n }\n}\n\n// ─── Public API ─────────────────────────────────────────────────────────────\n\n/**\n * Initialize the Risicare SDK. Call once at application startup.\n *\n * @example\n * import { init } from 'risicare';\n * init({ apiKey: 'rsk-...', serviceName: 'my-agent', environment: 'production' });\n */\nexport function init(config?: Partial<RisicareConfig>): void {\n if (getGlobalClient()) {\n debug('Already initialized. Call shutdown() first to re-initialize.');\n return;\n }\n\n const client = new RisicareClient(config);\n setGlobalClient(client);\n setGlobalTracer(client.tracer);\n}\n\n/**\n * Gracefully shut down the SDK. Flushes pending spans before resolving.\n */\nexport async function shutdown(): Promise<void> {\n // Stop FixRuntime first (stops background refresh)\n try {\n const { shutdownFixRuntime } = require('./runtime/runtime.js') as typeof import('./runtime/runtime.js');\n shutdownFixRuntime();\n } catch {\n // FixRuntime may not be initialized\n }\n\n const client = getGlobalClient() as RisicareClient | undefined;\n if (!client) return;\n await client.shutdown();\n setGlobalClient(undefined);\n setGlobalTracer(undefined);\n}\n\n/**\n * Flush all pending spans without shutting down.\n */\nexport async function flush(): Promise<void> {\n const client = getGlobalClient() as RisicareClient | undefined;\n if (!client) return;\n await client.flush();\n}\n\n/**\n * Enable tracing at runtime.\n */\nexport function enable(): void {\n const client = getGlobalClient() as RisicareClient | undefined;\n if (client) client.enabled = true;\n}\n\n/**\n * Disable tracing at runtime. Spans will not be created or exported.\n */\nexport function disable(): void {\n const client = getGlobalClient() as RisicareClient | undefined;\n if (client) client.enabled = false;\n}\n\n/**\n * Check whether tracing is currently enabled.\n */\nexport function isEnabled(): boolean {\n const client = getGlobalClient() as RisicareClient | undefined;\n return client?.enabled ?? false;\n}\n\n/**\n * Get the global tracer instance. Returns undefined if not initialized.\n */\nexport function getTracer(): Tracer | undefined {\n return getGlobalTracer() as Tracer | undefined;\n}\n\n/**\n * Get the global tracer, or throw if not initialized.\n * @internal Used by decorators and providers that require an active tracer.\n */\nexport function requireTracer(): Tracer {\n const tracer = getGlobalTracer() as Tracer | undefined;\n if (!tracer) {\n throw new Error(\n 'Risicare SDK not initialized. Call init() before using tracing features.',\n );\n }\n return tracer;\n}\n\n/**\n * Check whether content tracing (prompt/completion capture) is enabled.\n */\nexport function getTraceContent(): boolean {\n const tracer = getGlobalTracer() as Tracer | undefined;\n return tracer?.traceContent ?? true;\n}\n\n/**\n * Get SDK metrics: exported spans, dropped spans, failed exports, queue stats.\n * Returns zero-valued metrics if SDK is not initialized.\n */\nexport function getMetrics() {\n const client = getGlobalClient() as RisicareClient | undefined;\n return client?.processor.getMetrics() ?? {\n exportedSpans: 0,\n droppedSpans: 0,\n failedExports: 0,\n queueSize: 0,\n queueCapacity: 0,\n queueUtilization: 0,\n };\n}\n\n// ─── reportError ──────────────────────────────────────────────────────────\n\n// Error dedup: SHA256 fingerprint → timestamp. Prevents retry loops from\n// creating N duplicate diagnoses. Matches Python SDK's 5-minute TTL.\nconst _ERROR_DEDUP_TTL_MS = 5 * 60 * 1000; // 5 minutes\nconst _ERROR_DEDUP_MAX = 1000;\nconst _recentErrors = new Map<string, number>(); // fingerprint → Date.now()\n\nfunction _errorFingerprint(err: Error): string {\n const raw = `${err.constructor?.name ?? 'Error'}:${String(err.message ?? '').slice(0, 200)}`;\n const { createHash } = require('node:crypto') as typeof import('node:crypto');\n return createHash('sha256').update(raw).digest('hex').slice(0, 16);\n}\n\nfunction _isDuplicateError(fingerprint: string): boolean {\n const now = Date.now();\n // Evict expired entries\n for (const [fp, ts] of _recentErrors) {\n if (now - ts > _ERROR_DEDUP_TTL_MS) _recentErrors.delete(fp);\n else break; // Map iterates in insertion order\n }\n if (_recentErrors.has(fingerprint)) return true;\n // Enforce max size\n if (_recentErrors.size >= _ERROR_DEDUP_MAX) {\n const oldest = _recentErrors.keys().next().value;\n if (oldest !== undefined) _recentErrors.delete(oldest);\n }\n _recentErrors.set(fingerprint, now);\n return false;\n}\n\n/**\n * Report a caught exception to the self-healing pipeline.\n *\n * Creates an error span that triggers diagnosis and fix generation.\n * Deduplicates identical errors within a 5-minute window (SHA256 fingerprint).\n * This function never throws and is non-blocking.\n *\n * @param error - The caught exception (Error object or string)\n * @param options - Optional attributes and context overrides\n */\nexport function reportError(\n error: unknown,\n options?: { name?: string; attributes?: Record<string, unknown> },\n): void {\n try {\n const tracer = getTracer();\n if (!tracer) return;\n\n const err = error instanceof Error ? error : new Error(String(error ?? 'unknown'));\n const spanName = options?.name ?? `error:${err.constructor.name}`;\n\n // Dedup: suppress identical errors within TTL (matches Python SDK)\n const fp = _errorFingerprint(err);\n if (_isDuplicateError(fp)) {\n debug(`reportError: duplicate suppressed (fp=${fp.slice(0, 8)})`);\n return;\n }\n\n tracer.startSpan({ name: spanName, kind: SpanKind.INTERNAL }, (span) => {\n span.setStatus(SpanStatus.ERROR, err.message);\n span.setAttribute('error', true);\n span.setAttribute('error.type', err.constructor.name);\n span.setAttribute('error.message', err.message.slice(0, 2000));\n if (err.stack) span.setAttribute('error.stack', err.stack.slice(0, 4000));\n span.setAttribute('risicare.reported_error', true);\n if (options?.attributes) {\n for (const [k, v] of Object.entries(options.attributes)) {\n span.setAttribute(k, v);\n }\n }\n });\n } catch {\n // Never crash the host application\n debug('reportError failed');\n }\n}\n\n// ─── score ─────────────────────────────────────────────────────────────────\n\n/**\n * Record a custom evaluation score on a trace.\n *\n * Sends the score to the server in a fire-and-forget fashion.\n * This function never throws and is non-blocking.\n *\n * @param traceId - The trace to score\n * @param name - Score name (e.g., \"accuracy\", \"user_satisfaction\")\n * @param value - Score value between 0.0 and 1.0 inclusive\n * @param options - Optional span_id and comment\n */\nexport function score(\n traceId: string,\n name: string,\n value: number,\n options?: { spanId?: string; comment?: string },\n): void {\n try {\n if (typeof value !== 'number' || value < 0.0 || value > 1.0) {\n debug(`score: value must be in [0.0, 1.0], got ${value}. Score not sent.`);\n return;\n }\n if (!traceId || !name) {\n debug('score: traceId and name are required');\n return;\n }\n\n const client = getGlobalClient() as RisicareClient | undefined;\n if (!client?.enabled || !client.config.apiKey) return;\n\n const endpoint = client.config.endpoint.replace(/\\/$/, '');\n const url = `${endpoint}/api/v1/scores`;\n const body = JSON.stringify({\n trace_id: traceId,\n name,\n score: value,\n source: 'sdk',\n ...(options?.spanId && { span_id: options.spanId }),\n ...(options?.comment && { comment: options.comment }),\n });\n\n // Fire-and-forget — never blocks caller\n fetch(url, {\n method: 'POST',\n headers: {\n 'Content-Type': 'application/json',\n 'Authorization': `Bearer ${client.config.apiKey}`,\n },\n body,\n }).catch((err) => debug(`score: send failed: ${err}`));\n } catch {\n // Never crash the host application\n debug('score failed');\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;AAkCO,SAAS,YAAqB;AACnC,SAAO,EAAE,SAAS,QAAQ;AAC5B;AAkCO,SAAS,WAAoB;AAClC,SAAO,EAAE,SAAS,OAAO,KAAK;AAChC;AAxEA,IAkBA,yBAGM,GACA;AAtBN;AAAA;AAAA;AAkBA,8BAAkC;AAGlC,IAAM,IAAI;AACV,IAAM,SAAS;AAAA;AAAA;;;ACER,SAAS,MAAM,KAAmB;AACvC,MAAI,SAAS,GAAG;AACd,YAAQ,OAAO,MAAM,cAAc,GAAG;AAAA,CAAI;AAAA,EAC5C;AACF;AA5BA;AAAA;AAAA;AAUA;AAAA;AAAA;;;ACVA;AAAA;AAAA;AAAA;AAAA;;;ACSA,yBAA4B;;;ACIrB,IAAM,YAAY,OAAO,OAAO;AAAA,EACrC,SAAS;AAAA,EACT,QAAQ;AAAA,EACR,cAAc;AAAA,EACd,MAAM;AAAA,EACN;AAAA,EACA,WAAW;AAAA,EACX,aAAa;AAAA,EACb,SAAS;AAAA,EACT;AAAA,EACA,eAAe;AAAA,EACf,YAAY,OAAO,OAAO,CAAC,CAAC;AAAA,EAC5B,QAAQ,OAAO,OAAO,CAAC,CAAC;AAAA,EACxB,OAAO,OAAO,OAAO,CAAC,CAAC;AAAA,EACvB,WAAW;AAAA,EACX,SAAS;AAAA,EACT,WAAW;AAAA,EACX,eAAe;AAAA,EACf,aAAa;AAAA,EACb,UAAU;AAAA,EACV,iBAAiB;AAAA,EACjB,qBAAqB;AAAA,EACrB,gBAAgB;AAAA,EAChB,YAAY;AAAA,EACZ,UAAU;AAAA,EACV,aAAa;AAAA,EACb,SAAS;AAAA,EACT,YAAY;AAAA,EAEZ,eAAe;AAAE,WAAO;AAAA,EAAM;AAAA,EAC9B,gBAAgB;AAAE,WAAO;AAAA,EAAM;AAAA,EAC/B,YAAY;AAAE,WAAO;AAAA,EAAM;AAAA,EAC3B,WAAW;AAAE,WAAO;AAAA,EAAM;AAAA,EAC1B,UAAU;AAAE,WAAO;AAAA,EAAM;AAAA,EACzB,kBAAkB;AAAE,WAAO;AAAA,EAAM;AAAA,EACjC,eAAe;AAAE,WAAO;AAAA,EAAM;AAAA,EAC9B,gBAAgB;AAAE,WAAO;AAAA,EAAM;AAAA,EAC/B,MAAM;AAAA,EAAC;AAAA,EACP,YAAyB;AACvB,WAAO;AAAA,MACL,SAAS,KAAK;AAAA,MACd,QAAQ,KAAK;AAAA,MACb,MAAM,KAAK;AAAA,MACX,MAAM,KAAK;AAAA,MACX,WAAW,KAAK;AAAA,MAChB,QAAQ,KAAK;AAAA,MACb,YAAY,CAAC;AAAA,MACb,QAAQ,CAAC;AAAA,MACT,OAAO,CAAC;AAAA,IACV;AAAA,EACF;AACF,CAAC;;;AC/CD;;;ACJA;;;ACAA;;;ACOA;AACA;AA+MO,SAASA,aAAgC;AAC9C,SAAO,UAAgB;AACzB;AA4CA,IAAM,sBAAsB,IAAI,KAAK;;;AN/PrC;AAWO,SAAS,gBAAkC,QAAc;AAC9D,SAAO,IAAI,MAAM,QAAQ;AAAA,IACvB,IAAI,QAAQ,MAAM,UAAU;AAC1B,YAAM,QAAQ,QAAQ,IAAI,QAAQ,MAAM,QAAQ;AAEhD,UAAI,SAAS,YAAY,OAAO,UAAU,YAAY;AACpD,eAAO,SAAS,iBAEX,MACM;AACT,gBAAM,SAASC,WAAU;AACzB,cAAI,CAAC,QAAQ,SAAS;AACpB;AAAA,cACE;AAAA,YACF;AACA,mBAAO,MAAM,MAAM,MAAM,IAAI;AAAA,UAC/B;AAEA,gBAAM,SAAU,KAAK,CAAC,KAAK,CAAC;AAC5B,gBAAM,gBACJ,OAAO,kBAAkB,OAAO;AAGlC,cAAI,YAAY;AAChB,cAAI,iBAAiB,OAAO,kBAAkB,YAAY;AACxD,wBACG,cAAoC,QAAQ;AAAA,UACjD,WAAW,iBAAiB,OAAO,kBAAkB,UAAU;AAC7D,wBACI,cAA0C,QAC1C,cAA0C,eAC5C;AAAA,UACJ,WAAW,OAAO,kBAAkB,UAAU;AAC5C,wBAAY;AAAA,UACd;AAEA,iBAAO,OAAO;AAAA,YACZ;AAAA,cACE,MAAM;AAAA,cACN;AAAA,cACA,YAAY;AAAA,gBACV,WAAW;AAAA,gBACX,uBAAuB;AAAA,gBACvB,6BAA6B;AAAA,gBAC7B,GAAI,OAAO,eAAe,OACtB,EAAE,0BAA0B,OAAO,YAAY,IAC/C,CAAC;AAAA,gBACL,GAAI,OAAO,SAAS,OAChB,EAAE,wBAAwB,OAAO,MAAM,IACvC,CAAC;AAAA,cACP;AAAA,YACF;AAAA,YACA,CAAC,UAAU;AACT,oBAAM,SAAS,MAAM,MAAM,MAAM,IAAI;AAGrC,kBACE,UACA,OAAQ,OAA4B,SAAS,YAC7C;AACA,uBAAQ,OAA4B,KAAK,CAAC,MAAM,CAAC;AAAA,cACnD;AACA,qBAAO;AAAA,YACT;AAAA,UACF;AAAA,QACF;AAAA,MACF;AAEA,aAAO;AAAA,IACT;AAAA,EACF,CAAC;AACH;","names":["getTracer","getTracer"]}
|
|
@@ -1,3 +1,39 @@
|
|
|
1
|
+
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
2
|
+
var __esm = (fn, res) => function __init() {
|
|
3
|
+
return fn && (res = (0, fn[__getOwnPropNames(fn)[0]])(fn = 0)), res;
|
|
4
|
+
};
|
|
5
|
+
|
|
6
|
+
// src/globals.ts
|
|
7
|
+
import { AsyncLocalStorage } from "async_hooks";
|
|
8
|
+
function getTracer() {
|
|
9
|
+
return G[PREFIX + "tracer"];
|
|
10
|
+
}
|
|
11
|
+
function getDebug() {
|
|
12
|
+
return G[PREFIX + "debug"] ?? false;
|
|
13
|
+
}
|
|
14
|
+
var G, PREFIX;
|
|
15
|
+
var init_globals = __esm({
|
|
16
|
+
"src/globals.ts"() {
|
|
17
|
+
"use strict";
|
|
18
|
+
G = globalThis;
|
|
19
|
+
PREFIX = "__risicare_";
|
|
20
|
+
}
|
|
21
|
+
});
|
|
22
|
+
|
|
23
|
+
// src/utils/log.ts
|
|
24
|
+
function debug(msg) {
|
|
25
|
+
if (getDebug()) {
|
|
26
|
+
process.stderr.write(`[risicare] ${msg}
|
|
27
|
+
`);
|
|
28
|
+
}
|
|
29
|
+
}
|
|
30
|
+
var init_log = __esm({
|
|
31
|
+
"src/utils/log.ts"() {
|
|
32
|
+
"use strict";
|
|
33
|
+
init_globals();
|
|
34
|
+
}
|
|
35
|
+
});
|
|
36
|
+
|
|
1
37
|
// src/ids.ts
|
|
2
38
|
import { randomBytes } from "crypto";
|
|
3
39
|
|
|
@@ -71,31 +107,25 @@ var NOOP_SPAN = Object.freeze({
|
|
|
71
107
|
}
|
|
72
108
|
});
|
|
73
109
|
|
|
74
|
-
// src/
|
|
75
|
-
|
|
76
|
-
var G = globalThis;
|
|
77
|
-
var PREFIX = "__risicare_";
|
|
78
|
-
function getTracer() {
|
|
79
|
-
return G[PREFIX + "tracer"];
|
|
80
|
-
}
|
|
81
|
-
function getDebug() {
|
|
82
|
-
return G[PREFIX + "debug"] ?? false;
|
|
83
|
-
}
|
|
110
|
+
// src/context/storage.ts
|
|
111
|
+
init_globals();
|
|
84
112
|
|
|
85
|
-
// src/
|
|
86
|
-
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
}
|
|
91
|
-
}
|
|
113
|
+
// src/exporters/batch.ts
|
|
114
|
+
init_log();
|
|
115
|
+
|
|
116
|
+
// src/exporters/http.ts
|
|
117
|
+
init_log();
|
|
92
118
|
|
|
93
119
|
// src/client.ts
|
|
120
|
+
init_log();
|
|
121
|
+
init_globals();
|
|
94
122
|
function getTracer2() {
|
|
95
123
|
return getTracer();
|
|
96
124
|
}
|
|
125
|
+
var _ERROR_DEDUP_TTL_MS = 5 * 60 * 1e3;
|
|
97
126
|
|
|
98
127
|
// src/frameworks/instructor.ts
|
|
128
|
+
init_log();
|
|
99
129
|
function patchInstructor(client) {
|
|
100
130
|
return new Proxy(client, {
|
|
101
131
|
get(target, prop, receiver) {
|