@poncho-ai/harness 0.28.3 → 0.30.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,5 +1,5 @@
1
1
 
2
- > @poncho-ai/harness@0.28.3 build /home/runner/work/poncho-ai/poncho-ai/packages/harness
2
+ > @poncho-ai/harness@0.30.0 build /home/runner/work/poncho-ai/poncho-ai/packages/harness
3
3
  > node scripts/embed-docs.js && tsup src/index.ts --format esm --dts
4
4
 
5
5
  [embed-docs] Generated poncho-docs.ts with 4 topics
@@ -8,8 +8,8 @@
8
8
  CLI tsup v8.5.1
9
9
  CLI Target: es2022
10
10
  ESM Build start
11
- ESM dist/index.js 291.95 KB
12
- ESM ⚡️ Build success in 123ms
11
+ ESM dist/index.js 300.00 KB
12
+ ESM ⚡️ Build success in 135ms
13
13
  DTS Build start
14
- DTS ⚡️ Build success in 6599ms
15
- DTS dist/index.d.ts 29.62 KB
14
+ DTS ⚡️ Build success in 7526ms
15
+ DTS dist/index.d.ts 30.64 KB
package/CHANGELOG.md CHANGED
@@ -1,5 +1,27 @@
1
1
  # @poncho-ai/harness
2
2
 
3
+ ## 0.30.0
4
+
5
+ ### Minor Changes
6
+
7
+ - [`193c367`](https://github.com/cesr/poncho-ai/commit/193c367568dce22a470dff6acd022c221be3b722) Thanks [@cesr](https://github.com/cesr)! - Unified continuation logic across all entry points (chat, cron, subagents, SDK) with mid-stream soft deadline checkpointing and proper context preservation across continuation boundaries.
8
+
9
+ ### Patch Changes
10
+
11
+ - Updated dependencies [[`193c367`](https://github.com/cesr/poncho-ai/commit/193c367568dce22a470dff6acd022c221be3b722)]:
12
+ - @poncho-ai/sdk@1.6.3
13
+
14
+ ## 0.29.0
15
+
16
+ ### Minor Changes
17
+
18
+ - [#51](https://github.com/cesr/poncho-ai/pull/51) [`eb661a5`](https://github.com/cesr/poncho-ai/commit/eb661a554da6839702651671db8a8820ceb13f35) Thanks [@cesr](https://github.com/cesr)! - Add generic OTLP trace exporter for sending OpenTelemetry traces to any collector (Jaeger, Grafana Tempo, Honeycomb, etc.). Configure via `telemetry.otlp` as a URL string or `{ url, headers }` object. Works alongside or instead of Latitude telemetry.
19
+
20
+ ### Patch Changes
21
+
22
+ - Updated dependencies [[`eb661a5`](https://github.com/cesr/poncho-ai/commit/eb661a554da6839702651671db8a8820ceb13f35)]:
23
+ - @poncho-ai/sdk@1.6.2
24
+
3
25
  ## 0.28.3
4
26
 
5
27
  ### Patch Changes
package/dist/index.d.ts CHANGED
@@ -176,6 +176,15 @@ interface Conversation {
176
176
  /** Harness-internal message chain preserved across continuation runs.
177
177
  * Cleared when a run completes without continuation. */
178
178
  _continuationMessages?: Message[];
179
+ /** Number of continuation pickups for the current multi-step run.
180
+ * Reset when a run completes without continuation. Used to enforce
181
+ * a maximum continuation count across all entry points. */
182
+ _continuationCount?: number;
183
+ /** Full structured message chain from the last harness run, including
184
+ * tool-call and tool-result messages the model needs for context.
185
+ * Unlike `_continuationMessages`, this is always set after a run
186
+ * and does NOT signal that a continuation is pending. */
187
+ _harnessMessages?: Message[];
179
188
  createdAt: number;
180
189
  updatedAt: number;
181
190
  }
@@ -411,7 +420,10 @@ interface PonchoConfig extends McpConfig {
411
420
  };
412
421
  telemetry?: {
413
422
  enabled?: boolean;
414
- otlp?: string;
423
+ otlp?: string | {
424
+ url: string;
425
+ headers?: Record<string, string>;
426
+ };
415
427
  latitude?: {
416
428
  apiKeyEnv?: string;
417
429
  projectIdEnv?: string;
@@ -623,6 +635,9 @@ declare class AgentHarness {
623
635
  private readonly activeSkillNames;
624
636
  private readonly registeredMcpToolNames;
625
637
  private latitudeTelemetry?;
638
+ private otlpSpanProcessor?;
639
+ private otlpTracerProvider?;
640
+ private hasOtlpExporter;
626
641
  private insideTelemetryCapture;
627
642
  private _browserSession?;
628
643
  private _browserMod?;
@@ -686,8 +701,9 @@ declare class AgentHarness {
686
701
  shutdown(): Promise<void>;
687
702
  listTools(): ToolDefinition[];
688
703
  /**
689
- * Wraps the run() generator with Latitude telemetry capture for complete trace coverage
690
- * Streams events in real-time using an event queue pattern
704
+ * Wraps the run() generator with telemetry capture for complete trace coverage.
705
+ * Supports Latitude, generic OTLP, or both simultaneously.
706
+ * Streams events in real-time using an event queue pattern.
691
707
  */
692
708
  runWithTelemetry(input: RunInput): AsyncGenerator<AgentEvent>;
693
709
  compact(messages: Message[], options?: CompactMessagesOptions): Promise<CompactResult>;
@@ -804,9 +820,15 @@ declare const createSkillTools: (skills: SkillMetadata[], options?: {
804
820
  }) => ToolDefinition[];
805
821
  declare const normalizeScriptPolicyPath: (relativePath: string) => string;
806
822
 
823
+ interface OtlpConfig {
824
+ url: string;
825
+ headers?: Record<string, string>;
826
+ }
827
+ type OtlpOption = string | OtlpConfig;
828
+ declare function normalizeOtlp(opt: OtlpOption | undefined): OtlpConfig | undefined;
807
829
  interface TelemetryConfig {
808
830
  enabled?: boolean;
809
- otlp?: string;
831
+ otlp?: OtlpOption;
810
832
  latitude?: {
811
833
  apiKeyEnv?: string;
812
834
  projectIdEnv?: string;
@@ -824,4 +846,4 @@ declare class TelemetryEmitter {
824
846
 
825
847
  declare const createSubagentTools: (manager: SubagentManager) => ToolDefinition[];
826
848
 
827
- export { type AgentFrontmatter, AgentHarness, type AgentIdentity, type AgentLimitsConfig, type AgentModelConfig, type BuiltInToolToggles, type CompactMessagesOptions, type CompactResult, type CompactionConfig, type Conversation, type ConversationState, type ConversationStore, type ConversationSummary, type CronJobConfig, type HarnessOptions, type HarnessRunOutput, InMemoryConversationStore, InMemoryStateStore, LatitudeCapture, type LatitudeCaptureConfig, LocalMcpBridge, LocalUploadStore, type MainMemory, type McpConfig, type MemoryConfig, type MemoryStore, type MessagingChannelConfig, type ModelProviderFactory, PONCHO_UPLOAD_SCHEME, type ParsedAgent, type PendingSubagentResult, type PonchoConfig, type ProviderConfig, type RemoteMcpServerConfig, type RuntimeRenderContext, S3UploadStore, STORAGE_SCHEMA_VERSION, type SkillContextEntry, type SkillMetadata, type StateConfig, type StateProviderName, type StateStore, type StorageConfig, type SubagentManager, type SubagentResult, type SubagentSpawnResult, type SubagentSummary, type TelemetryConfig, TelemetryEmitter, type ToolAccess, type ToolCall, ToolDispatcher, type ToolExecutionResult, type UploadStore, type UploadsConfig, VercelBlobUploadStore, buildAgentDirectoryName, buildSkillContextWindow, compactMessages, createConversationStore, createDefaultTools, createDeleteDirectoryTool, createDeleteTool, createEditTool, createMemoryStore, createMemoryTools, createModelProvider, createSearchTools, createSkillTools, createStateStore, createSubagentTools, createUploadStore, createWriteTool, deriveUploadKey, ensureAgentIdentity, estimateTokens, estimateTotalTokens, findSafeSplitPoint, generateAgentId, getAgentStoreDirectory, getModelContextWindow, getPonchoStoreRoot, jsonSchemaToZod, loadPonchoConfig, loadSkillContext, loadSkillInstructions, loadSkillMetadata, normalizeScriptPolicyPath, parseAgentFile, parseAgentMarkdown, ponchoDocsTool, readSkillResource, renderAgentPrompt, resolveAgentIdentity, resolveCompactionConfig, resolveMemoryConfig, resolveSkillDirs, resolveStateConfig, slugifyStorageComponent };
849
+ export { type AgentFrontmatter, AgentHarness, type AgentIdentity, type AgentLimitsConfig, type AgentModelConfig, type BuiltInToolToggles, type CompactMessagesOptions, type CompactResult, type CompactionConfig, type Conversation, type ConversationState, type ConversationStore, type ConversationSummary, type CronJobConfig, type HarnessOptions, type HarnessRunOutput, InMemoryConversationStore, InMemoryStateStore, LatitudeCapture, type LatitudeCaptureConfig, LocalMcpBridge, LocalUploadStore, type MainMemory, type McpConfig, type MemoryConfig, type MemoryStore, type MessagingChannelConfig, type ModelProviderFactory, type OtlpConfig, type OtlpOption, PONCHO_UPLOAD_SCHEME, type ParsedAgent, type PendingSubagentResult, type PonchoConfig, type ProviderConfig, type RemoteMcpServerConfig, type RuntimeRenderContext, S3UploadStore, STORAGE_SCHEMA_VERSION, type SkillContextEntry, type SkillMetadata, type StateConfig, type StateProviderName, type StateStore, type StorageConfig, type SubagentManager, type SubagentResult, type SubagentSpawnResult, type SubagentSummary, type TelemetryConfig, TelemetryEmitter, type ToolAccess, type ToolCall, ToolDispatcher, type ToolExecutionResult, type UploadStore, type UploadsConfig, VercelBlobUploadStore, buildAgentDirectoryName, buildSkillContextWindow, compactMessages, createConversationStore, createDefaultTools, createDeleteDirectoryTool, createDeleteTool, createEditTool, createMemoryStore, createMemoryTools, createModelProvider, createSearchTools, createSkillTools, createStateStore, createSubagentTools, createUploadStore, createWriteTool, deriveUploadKey, ensureAgentIdentity, estimateTokens, estimateTotalTokens, findSafeSplitPoint, generateAgentId, getAgentStoreDirectory, getModelContextWindow, getPonchoStoreRoot, jsonSchemaToZod, loadPonchoConfig, loadSkillContext, loadSkillInstructions, loadSkillMetadata, normalizeOtlp, normalizeScriptPolicyPath, parseAgentFile, parseAgentMarkdown, ponchoDocsTool, readSkillResource, renderAgentPrompt, resolveAgentIdentity, resolveCompactionConfig, resolveMemoryConfig, resolveSkillDirs, resolveStateConfig, slugifyStorageComponent };
package/dist/index.js CHANGED
@@ -1529,11 +1529,17 @@ export default {
1529
1529
  },
1530
1530
  },
1531
1531
 
1532
- // Telemetry destination
1532
+ // Telemetry destination \u2014 generic OTLP and/or Latitude
1533
1533
  telemetry: {
1534
1534
  enabled: true,
1535
+ // Generic OTLP: string shorthand or { url, headers? } object
1535
1536
  otlp: process.env.OTEL_EXPORTER_OTLP_ENDPOINT,
1536
- // Or use Latitude (reads from LATITUDE_API_KEY and LATITUDE_PROJECT_ID env vars by default)
1537
+ // With auth headers (Honeycomb, Grafana Cloud, etc.):
1538
+ // otlp: {
1539
+ // url: 'https://api.honeycomb.io/v1/traces',
1540
+ // headers: { 'x-honeycomb-team': process.env.HONEYCOMB_API_KEY },
1541
+ // },
1542
+ // Latitude (reads from LATITUDE_API_KEY and LATITUDE_PROJECT_ID env vars by default)
1537
1543
  latitude: {
1538
1544
  // apiKeyEnv: 'LATITUDE_API_KEY', // default
1539
1545
  // projectIdEnv: 'LATITUDE_PROJECT_ID', // default
@@ -1606,7 +1612,7 @@ Remote storage keys are namespaced and versioned, for example \`poncho:v1:<agent
1606
1612
  | \`PONCHO_AUTH_TOKEN\` | No | Unified auth token (Web UI passphrase + API Bearer token) |
1607
1613
  | \`PONCHO_INTERNAL_SECRET\` | No | Shared secret used by internal serverless callbacks (recommended for Vercel/Lambda) |
1608
1614
  | \`PONCHO_SELF_BASE_URL\` | No | Explicit base URL for internal self-callbacks when auto-detection is unavailable |
1609
- | \`OTEL_EXPORTER_OTLP_ENDPOINT\` | No | Telemetry destination |
1615
+ | \`OTEL_EXPORTER_OTLP_ENDPOINT\` | No | OTLP trace endpoint (Jaeger, Tempo, Honeycomb, etc.) |
1610
1616
  | \`LATITUDE_API_KEY\` | No | Latitude dashboard integration |
1611
1617
  | \`LATITUDE_PROJECT_ID\` | No | Latitude project identifier for capture traces |
1612
1618
  | \`LATITUDE_PATH\` | No | Latitude prompt path for grouping traces |
@@ -1641,23 +1647,45 @@ Logs print to console:
1641
1647
  [event] run:completed {"type":"run:completed","runId":"run_abc123","result":{"status":"completed","response":"...","steps":3,"tokens":{"input":1500,"output":840}}}
1642
1648
  \`\`\`
1643
1649
 
1644
- ### Production telemetry
1650
+ ### Production telemetry (generic OTLP)
1645
1651
 
1646
- Send events to your observability stack:
1652
+ Send full OpenTelemetry traces (agent runs, LLM calls, tool executions) to any
1653
+ OTLP-compatible collector \u2014 Jaeger, Grafana Tempo, Honeycomb, Datadog, etc.
1647
1654
 
1648
1655
  \`\`\`bash
1649
- # Environment variable
1650
- OTEL_EXPORTER_OTLP_ENDPOINT=https://otel.example.com
1656
+ # Simple: just a URL
1657
+ OTEL_EXPORTER_OTLP_ENDPOINT=https://otel.example.com/v1/traces
1651
1658
  \`\`\`
1652
1659
 
1653
- Or configure in code:
1660
+ \`\`\`javascript
1661
+ // poncho.config.js \u2014 string shorthand
1662
+ export default {
1663
+ telemetry: {
1664
+ otlp: 'https://otel.example.com/v1/traces',
1665
+ }
1666
+ }
1667
+ \`\`\`
1668
+
1669
+ \`\`\`javascript
1670
+ // poncho.config.js \u2014 with auth headers (Honeycomb, Grafana Cloud, etc.)
1671
+ export default {
1672
+ telemetry: {
1673
+ otlp: {
1674
+ url: 'https://api.honeycomb.io/v1/traces',
1675
+ headers: {
1676
+ 'x-honeycomb-team': process.env.HONEYCOMB_API_KEY,
1677
+ },
1678
+ },
1679
+ }
1680
+ }
1681
+ \`\`\`
1682
+
1683
+ You can also use a custom event handler for non-OTLP destinations:
1654
1684
 
1655
1685
  \`\`\`javascript
1656
1686
  // poncho.config.js
1657
1687
  export default {
1658
1688
  telemetry: {
1659
- otlp: 'https://otel.example.com',
1660
- // Or custom handler
1661
1689
  handler: async (event) => {
1662
1690
  await sendToMyLoggingService(event)
1663
1691
  }
@@ -1687,6 +1715,8 @@ telemetry: {
1687
1715
  }
1688
1716
  \`\`\`
1689
1717
 
1718
+ Both \`otlp\` and \`latitude\` can be configured simultaneously \u2014 all spans flow to both destinations.
1719
+
1690
1720
  ## Security
1691
1721
 
1692
1722
  ### Protect your endpoint
@@ -4581,6 +4611,78 @@ var createSubagentTools = (manager) => [
4581
4611
 
4582
4612
  // src/harness.ts
4583
4613
  import { LatitudeTelemetry } from "@latitude-data/telemetry";
4614
+ import { trace, context as otelContext, SpanStatusCode } from "@opentelemetry/api";
4615
+ import { NodeTracerProvider, BatchSpanProcessor } from "@opentelemetry/sdk-trace-node";
4616
+ import { OTLPTraceExporter } from "@opentelemetry/exporter-trace-otlp-http";
4617
+
4618
+ // src/telemetry.ts
4619
+ var MAX_FIELD_LENGTH = 200;
4620
+ function sanitizeEventForLog(event) {
4621
+ return JSON.stringify(event, (_key, value) => {
4622
+ if (typeof value === "string" && value.length > MAX_FIELD_LENGTH) {
4623
+ return `${value.slice(0, 80)}...[${value.length} chars]`;
4624
+ }
4625
+ return value;
4626
+ });
4627
+ }
4628
+ function normalizeOtlp(opt) {
4629
+ if (!opt) return void 0;
4630
+ if (typeof opt === "string") return opt ? { url: opt } : void 0;
4631
+ return opt.url ? opt : void 0;
4632
+ }
4633
+ var TelemetryEmitter = class {
4634
+ config;
4635
+ constructor(config) {
4636
+ this.config = config;
4637
+ }
4638
+ async emit(event) {
4639
+ if (this.config?.enabled === false) {
4640
+ return;
4641
+ }
4642
+ if (this.config?.handler) {
4643
+ await this.config.handler(event);
4644
+ return;
4645
+ }
4646
+ const otlp = normalizeOtlp(this.config?.otlp);
4647
+ if (otlp) {
4648
+ await this.sendOtlp(event, otlp);
4649
+ }
4650
+ process.stdout.write(`[event] ${event.type} ${sanitizeEventForLog(event)}
4651
+ `);
4652
+ }
4653
+ async sendOtlp(event, otlp) {
4654
+ try {
4655
+ await fetch(otlp.url, {
4656
+ method: "POST",
4657
+ headers: { "Content-Type": "application/json", ...otlp.headers },
4658
+ body: JSON.stringify({
4659
+ resourceLogs: [
4660
+ {
4661
+ scopeLogs: [
4662
+ {
4663
+ logRecords: [
4664
+ {
4665
+ timeUnixNano: String(Date.now() * 1e6),
4666
+ severityText: "INFO",
4667
+ body: { stringValue: event.type },
4668
+ attributes: [
4669
+ {
4670
+ key: "event.payload",
4671
+ value: { stringValue: JSON.stringify(event) }
4672
+ }
4673
+ ]
4674
+ }
4675
+ ]
4676
+ }
4677
+ ]
4678
+ }
4679
+ ]
4680
+ })
4681
+ });
4682
+ } catch {
4683
+ }
4684
+ }
4685
+ };
4584
4686
 
4585
4687
  // src/tool-dispatcher.ts
4586
4688
  var ToolDispatcher = class {
@@ -5113,6 +5215,9 @@ var AgentHarness = class _AgentHarness {
5113
5215
  activeSkillNames = /* @__PURE__ */ new Set();
5114
5216
  registeredMcpToolNames = /* @__PURE__ */ new Set();
5115
5217
  latitudeTelemetry;
5218
+ otlpSpanProcessor;
5219
+ otlpTracerProvider;
5220
+ hasOtlpExporter = false;
5116
5221
  insideTelemetryCapture = false;
5117
5222
  _browserSession;
5118
5223
  _browserMod;
@@ -5560,6 +5665,31 @@ var AgentHarness = class _AgentHarness {
5560
5665
  `[poncho][telemetry] Latitude telemetry is configured but missing: ${missing.join(", ")}. Traces will NOT be sent.`
5561
5666
  );
5562
5667
  }
5668
+ const otlpConfig = telemetryEnabled ? normalizeOtlp(config?.telemetry?.otlp) : void 0;
5669
+ if (otlpConfig) {
5670
+ const exporter = new OTLPTraceExporter({
5671
+ url: otlpConfig.url,
5672
+ headers: otlpConfig.headers
5673
+ });
5674
+ const processor = new BatchSpanProcessor(exporter);
5675
+ this.otlpSpanProcessor = processor;
5676
+ if (this.latitudeTelemetry) {
5677
+ const globalProvider = trace.getTracerProvider();
5678
+ const delegate = globalProvider.getDelegate?.() ?? globalProvider;
5679
+ if (typeof delegate.addSpanProcessor === "function") {
5680
+ delegate.addSpanProcessor(processor);
5681
+ }
5682
+ console.info(`[poncho][telemetry] OTLP exporter added (piggybacking on Latitude provider) \u2192 ${otlpConfig.url}`);
5683
+ } else {
5684
+ const provider2 = new NodeTracerProvider({
5685
+ spanProcessors: [processor]
5686
+ });
5687
+ provider2.register();
5688
+ this.otlpTracerProvider = provider2;
5689
+ console.info(`[poncho][telemetry] OTLP exporter active (standalone provider) \u2192 ${otlpConfig.url}`);
5690
+ }
5691
+ this.hasOtlpExporter = true;
5692
+ }
5563
5693
  }
5564
5694
  async buildBrowserStoragePersistence(config, sessionId) {
5565
5695
  const provider = config.storage?.provider ?? config.state?.provider ?? "local";
@@ -5710,13 +5840,31 @@ var AgentHarness = class _AgentHarness {
5710
5840
  });
5711
5841
  this.latitudeTelemetry = void 0;
5712
5842
  }
5843
+ if (this.otlpSpanProcessor) {
5844
+ await this.otlpSpanProcessor.shutdown().catch((err) => {
5845
+ console.warn(
5846
+ `[poncho][telemetry] OTLP span processor shutdown error: ${err instanceof Error ? err.message : String(err)}`
5847
+ );
5848
+ });
5849
+ this.otlpSpanProcessor = void 0;
5850
+ }
5851
+ if (this.otlpTracerProvider) {
5852
+ await this.otlpTracerProvider.shutdown().catch((err) => {
5853
+ console.warn(
5854
+ `[poncho][telemetry] OTLP tracer provider shutdown error: ${err instanceof Error ? err.message : String(err)}`
5855
+ );
5856
+ });
5857
+ this.otlpTracerProvider = void 0;
5858
+ }
5859
+ this.hasOtlpExporter = false;
5713
5860
  }
5714
5861
  listTools() {
5715
5862
  return this.dispatcher.list();
5716
5863
  }
5717
5864
  /**
5718
- * Wraps the run() generator with Latitude telemetry capture for complete trace coverage
5719
- * Streams events in real-time using an event queue pattern
5865
+ * Wraps the run() generator with telemetry capture for complete trace coverage.
5866
+ * Supports Latitude, generic OTLP, or both simultaneously.
5867
+ * Streams events in real-time using an event queue pattern.
5720
5868
  */
5721
5869
  async *runWithTelemetry(input) {
5722
5870
  const config = this.loadedConfig;
@@ -5783,6 +5931,39 @@ var AgentHarness = class _AgentHarness {
5783
5931
  }
5784
5932
  }
5785
5933
  }
5934
+ } else if (this.hasOtlpExporter) {
5935
+ const tracer = trace.getTracer("poncho");
5936
+ const agentName = this.parsedAgent?.frontmatter.name ?? "agent";
5937
+ const rootSpan = tracer.startSpan(`agent.run ${agentName}`);
5938
+ rootSpan.setAttribute("poncho.agent.name", agentName);
5939
+ if (input.conversationId) {
5940
+ rootSpan.setAttribute("poncho.conversation.id", input.conversationId);
5941
+ }
5942
+ const spanContext = trace.setSpan(otelContext.active(), rootSpan);
5943
+ this.insideTelemetryCapture = true;
5944
+ try {
5945
+ const gen = this.run(input);
5946
+ let next;
5947
+ do {
5948
+ next = await otelContext.with(spanContext, () => gen.next());
5949
+ if (!next.done) yield next.value;
5950
+ } while (!next.done);
5951
+ rootSpan.setStatus({ code: SpanStatusCode.OK });
5952
+ } catch (error) {
5953
+ rootSpan.setStatus({
5954
+ code: SpanStatusCode.ERROR,
5955
+ message: error instanceof Error ? error.message : String(error)
5956
+ });
5957
+ rootSpan.recordException(error instanceof Error ? error : new Error(String(error)));
5958
+ throw error;
5959
+ } finally {
5960
+ this.insideTelemetryCapture = false;
5961
+ rootSpan.end();
5962
+ try {
5963
+ await this.otlpSpanProcessor?.forceFlush();
5964
+ } catch {
5965
+ }
5966
+ }
5786
5967
  } else {
5787
5968
  yield* this.run(input);
5788
5969
  }
@@ -5950,7 +6131,7 @@ ${this.skillFingerprint}`;
5950
6131
  if (lastMsg && lastMsg.role !== "user") {
5951
6132
  messages.push({
5952
6133
  role: "user",
5953
- content: "[System: Your previous turn was interrupted by a time limit. Continue from where you left off \u2014 do NOT repeat what you already said. Proceed directly with the next action or tool call.]",
6134
+ content: "[System: Your previous turn was interrupted by a time limit. Your partial response above is already visible to the user. Continue EXACTLY from where you left off \u2014 do NOT restart, re-summarize, or repeat any content you already produced. If you were mid-sentence or mid-table, continue that sentence or table. Proceed directly with the next action or output.]",
5954
6135
  metadata: { timestamp: now(), id: randomUUID3() }
5955
6136
  });
5956
6137
  }
@@ -6271,7 +6452,7 @@ ${textContent}` };
6271
6452
  abortSignal: input.abortSignal,
6272
6453
  ...typeof maxTokens === "number" ? { maxTokens } : {},
6273
6454
  experimental_telemetry: {
6274
- isEnabled: telemetryEnabled && !!this.latitudeTelemetry,
6455
+ isEnabled: telemetryEnabled && !!(this.latitudeTelemetry || this.hasOtlpExporter),
6275
6456
  recordInputs: true,
6276
6457
  recordOutputs: true
6277
6458
  }
@@ -6280,7 +6461,10 @@ ${textContent}` };
6280
6461
  let chunkCount = 0;
6281
6462
  const hasRunTimeout = timeoutMs > 0;
6282
6463
  const streamDeadline = hasRunTimeout ? start + timeoutMs : 0;
6464
+ const hasSoftDeadline = softDeadlineMs > 0;
6465
+ const INTER_CHUNK_TIMEOUT_MS = 6e4;
6283
6466
  const fullStreamIterator = result.fullStream[Symbol.asyncIterator]();
6467
+ let softDeadlineFiredDuringStream = false;
6284
6468
  try {
6285
6469
  while (true) {
6286
6470
  if (isCancelled()) {
@@ -6288,8 +6472,8 @@ ${textContent}` };
6288
6472
  return;
6289
6473
  }
6290
6474
  if (hasRunTimeout) {
6291
- const remaining2 = streamDeadline - now();
6292
- if (remaining2 <= 0) {
6475
+ const remaining = streamDeadline - now();
6476
+ if (remaining <= 0) {
6293
6477
  yield pushEvent({
6294
6478
  type: "run:error",
6295
6479
  runId,
@@ -6304,22 +6488,33 @@ ${textContent}` };
6304
6488
  return;
6305
6489
  }
6306
6490
  }
6307
- const remaining = hasRunTimeout ? streamDeadline - now() : Infinity;
6308
- const timeout = chunkCount === 0 ? Math.min(remaining, FIRST_CHUNK_TIMEOUT_MS) : hasRunTimeout ? remaining : 0;
6491
+ if (hasSoftDeadline && chunkCount > 0 && now() - start >= softDeadlineMs) {
6492
+ softDeadlineFiredDuringStream = true;
6493
+ break;
6494
+ }
6495
+ const hardRemaining = hasRunTimeout ? streamDeadline - now() : Infinity;
6496
+ const softRemaining = hasSoftDeadline ? Math.max(0, start + softDeadlineMs - now()) : Infinity;
6497
+ const deadlineRemaining = Math.min(hardRemaining, softRemaining);
6498
+ const timeout = chunkCount === 0 ? Math.min(deadlineRemaining, FIRST_CHUNK_TIMEOUT_MS) : Math.min(deadlineRemaining, INTER_CHUNK_TIMEOUT_MS);
6309
6499
  let nextPart;
6310
- if (timeout <= 0 && chunkCount > 0) {
6500
+ if (timeout <= 0 && chunkCount > 0 && !hasSoftDeadline) {
6311
6501
  nextPart = await fullStreamIterator.next();
6312
6502
  } else {
6503
+ const effectiveTimeout = Math.max(timeout, 1);
6313
6504
  let timer;
6314
6505
  nextPart = await Promise.race([
6315
6506
  fullStreamIterator.next(),
6316
6507
  new Promise((resolve12) => {
6317
- timer = setTimeout(() => resolve12(null), timeout);
6508
+ timer = setTimeout(() => resolve12(null), effectiveTimeout);
6318
6509
  })
6319
6510
  ]);
6320
6511
  clearTimeout(timer);
6321
6512
  }
6322
6513
  if (nextPart === null) {
6514
+ if (hasSoftDeadline && deadlineRemaining <= INTER_CHUNK_TIMEOUT_MS) {
6515
+ softDeadlineFiredDuringStream = true;
6516
+ break;
6517
+ }
6323
6518
  const isFirstChunk = chunkCount === 0;
6324
6519
  console.error(
6325
6520
  `[poncho][harness] Stream timeout waiting for ${isFirstChunk ? "first" : "next"} chunk: model="${modelName}", step=${step}, chunks=${chunkCount}, elapsed=${now() - start}ms`
@@ -6352,11 +6547,42 @@ ${textContent}` };
6352
6547
  fullStreamIterator.return?.(void 0)?.catch?.(() => {
6353
6548
  });
6354
6549
  }
6550
+ if (softDeadlineFiredDuringStream) {
6551
+ if (fullText.length > 0) {
6552
+ messages.push({
6553
+ role: "assistant",
6554
+ content: fullText,
6555
+ metadata: { timestamp: now(), id: randomUUID3(), step }
6556
+ });
6557
+ }
6558
+ const result_ = {
6559
+ status: "completed",
6560
+ response: responseText + fullText,
6561
+ steps: step,
6562
+ tokens: { input: totalInputTokens, output: totalOutputTokens, cached: totalCachedTokens },
6563
+ duration: now() - start,
6564
+ continuation: true,
6565
+ continuationMessages: [...messages],
6566
+ maxSteps,
6567
+ contextTokens: latestContextTokens + toolOutputEstimateSinceModel,
6568
+ contextWindow
6569
+ };
6570
+ console.info(`[poncho][harness] Soft deadline fired mid-stream at step ${step} (${(now() - start).toFixed(0)}ms). Checkpointing with ${fullText.length} chars of partial text.`);
6571
+ yield pushEvent({ type: "run:completed", runId, result: result_ });
6572
+ return;
6573
+ }
6355
6574
  if (isCancelled()) {
6356
6575
  yield emitCancellation();
6357
6576
  return;
6358
6577
  }
6359
6578
  if (softDeadlineMs > 0 && now() - start > softDeadlineMs) {
6579
+ if (fullText.length > 0) {
6580
+ messages.push({
6581
+ role: "assistant",
6582
+ content: fullText,
6583
+ metadata: { timestamp: now(), id: randomUUID3(), step }
6584
+ });
6585
+ }
6360
6586
  const result_ = {
6361
6587
  status: "completed",
6362
6588
  response: responseText + fullText,
@@ -6442,6 +6668,13 @@ ${textContent}` };
6442
6668
  `[poncho][harness] Model "${modelName}" returned an empty response with finishReason="stop" on step ${step}.`
6443
6669
  );
6444
6670
  }
6671
+ if (fullText.length > 0) {
6672
+ messages.push({
6673
+ role: "assistant",
6674
+ content: fullText,
6675
+ metadata: { timestamp: now(), id: randomUUID3(), step }
6676
+ });
6677
+ }
6445
6678
  responseText = fullText;
6446
6679
  yield pushEvent({
6447
6680
  type: "step:completed",
@@ -6459,7 +6692,8 @@ ${textContent}` };
6459
6692
  },
6460
6693
  duration: now() - start,
6461
6694
  contextTokens: latestContextTokens + toolOutputEstimateSinceModel,
6462
- contextWindow
6695
+ contextWindow,
6696
+ continuationMessages: [...messages]
6463
6697
  };
6464
6698
  yield pushEvent({ type: "run:completed", runId, result: result2 });
6465
6699
  return;
@@ -6555,6 +6789,27 @@ ${textContent}` };
6555
6789
  })
6556
6790
  );
6557
6791
  }
6792
+ } else if (this.insideTelemetryCapture && this.hasOtlpExporter) {
6793
+ const tracer = trace.getTracer("poncho");
6794
+ for (const call of approvedCalls) {
6795
+ const span = tracer.startSpan(`tool ${call.name}`, {
6796
+ attributes: {
6797
+ "poncho.tool.name": call.name,
6798
+ "poncho.tool.call_id": call.id,
6799
+ "poncho.tool.arguments": JSON.stringify(call.input)
6800
+ }
6801
+ });
6802
+ toolSpans.set(call.id, {
6803
+ end(opts) {
6804
+ if (opts.result.isError) {
6805
+ span.setStatus({ code: SpanStatusCode.ERROR, message: String(opts.result.value) });
6806
+ } else {
6807
+ span.setStatus({ code: SpanStatusCode.OK });
6808
+ }
6809
+ span.end();
6810
+ }
6811
+ });
6812
+ }
6558
6813
  }
6559
6814
  const TOOL_DEADLINE_SENTINEL = /* @__PURE__ */ Symbol("tool_deadline");
6560
6815
  const toolDeadlineRemainingMs = softDeadlineMs > 0 ? softDeadlineMs - (now() - start) : Infinity;
@@ -6579,6 +6834,13 @@ ${textContent}` };
6579
6834
  batchResults = await this.dispatcher.executeBatch(approvedCalls, toolContext);
6580
6835
  }
6581
6836
  if (batchResults === TOOL_DEADLINE_SENTINEL) {
6837
+ if (fullText.length > 0) {
6838
+ messages.push({
6839
+ role: "assistant",
6840
+ content: fullText,
6841
+ metadata: { timestamp: now(), id: randomUUID3(), step }
6842
+ });
6843
+ }
6582
6844
  const result_ = {
6583
6845
  status: "completed",
6584
6846
  response: responseText + fullText,
@@ -8066,73 +8328,6 @@ var createConversationStore = (config, options) => {
8066
8328
  return new InMemoryConversationStore(ttl);
8067
8329
  };
8068
8330
 
8069
- // src/telemetry.ts
8070
- var MAX_FIELD_LENGTH = 200;
8071
- function sanitizeEventForLog(event) {
8072
- return JSON.stringify(event, (_key, value) => {
8073
- if (typeof value === "string" && value.length > MAX_FIELD_LENGTH) {
8074
- return `${value.slice(0, 80)}...[${value.length} chars]`;
8075
- }
8076
- return value;
8077
- });
8078
- }
8079
- var TelemetryEmitter = class {
8080
- config;
8081
- constructor(config) {
8082
- this.config = config;
8083
- }
8084
- async emit(event) {
8085
- if (this.config?.enabled === false) {
8086
- return;
8087
- }
8088
- if (this.config?.handler) {
8089
- await this.config.handler(event);
8090
- return;
8091
- }
8092
- if (this.config?.otlp) {
8093
- await this.sendOtlp(event);
8094
- }
8095
- process.stdout.write(`[event] ${event.type} ${sanitizeEventForLog(event)}
8096
- `);
8097
- }
8098
- async sendOtlp(event) {
8099
- const endpoint = this.config?.otlp;
8100
- if (!endpoint) {
8101
- return;
8102
- }
8103
- try {
8104
- await fetch(endpoint, {
8105
- method: "POST",
8106
- headers: { "Content-Type": "application/json" },
8107
- body: JSON.stringify({
8108
- resourceLogs: [
8109
- {
8110
- scopeLogs: [
8111
- {
8112
- logRecords: [
8113
- {
8114
- timeUnixNano: String(Date.now() * 1e6),
8115
- severityText: "INFO",
8116
- body: { stringValue: event.type },
8117
- attributes: [
8118
- {
8119
- key: "event.payload",
8120
- value: { stringValue: JSON.stringify(event) }
8121
- }
8122
- ]
8123
- }
8124
- ]
8125
- }
8126
- ]
8127
- }
8128
- ]
8129
- })
8130
- });
8131
- } catch {
8132
- }
8133
- }
8134
- };
8135
-
8136
8331
  // src/index.ts
8137
8332
  import { defineTool as defineTool7 } from "@poncho-ai/sdk";
8138
8333
  export {
@@ -8180,6 +8375,7 @@ export {
8180
8375
  loadSkillContext,
8181
8376
  loadSkillInstructions,
8182
8377
  loadSkillMetadata,
8378
+ normalizeOtlp,
8183
8379
  normalizeScriptPolicyPath,
8184
8380
  parseAgentFile,
8185
8381
  parseAgentMarkdown,
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@poncho-ai/harness",
3
- "version": "0.28.3",
3
+ "version": "0.30.0",
4
4
  "description": "Agent execution runtime - conversation loop, tool dispatch, streaming",
5
5
  "repository": {
6
6
  "type": "git",
@@ -25,6 +25,8 @@
25
25
  "@aws-sdk/client-dynamodb": "^3.988.0",
26
26
  "@latitude-data/telemetry": "^2.0.4",
27
27
  "@opentelemetry/api": "1.9.0",
28
+ "@opentelemetry/exporter-trace-otlp-http": "^0.213.0",
29
+ "@opentelemetry/sdk-trace-node": "^2.6.0",
28
30
  "ai": "^6.0.86",
29
31
  "cheerio": "^1.2.0",
30
32
  "jiti": "^2.6.1",
@@ -32,7 +34,7 @@
32
34
  "redis": "^5.10.0",
33
35
  "yaml": "^2.4.0",
34
36
  "zod": "^3.22.0",
35
- "@poncho-ai/sdk": "1.6.1"
37
+ "@poncho-ai/sdk": "1.6.3"
36
38
  },
37
39
  "devDependencies": {
38
40
  "@types/mustache": "^4.2.6",
package/src/config.ts CHANGED
@@ -104,7 +104,10 @@ export interface PonchoConfig extends McpConfig {
104
104
  };
105
105
  telemetry?: {
106
106
  enabled?: boolean;
107
- otlp?: string;
107
+ otlp?: string | {
108
+ url: string;
109
+ headers?: Record<string, string>;
110
+ };
108
111
  latitude?: {
109
112
  apiKeyEnv?: string;
110
113
  projectIdEnv?: string;
package/src/harness.ts CHANGED
@@ -36,6 +36,10 @@ import { createSearchTools } from "./search-tools.js";
36
36
  import { createSubagentTools } from "./subagent-tools.js";
37
37
  import type { SubagentManager } from "./subagent-manager.js";
38
38
  import { LatitudeTelemetry } from "@latitude-data/telemetry";
39
+ import { trace, context as otelContext, SpanStatusCode } from "@opentelemetry/api";
40
+ import { NodeTracerProvider, BatchSpanProcessor } from "@opentelemetry/sdk-trace-node";
41
+ import { OTLPTraceExporter } from "@opentelemetry/exporter-trace-otlp-http";
42
+ import { normalizeOtlp } from "./telemetry.js";
39
43
  import {
40
44
  isSiblingScriptsPattern,
41
45
  matchesRelativeScriptPattern,
@@ -560,6 +564,9 @@ export class AgentHarness {
560
564
  private readonly activeSkillNames = new Set<string>();
561
565
  private readonly registeredMcpToolNames = new Set<string>();
562
566
  private latitudeTelemetry?: LatitudeTelemetry;
567
+ private otlpSpanProcessor?: BatchSpanProcessor;
568
+ private otlpTracerProvider?: NodeTracerProvider;
569
+ private hasOtlpExporter = false;
563
570
  private insideTelemetryCapture = false;
564
571
  private _browserSession?: unknown;
565
572
  private _browserMod?: {
@@ -1079,6 +1086,37 @@ export class AgentHarness {
1079
1086
  `[poncho][telemetry] Latitude telemetry is configured but missing: ${missing.join(", ")}. Traces will NOT be sent.`,
1080
1087
  );
1081
1088
  }
1089
+
1090
+ // Generic OTLP trace exporter — works alongside or instead of Latitude.
1091
+ const otlpConfig = telemetryEnabled ? normalizeOtlp(config?.telemetry?.otlp) : undefined;
1092
+ if (otlpConfig) {
1093
+ const exporter = new OTLPTraceExporter({
1094
+ url: otlpConfig.url,
1095
+ headers: otlpConfig.headers,
1096
+ });
1097
+ const processor = new BatchSpanProcessor(exporter);
1098
+ this.otlpSpanProcessor = processor;
1099
+
1100
+ if (this.latitudeTelemetry) {
1101
+ // Latitude already registered a global TracerProvider (v1.x) — add our
1102
+ // processor to it so every span flows to both destinations.
1103
+ const globalProvider = trace.getTracerProvider();
1104
+ const delegate = (globalProvider as unknown as { getDelegate?: () => unknown })
1105
+ .getDelegate?.() ?? globalProvider;
1106
+ if (typeof (delegate as Record<string, unknown>).addSpanProcessor === "function") {
1107
+ (delegate as unknown as { addSpanProcessor(p: BatchSpanProcessor): void }).addSpanProcessor(processor);
1108
+ }
1109
+ console.info(`[poncho][telemetry] OTLP exporter added (piggybacking on Latitude provider) → ${otlpConfig.url}`);
1110
+ } else {
1111
+ const provider = new NodeTracerProvider({
1112
+ spanProcessors: [processor],
1113
+ });
1114
+ provider.register();
1115
+ this.otlpTracerProvider = provider;
1116
+ console.info(`[poncho][telemetry] OTLP exporter active (standalone provider) → ${otlpConfig.url}`);
1117
+ }
1118
+ this.hasOtlpExporter = true;
1119
+ }
1082
1120
  }
1083
1121
 
1084
1122
  private async buildBrowserStoragePersistence(
@@ -1250,6 +1288,27 @@ export class AgentHarness {
1250
1288
  });
1251
1289
  this.latitudeTelemetry = undefined;
1252
1290
  }
1291
+ if (this.otlpSpanProcessor) {
1292
+ await this.otlpSpanProcessor.shutdown().catch((err) => {
1293
+ console.warn(
1294
+ `[poncho][telemetry] OTLP span processor shutdown error: ${
1295
+ err instanceof Error ? err.message : String(err)
1296
+ }`,
1297
+ );
1298
+ });
1299
+ this.otlpSpanProcessor = undefined;
1300
+ }
1301
+ if (this.otlpTracerProvider) {
1302
+ await this.otlpTracerProvider.shutdown().catch((err) => {
1303
+ console.warn(
1304
+ `[poncho][telemetry] OTLP tracer provider shutdown error: ${
1305
+ err instanceof Error ? err.message : String(err)
1306
+ }`,
1307
+ );
1308
+ });
1309
+ this.otlpTracerProvider = undefined;
1310
+ }
1311
+ this.hasOtlpExporter = false;
1253
1312
  }
1254
1313
 
1255
1314
  listTools(): ToolDefinition[] {
@@ -1257,18 +1316,20 @@ export class AgentHarness {
1257
1316
  }
1258
1317
 
1259
1318
  /**
1260
- * Wraps the run() generator with Latitude telemetry capture for complete trace coverage
1261
- * Streams events in real-time using an event queue pattern
1319
+ * Wraps the run() generator with telemetry capture for complete trace coverage.
1320
+ * Supports Latitude, generic OTLP, or both simultaneously.
1321
+ * Streams events in real-time using an event queue pattern.
1262
1322
  */
1263
1323
  async *runWithTelemetry(input: RunInput): AsyncGenerator<AgentEvent> {
1264
1324
  const config = this.loadedConfig;
1265
1325
  const telemetry = this.latitudeTelemetry;
1266
1326
 
1267
1327
  if (telemetry) {
1328
+ // Latitude capture path — wraps run() inside telemetry.capture().
1329
+ // If OTLP is also configured, spans flow to both via the shared provider.
1268
1330
  const latProjectIdEnv2 = config?.telemetry?.latitude?.projectIdEnv ?? "LATITUDE_PROJECT_ID";
1269
1331
  const projectId = parseInt(process.env[latProjectIdEnv2] ?? "", 10) as number;
1270
1332
  const rawPath = config?.telemetry?.latitude?.path ?? this.parsedAgent?.frontmatter.name ?? 'agent';
1271
- // Sanitize path for Latitude's DOCUMENT_PATH_REGEXP: /^([\w-]+\/)*([\w-.])+$/
1272
1333
  const path = rawPath.replace(/[^\w\-./]/g, '-').replace(/-+/g, '-').replace(/^-+|-+$/g, '') || 'agent';
1273
1334
 
1274
1335
  const rawConversationId = input.conversationId ?? (
@@ -1276,7 +1337,6 @@ export class AgentHarness {
1276
1337
  ? input.parameters.__activeConversationId
1277
1338
  : undefined
1278
1339
  );
1279
- // Latitude expects a UUID v4 for documentLogUuid; only pass it if valid
1280
1340
  const UUID_RE = /^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$/i;
1281
1341
  const conversationUuid = rawConversationId && UUID_RE.test(rawConversationId)
1282
1342
  ? rawConversationId
@@ -1286,13 +1346,11 @@ export class AgentHarness {
1286
1346
  `[poncho][telemetry] Latitude telemetry active – projectId=${projectId}, path="${path}"${conversationUuid ? `, conversation="${conversationUuid}"` : ""}`,
1287
1347
  );
1288
1348
 
1289
- // Event queue for streaming events in real-time
1290
1349
  const eventQueue: AgentEvent[] = [];
1291
1350
  let queueResolve: ((value: void) => void) | null = null;
1292
1351
  let generatorDone = false;
1293
1352
  let generatorError: Error | null = null;
1294
1353
 
1295
- // Start the generator inside telemetry.capture() (runs in background)
1296
1354
  const capturePromise = telemetry.capture({ projectId, path, conversationUuid }, async () => {
1297
1355
  this.insideTelemetryCapture = true;
1298
1356
  try {
@@ -1316,13 +1374,11 @@ export class AgentHarness {
1316
1374
  }
1317
1375
  });
1318
1376
 
1319
- // Yield events from the queue as they arrive
1320
1377
  try {
1321
1378
  while (!generatorDone || eventQueue.length > 0) {
1322
1379
  if (eventQueue.length > 0) {
1323
1380
  yield eventQueue.shift()!;
1324
1381
  } else if (!generatorDone) {
1325
- // Wait for next event
1326
1382
  await new Promise<void>((resolve) => {
1327
1383
  queueResolve = resolve;
1328
1384
  });
@@ -1344,8 +1400,47 @@ export class AgentHarness {
1344
1400
  }
1345
1401
  }
1346
1402
  }
1403
+ } else if (this.hasOtlpExporter) {
1404
+ // Standalone OTLP path — create a root span for the agent run so all
1405
+ // child spans (LLM calls via Vercel AI SDK, tool spans) are grouped
1406
+ // under a single trace.
1407
+ const tracer = trace.getTracer("poncho");
1408
+ const agentName = this.parsedAgent?.frontmatter.name ?? "agent";
1409
+
1410
+ const rootSpan = tracer.startSpan(`agent.run ${agentName}`);
1411
+ rootSpan.setAttribute("poncho.agent.name", agentName);
1412
+ if (input.conversationId) {
1413
+ rootSpan.setAttribute("poncho.conversation.id", input.conversationId);
1414
+ }
1415
+
1416
+ // Bind the root span's context so every async step (including
1417
+ // streamText and tool calls) sees it as the parent span.
1418
+ const spanContext = trace.setSpan(otelContext.active(), rootSpan);
1419
+ this.insideTelemetryCapture = true;
1420
+
1421
+ try {
1422
+ const gen = this.run(input);
1423
+ let next: IteratorResult<AgentEvent>;
1424
+ do {
1425
+ next = await otelContext.with(spanContext, () => gen.next());
1426
+ if (!next.done) yield next.value;
1427
+ } while (!next.done);
1428
+ rootSpan.setStatus({ code: SpanStatusCode.OK });
1429
+ } catch (error) {
1430
+ rootSpan.setStatus({
1431
+ code: SpanStatusCode.ERROR,
1432
+ message: error instanceof Error ? error.message : String(error),
1433
+ });
1434
+ rootSpan.recordException(error instanceof Error ? error : new Error(String(error)));
1435
+ throw error;
1436
+ } finally {
1437
+ this.insideTelemetryCapture = false;
1438
+ rootSpan.end();
1439
+ try {
1440
+ await this.otlpSpanProcessor?.forceFlush();
1441
+ } catch { /* best-effort */ }
1442
+ }
1347
1443
  } else {
1348
- // No telemetry configured, just pass through
1349
1444
  yield* this.run(input);
1350
1445
  }
1351
1446
  }
@@ -1548,7 +1643,7 @@ ${boundedMainMemory.trim()}`
1548
1643
  if (lastMsg && lastMsg.role !== "user") {
1549
1644
  messages.push({
1550
1645
  role: "user",
1551
- content: "[System: Your previous turn was interrupted by a time limit. Continue from where you left off — do NOT repeat what you already said. Proceed directly with the next action or tool call.]",
1646
+ content: "[System: Your previous turn was interrupted by a time limit. Your partial response above is already visible to the user. Continue EXACTLY from where you left off — do NOT restart, re-summarize, or repeat any content you already produced. If you were mid-sentence or mid-table, continue that sentence or table. Proceed directly with the next action or output.]",
1552
1647
  metadata: { timestamp: now(), id: randomUUID() },
1553
1648
  });
1554
1649
  }
@@ -1941,7 +2036,7 @@ ${boundedMainMemory.trim()}`
1941
2036
  abortSignal: input.abortSignal,
1942
2037
  ...(typeof maxTokens === "number" ? { maxTokens } : {}),
1943
2038
  experimental_telemetry: {
1944
- isEnabled: telemetryEnabled && !!this.latitudeTelemetry,
2039
+ isEnabled: telemetryEnabled && !!(this.latitudeTelemetry || this.hasOtlpExporter),
1945
2040
  recordInputs: true,
1946
2041
  recordOutputs: true,
1947
2042
  },
@@ -1953,7 +2048,10 @@ ${boundedMainMemory.trim()}`
1953
2048
  let chunkCount = 0;
1954
2049
  const hasRunTimeout = timeoutMs > 0;
1955
2050
  const streamDeadline = hasRunTimeout ? start + timeoutMs : 0;
2051
+ const hasSoftDeadline = softDeadlineMs > 0;
2052
+ const INTER_CHUNK_TIMEOUT_MS = 60_000;
1956
2053
  const fullStreamIterator = result.fullStream[Symbol.asyncIterator]();
2054
+ let softDeadlineFiredDuringStream = false;
1957
2055
  try {
1958
2056
  while (true) {
1959
2057
  if (isCancelled()) {
@@ -1977,25 +2075,36 @@ ${boundedMainMemory.trim()}`
1977
2075
  return;
1978
2076
  }
1979
2077
  }
1980
- const remaining = hasRunTimeout ? streamDeadline - now() : Infinity;
2078
+ if (hasSoftDeadline && chunkCount > 0 && now() - start >= softDeadlineMs) {
2079
+ softDeadlineFiredDuringStream = true;
2080
+ break;
2081
+ }
2082
+ const hardRemaining = hasRunTimeout ? streamDeadline - now() : Infinity;
2083
+ const softRemaining = hasSoftDeadline ? Math.max(0, (start + softDeadlineMs) - now()) : Infinity;
2084
+ const deadlineRemaining = Math.min(hardRemaining, softRemaining);
1981
2085
  const timeout = chunkCount === 0
1982
- ? Math.min(remaining, FIRST_CHUNK_TIMEOUT_MS)
1983
- : hasRunTimeout ? remaining : 0;
2086
+ ? Math.min(deadlineRemaining, FIRST_CHUNK_TIMEOUT_MS)
2087
+ : Math.min(deadlineRemaining, INTER_CHUNK_TIMEOUT_MS);
1984
2088
  let nextPart: IteratorResult<(typeof result.fullStream) extends AsyncIterable<infer T> ? T : never> | null;
1985
- if (timeout <= 0 && chunkCount > 0) {
2089
+ if (timeout <= 0 && chunkCount > 0 && !hasSoftDeadline) {
1986
2090
  nextPart = await fullStreamIterator.next();
1987
2091
  } else {
2092
+ const effectiveTimeout = Math.max(timeout, 1);
1988
2093
  let timer: ReturnType<typeof setTimeout> | undefined;
1989
2094
  nextPart = await Promise.race([
1990
2095
  fullStreamIterator.next(),
1991
2096
  new Promise<null>((resolve) => {
1992
- timer = setTimeout(() => resolve(null), timeout);
2097
+ timer = setTimeout(() => resolve(null), effectiveTimeout);
1993
2098
  }),
1994
2099
  ]);
1995
2100
  clearTimeout(timer);
1996
2101
  }
1997
2102
 
1998
2103
  if (nextPart === null) {
2104
+ if (hasSoftDeadline && deadlineRemaining <= INTER_CHUNK_TIMEOUT_MS) {
2105
+ softDeadlineFiredDuringStream = true;
2106
+ break;
2107
+ }
1999
2108
  const isFirstChunk = chunkCount === 0;
2000
2109
  console.error(
2001
2110
  `[poncho][harness] Stream timeout waiting for ${isFirstChunk ? "first" : "next"} chunk: model="${modelName}", step=${step}, chunks=${chunkCount}, elapsed=${now() - start}ms`,
@@ -2030,6 +2139,31 @@ ${boundedMainMemory.trim()}`
2030
2139
  fullStreamIterator.return?.(undefined)?.catch?.(() => {});
2031
2140
  }
2032
2141
 
2142
+ if (softDeadlineFiredDuringStream) {
2143
+ if (fullText.length > 0) {
2144
+ messages.push({
2145
+ role: "assistant",
2146
+ content: fullText,
2147
+ metadata: { timestamp: now(), id: randomUUID(), step },
2148
+ });
2149
+ }
2150
+ const result_: RunResult = {
2151
+ status: "completed",
2152
+ response: responseText + fullText,
2153
+ steps: step,
2154
+ tokens: { input: totalInputTokens, output: totalOutputTokens, cached: totalCachedTokens },
2155
+ duration: now() - start,
2156
+ continuation: true,
2157
+ continuationMessages: [...messages],
2158
+ maxSteps,
2159
+ contextTokens: latestContextTokens + toolOutputEstimateSinceModel,
2160
+ contextWindow,
2161
+ };
2162
+ console.info(`[poncho][harness] Soft deadline fired mid-stream at step ${step} (${(now() - start).toFixed(0)}ms). Checkpointing with ${fullText.length} chars of partial text.`);
2163
+ yield pushEvent({ type: "run:completed", runId, result: result_ });
2164
+ return;
2165
+ }
2166
+
2033
2167
  if (isCancelled()) {
2034
2168
  yield emitCancellation();
2035
2169
  return;
@@ -2038,6 +2172,13 @@ ${boundedMainMemory.trim()}`
2038
2172
  // Post-streaming soft deadline: if the model stream took long enough to
2039
2173
  // push past the soft deadline, checkpoint now before tool execution.
2040
2174
  if (softDeadlineMs > 0 && now() - start > softDeadlineMs) {
2175
+ if (fullText.length > 0) {
2176
+ messages.push({
2177
+ role: "assistant",
2178
+ content: fullText,
2179
+ metadata: { timestamp: now(), id: randomUUID(), step },
2180
+ });
2181
+ }
2041
2182
  const result_: RunResult = {
2042
2183
  status: "completed",
2043
2184
  response: responseText + fullText,
@@ -2138,6 +2279,13 @@ ${boundedMainMemory.trim()}`
2138
2279
  `[poncho][harness] Model "${modelName}" returned an empty response with finishReason="stop" on step ${step}.`,
2139
2280
  );
2140
2281
  }
2282
+ if (fullText.length > 0) {
2283
+ messages.push({
2284
+ role: "assistant",
2285
+ content: fullText,
2286
+ metadata: { timestamp: now(), id: randomUUID(), step },
2287
+ });
2288
+ }
2141
2289
  responseText = fullText;
2142
2290
  yield pushEvent({
2143
2291
  type: "step:completed",
@@ -2156,6 +2304,7 @@ ${boundedMainMemory.trim()}`
2156
2304
  duration: now() - start,
2157
2305
  contextTokens: latestContextTokens + toolOutputEstimateSinceModel,
2158
2306
  contextWindow,
2307
+ continuationMessages: [...messages],
2159
2308
  };
2160
2309
  yield pushEvent({ type: "run:completed", runId, result });
2161
2310
  return;
@@ -2275,7 +2424,7 @@ ${boundedMainMemory.trim()}`
2275
2424
  return;
2276
2425
  }
2277
2426
 
2278
- // Create telemetry tool spans so tool calls appear in Latitude traces
2427
+ // Create telemetry tool spans so tool calls appear in traces
2279
2428
  type ToolSpanHandle = { end: (opts: { result: { value: unknown; isError: boolean } }) => void };
2280
2429
  const toolSpans = new Map<string, ToolSpanHandle>();
2281
2430
  if (this.insideTelemetryCapture && this.latitudeTelemetry) {
@@ -2288,6 +2437,27 @@ ${boundedMainMemory.trim()}`
2288
2437
  }),
2289
2438
  );
2290
2439
  }
2440
+ } else if (this.insideTelemetryCapture && this.hasOtlpExporter) {
2441
+ const tracer = trace.getTracer("poncho");
2442
+ for (const call of approvedCalls) {
2443
+ const span = tracer.startSpan(`tool ${call.name}`, {
2444
+ attributes: {
2445
+ "poncho.tool.name": call.name,
2446
+ "poncho.tool.call_id": call.id,
2447
+ "poncho.tool.arguments": JSON.stringify(call.input),
2448
+ },
2449
+ });
2450
+ toolSpans.set(call.id, {
2451
+ end(opts: { result: { value: unknown; isError: boolean } }) {
2452
+ if (opts.result.isError) {
2453
+ span.setStatus({ code: SpanStatusCode.ERROR, message: String(opts.result.value) });
2454
+ } else {
2455
+ span.setStatus({ code: SpanStatusCode.OK });
2456
+ }
2457
+ span.end();
2458
+ },
2459
+ });
2460
+ }
2291
2461
  }
2292
2462
 
2293
2463
  // Race tool execution against the soft deadline so long-running tool
@@ -2322,6 +2492,13 @@ ${boundedMainMemory.trim()}`
2322
2492
  }
2323
2493
 
2324
2494
  if ((batchResults as unknown) === TOOL_DEADLINE_SENTINEL) {
2495
+ if (fullText.length > 0) {
2496
+ messages.push({
2497
+ role: "assistant",
2498
+ content: fullText,
2499
+ metadata: { timestamp: now(), id: randomUUID(), step },
2500
+ });
2501
+ }
2325
2502
  const result_: RunResult = {
2326
2503
  status: "completed",
2327
2504
  response: responseText + fullText,
package/src/state.ts CHANGED
@@ -71,6 +71,15 @@ export interface Conversation {
71
71
  /** Harness-internal message chain preserved across continuation runs.
72
72
  * Cleared when a run completes without continuation. */
73
73
  _continuationMessages?: Message[];
74
+ /** Number of continuation pickups for the current multi-step run.
75
+ * Reset when a run completes without continuation. Used to enforce
76
+ * a maximum continuation count across all entry points. */
77
+ _continuationCount?: number;
78
+ /** Full structured message chain from the last harness run, including
79
+ * tool-call and tool-result messages the model needs for context.
80
+ * Unlike `_continuationMessages`, this is always set after a run
81
+ * and does NOT signal that a continuation is pending. */
82
+ _harnessMessages?: Message[];
74
83
  createdAt: number;
75
84
  updatedAt: number;
76
85
  }
package/src/telemetry.ts CHANGED
@@ -11,9 +11,22 @@ function sanitizeEventForLog(event: AgentEvent): string {
11
11
  });
12
12
  }
13
13
 
14
+ export interface OtlpConfig {
15
+ url: string;
16
+ headers?: Record<string, string>;
17
+ }
18
+
19
+ export type OtlpOption = string | OtlpConfig;
20
+
21
+ export function normalizeOtlp(opt: OtlpOption | undefined): OtlpConfig | undefined {
22
+ if (!opt) return undefined;
23
+ if (typeof opt === "string") return opt ? { url: opt } : undefined;
24
+ return opt.url ? opt : undefined;
25
+ }
26
+
14
27
  export interface TelemetryConfig {
15
28
  enabled?: boolean;
16
- otlp?: string;
29
+ otlp?: OtlpOption;
17
30
  latitude?: {
18
31
  apiKeyEnv?: string;
19
32
  projectIdEnv?: string;
@@ -38,8 +51,9 @@ export class TelemetryEmitter {
38
51
  await this.config.handler(event);
39
52
  return;
40
53
  }
41
- if (this.config?.otlp) {
42
- await this.sendOtlp(event);
54
+ const otlp = normalizeOtlp(this.config?.otlp);
55
+ if (otlp) {
56
+ await this.sendOtlp(event, otlp);
43
57
  }
44
58
  // Latitude telemetry is handled by LatitudeTelemetry (from
45
59
  // @latitude-data/telemetry) via harness.runWithTelemetry().
@@ -48,15 +62,11 @@ export class TelemetryEmitter {
48
62
  process.stdout.write(`[event] ${event.type} ${sanitizeEventForLog(event)}\n`);
49
63
  }
50
64
 
51
- private async sendOtlp(event: AgentEvent): Promise<void> {
52
- const endpoint = this.config?.otlp;
53
- if (!endpoint) {
54
- return;
55
- }
65
+ private async sendOtlp(event: AgentEvent, otlp: OtlpConfig): Promise<void> {
56
66
  try {
57
- await fetch(endpoint, {
67
+ await fetch(otlp.url, {
58
68
  method: "POST",
59
- headers: { "Content-Type": "application/json" },
69
+ headers: { "Content-Type": "application/json", ...otlp.headers },
60
70
  body: JSON.stringify({
61
71
  resourceLogs: [
62
72
  {