@upstash/workflow 0.2.5-agents → 0.2.5-agents-2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/cloudflare.js CHANGED
@@ -826,29 +826,16 @@ var triggerWorkflowDelete = async (workflowContext, debug, cancel = false) => {
826
826
  await debug?.log("SUBMIT", "SUBMIT_CLEANUP", {
827
827
  deletedWorkflowRunId: workflowContext.workflowRunId
828
828
  });
829
- try {
830
- await workflowContext.qstashClient.http.request({
831
- path: ["v2", "workflows", "runs", `${workflowContext.workflowRunId}?cancel=${cancel}`],
832
- method: "DELETE",
833
- parseResponseAsJson: false
834
- });
835
- await debug?.log(
836
- "SUBMIT",
837
- "SUBMIT_CLEANUP",
838
- `workflow run ${workflowContext.workflowRunId} deleted.`
839
- );
840
- return { deleted: true };
841
- } catch (error) {
842
- if (error instanceof import_qstash3.QstashError && error.status === 404) {
843
- await debug?.log("WARN", "SUBMIT_CLEANUP", {
844
- message: `Failed to remove workflow run ${workflowContext.workflowRunId} as it doesn't exist.`,
845
- name: error.name,
846
- errorMessage: error.message
847
- });
848
- return { deleted: false };
849
- }
850
- throw error;
851
- }
829
+ await workflowContext.qstashClient.http.request({
830
+ path: ["v2", "workflows", "runs", `${workflowContext.workflowRunId}?cancel=${cancel}`],
831
+ method: "DELETE",
832
+ parseResponseAsJson: false
833
+ });
834
+ await debug?.log(
835
+ "SUBMIT",
836
+ "SUBMIT_CLEANUP",
837
+ `workflow run ${workflowContext.workflowRunId} deleted.`
838
+ );
852
839
  };
853
840
  var recreateUserHeaders = (headers) => {
854
841
  const filteredHeaders = new Headers();
@@ -1635,7 +1622,22 @@ var WorkflowApi = class extends BaseWorkflowApi {
1635
1622
  // src/agents/adapters.ts
1636
1623
  var import_openai2 = require("@ai-sdk/openai");
1637
1624
  var import_ai = require("ai");
1625
+
1626
+ // src/agents/constants.ts
1638
1627
  var AGENT_NAME_HEADER = "upstash-agent-name";
1628
+ var MANAGER_AGENT_PROMPT = `You are an agent orchestrating other AI Agents.
1629
+
1630
+ These other agents have tools available to them.
1631
+
1632
+ Given a prompt, utilize these agents to address requests.
1633
+
1634
+ Don't always call all the agents provided to you at the same time. You can call one and use it's response to call another.
1635
+
1636
+ Avoid calling the same agent twice in one turn. Instead, prefer to call it once but provide everything
1637
+ you need from that agent.
1638
+ `;
1639
+
1640
+ // src/agents/adapters.ts
1639
1641
  var createWorkflowOpenAI = (context) => {
1640
1642
  return (0, import_openai2.createOpenAI)({
1641
1643
  compatibility: "strict",
@@ -1702,8 +1704,7 @@ var convertLangchainTool = (langchainTool) => {
1702
1704
  return (0, import_ai.tool)({
1703
1705
  description: langchainTool.description,
1704
1706
  parameters: langchainTool.schema,
1705
- // eslint-disable-next-line @typescript-eslint/no-explicit-any
1706
- execute: async (param) => langchainTool.invoke(param)
1707
+ execute: async (...param) => langchainTool.invoke(...param)
1707
1708
  });
1708
1709
  };
1709
1710
 
@@ -1716,16 +1717,24 @@ var Agent = class {
1716
1717
  maxSteps;
1717
1718
  background;
1718
1719
  model;
1719
- constructor({ tools, maxSteps, background, name, model }) {
1720
+ temparature;
1721
+ constructor({ tools, maxSteps, background, name, model, temparature = 0.1 }) {
1720
1722
  this.name = name;
1721
1723
  this.tools = tools ?? {};
1722
1724
  this.maxSteps = maxSteps;
1723
1725
  this.background = background;
1724
1726
  this.model = model;
1727
+ this.temparature = temparature;
1725
1728
  }
1729
+ /**
1730
+ * Trigger the agent by passing a prompt
1731
+ *
1732
+ * @param prompt task to assign to the agent
1733
+ * @returns Response as `{ text: string }`
1734
+ */
1726
1735
  async call({ prompt }) {
1727
1736
  try {
1728
- return await (0, import_ai2.generateText)({
1737
+ const result = await (0, import_ai2.generateText)({
1729
1738
  model: this.model,
1730
1739
  tools: this.tools,
1731
1740
  maxSteps: this.maxSteps,
@@ -1733,8 +1742,10 @@ var Agent = class {
1733
1742
  prompt,
1734
1743
  headers: {
1735
1744
  [AGENT_NAME_HEADER]: this.name
1736
- }
1745
+ },
1746
+ temperature: this.temparature
1737
1747
  });
1748
+ return { text: result.text };
1738
1749
  } catch (error) {
1739
1750
  if (error instanceof import_ai2.ToolExecutionError) {
1740
1751
  if (error.cause instanceof Error && error.cause.name === "WorkflowAbort") {
@@ -1749,6 +1760,11 @@ var Agent = class {
1749
1760
  }
1750
1761
  }
1751
1762
  }
1763
+ /**
1764
+ * Convert the agent to a tool which can be used by other agents.
1765
+ *
1766
+ * @returns the agent as a tool
1767
+ */
1752
1768
  asTool() {
1753
1769
  const toolDescriptions = Object.values(this.tools).map((tool3) => tool3.description).join("\n");
1754
1770
  return (0, import_ai2.tool)({
@@ -1760,18 +1776,25 @@ var Agent = class {
1760
1776
  });
1761
1777
  }
1762
1778
  };
1763
- var MANAGER_AGENT_PROMPT = `You are an AI agent who orchestrates other AI Agents.
1764
- These other agents have tools available to them.
1765
- Given a prompt, utilize these agents to address requests.
1766
- Don't always call all the agents provided to you at the same time. You can call one and use it's response to call another.
1767
- `;
1768
1779
  var ManagerAgent = class extends Agent {
1769
1780
  agents;
1781
+ /**
1782
+ * A manager agent which coordinates agents available to it to achieve a
1783
+ * given task
1784
+ *
1785
+ * @param name Name of the agent
1786
+ * @param background Background of the agent. If not passed, default will be used.
1787
+ * @param model LLM model to use
1788
+ * @param agents: List of agents available to the agent
1789
+ * @param maxSteps number of times the manager agent can call the LLM at most.
1790
+ * If the agent abruptly stops execution after calling other agents, you may
1791
+ * need to increase maxSteps
1792
+ */
1770
1793
  constructor({
1771
- maxSteps,
1772
- background = MANAGER_AGENT_PROMPT,
1773
1794
  agents,
1795
+ background = MANAGER_AGENT_PROMPT,
1774
1796
  model,
1797
+ maxSteps,
1775
1798
  name = "manager llm"
1776
1799
  }) {
1777
1800
  super({
@@ -1796,6 +1819,11 @@ var Task = class {
1796
1819
  this.context = context;
1797
1820
  this.taskParameters = taskParameters;
1798
1821
  }
1822
+ /**
1823
+ * Run the agents to complete the task
1824
+ *
1825
+ * @returns Result of the task as { text: string }
1826
+ */
1799
1827
  async run() {
1800
1828
  const { prompt, ...otherParams } = this.taskParameters;
1801
1829
  const safePrompt = await this.context.run("Get Prompt", () => prompt);
@@ -1826,6 +1854,29 @@ var WorkflowAgents = class {
1826
1854
  constructor({ context }) {
1827
1855
  this.context = context;
1828
1856
  }
1857
+ /**
1858
+ * Defines an agent
1859
+ *
1860
+ * ```ts
1861
+ * const researcherAgent = context.agents.agent({
1862
+ * model,
1863
+ * name: 'academic',
1864
+ * maxSteps: 2,
1865
+ * tools: {
1866
+ * wikiTool: new WikipediaQueryRun({
1867
+ * topKResults: 1,
1868
+ * maxDocContentLength: 500,
1869
+ * })
1870
+ * },
1871
+ * background:
1872
+ * 'You are researcher agent with access to Wikipedia. ' +
1873
+ * 'Utilize Wikipedia as much as possible for correct information',
1874
+ * });
1875
+ * ```
1876
+ *
1877
+ * @param params agent parameters
1878
+ * @returns
1879
+ */
1829
1880
  agent(params) {
1830
1881
  const wrappedTools = wrapTools({ context: this.context, tools: params.tools });
1831
1882
  return new Agent({
@@ -1836,6 +1887,9 @@ var WorkflowAgents = class {
1836
1887
  task(taskParameters) {
1837
1888
  return new Task({ context: this.context, taskParameters });
1838
1889
  }
1890
+ /**
1891
+ * creates an openai model for agents
1892
+ */
1839
1893
  openai(...params) {
1840
1894
  const openai2 = createWorkflowOpenAI(this.context);
1841
1895
  return openai2(...params);
@@ -2436,7 +2490,6 @@ var checkIfLastOneIsDuplicate = async (steps, debug) => {
2436
2490
  if (step.stepId === lastStepId && step.targetStep === lastTargetStepId) {
2437
2491
  const message = `Upstash Workflow: The step '${step.stepName}' with id '${step.stepId}' has run twice during workflow execution. Rest of the workflow will continue running as usual.`;
2438
2492
  await debug?.log("WARN", "RESPONSE_DEFAULT", message);
2439
- console.log(steps);
2440
2493
  console.warn(message);
2441
2494
  return true;
2442
2495
  }
package/cloudflare.mjs CHANGED
@@ -1,8 +1,7 @@
1
1
  import {
2
2
  SDK_TELEMETRY,
3
3
  serveBase
4
- } from "./chunk-RFX5YRRT.mjs";
5
- import "./chunk-PU5J4TNC.mjs";
4
+ } from "./chunk-VOM3CFYZ.mjs";
6
5
 
7
6
  // platforms/cloudflare.ts
8
7
  var getArgs = (args) => {
package/express.d.mts CHANGED
@@ -1,9 +1,8 @@
1
- import { R as RouteFunction, W as WorkflowServeOptions } from './types-BEyIoCRe.mjs';
1
+ import { R as RouteFunction, W as WorkflowServeOptions } from './types-D9gwTj2n.mjs';
2
2
  import { Router } from 'express';
3
3
  import '@upstash/qstash';
4
4
  import 'ai';
5
5
  import '@ai-sdk/openai';
6
- import 'langchain/tools';
7
6
 
8
7
  declare function serve<TInitialPayload = unknown>(routeFunction: RouteFunction<TInitialPayload>, options?: Omit<WorkflowServeOptions<globalThis.Response, TInitialPayload>, "onStepFinish">): Router;
9
8
 
package/express.d.ts CHANGED
@@ -1,9 +1,8 @@
1
- import { R as RouteFunction, W as WorkflowServeOptions } from './types-BEyIoCRe.js';
1
+ import { R as RouteFunction, W as WorkflowServeOptions } from './types-D9gwTj2n.js';
2
2
  import { Router } from 'express';
3
3
  import '@upstash/qstash';
4
4
  import 'ai';
5
5
  import '@ai-sdk/openai';
6
- import 'langchain/tools';
7
6
 
8
7
  declare function serve<TInitialPayload = unknown>(routeFunction: RouteFunction<TInitialPayload>, options?: Omit<WorkflowServeOptions<globalThis.Response, TInitialPayload>, "onStepFinish">): Router;
9
8
 
package/express.js CHANGED
@@ -24498,29 +24498,16 @@ var triggerWorkflowDelete = async (workflowContext, debug, cancel = false) => {
24498
24498
  await debug?.log("SUBMIT", "SUBMIT_CLEANUP", {
24499
24499
  deletedWorkflowRunId: workflowContext.workflowRunId
24500
24500
  });
24501
- try {
24502
- await workflowContext.qstashClient.http.request({
24503
- path: ["v2", "workflows", "runs", `${workflowContext.workflowRunId}?cancel=${cancel}`],
24504
- method: "DELETE",
24505
- parseResponseAsJson: false
24506
- });
24507
- await debug?.log(
24508
- "SUBMIT",
24509
- "SUBMIT_CLEANUP",
24510
- `workflow run ${workflowContext.workflowRunId} deleted.`
24511
- );
24512
- return { deleted: true };
24513
- } catch (error) {
24514
- if (error instanceof import_qstash3.QstashError && error.status === 404) {
24515
- await debug?.log("WARN", "SUBMIT_CLEANUP", {
24516
- message: `Failed to remove workflow run ${workflowContext.workflowRunId} as it doesn't exist.`,
24517
- name: error.name,
24518
- errorMessage: error.message
24519
- });
24520
- return { deleted: false };
24521
- }
24522
- throw error;
24523
- }
24501
+ await workflowContext.qstashClient.http.request({
24502
+ path: ["v2", "workflows", "runs", `${workflowContext.workflowRunId}?cancel=${cancel}`],
24503
+ method: "DELETE",
24504
+ parseResponseAsJson: false
24505
+ });
24506
+ await debug?.log(
24507
+ "SUBMIT",
24508
+ "SUBMIT_CLEANUP",
24509
+ `workflow run ${workflowContext.workflowRunId} deleted.`
24510
+ );
24524
24511
  };
24525
24512
  var recreateUserHeaders = (headers) => {
24526
24513
  const filteredHeaders = new Headers();
@@ -25307,7 +25294,22 @@ var WorkflowApi = class extends BaseWorkflowApi {
25307
25294
  // src/agents/adapters.ts
25308
25295
  var import_openai2 = require("@ai-sdk/openai");
25309
25296
  var import_ai = require("ai");
25297
+
25298
+ // src/agents/constants.ts
25310
25299
  var AGENT_NAME_HEADER = "upstash-agent-name";
25300
+ var MANAGER_AGENT_PROMPT = `You are an agent orchestrating other AI Agents.
25301
+
25302
+ These other agents have tools available to them.
25303
+
25304
+ Given a prompt, utilize these agents to address requests.
25305
+
25306
+ Don't always call all the agents provided to you at the same time. You can call one and use it's response to call another.
25307
+
25308
+ Avoid calling the same agent twice in one turn. Instead, prefer to call it once but provide everything
25309
+ you need from that agent.
25310
+ `;
25311
+
25312
+ // src/agents/adapters.ts
25311
25313
  var createWorkflowOpenAI = (context) => {
25312
25314
  return (0, import_openai2.createOpenAI)({
25313
25315
  compatibility: "strict",
@@ -25374,8 +25376,7 @@ var convertLangchainTool = (langchainTool) => {
25374
25376
  return (0, import_ai.tool)({
25375
25377
  description: langchainTool.description,
25376
25378
  parameters: langchainTool.schema,
25377
- // eslint-disable-next-line @typescript-eslint/no-explicit-any
25378
- execute: async (param) => langchainTool.invoke(param)
25379
+ execute: async (...param) => langchainTool.invoke(...param)
25379
25380
  });
25380
25381
  };
25381
25382
 
@@ -25388,16 +25389,24 @@ var Agent = class {
25388
25389
  maxSteps;
25389
25390
  background;
25390
25391
  model;
25391
- constructor({ tools, maxSteps, background, name, model }) {
25392
+ temparature;
25393
+ constructor({ tools, maxSteps, background, name, model, temparature = 0.1 }) {
25392
25394
  this.name = name;
25393
25395
  this.tools = tools ?? {};
25394
25396
  this.maxSteps = maxSteps;
25395
25397
  this.background = background;
25396
25398
  this.model = model;
25399
+ this.temparature = temparature;
25397
25400
  }
25401
+ /**
25402
+ * Trigger the agent by passing a prompt
25403
+ *
25404
+ * @param prompt task to assign to the agent
25405
+ * @returns Response as `{ text: string }`
25406
+ */
25398
25407
  async call({ prompt }) {
25399
25408
  try {
25400
- return await (0, import_ai2.generateText)({
25409
+ const result = await (0, import_ai2.generateText)({
25401
25410
  model: this.model,
25402
25411
  tools: this.tools,
25403
25412
  maxSteps: this.maxSteps,
@@ -25405,8 +25414,10 @@ var Agent = class {
25405
25414
  prompt,
25406
25415
  headers: {
25407
25416
  [AGENT_NAME_HEADER]: this.name
25408
- }
25417
+ },
25418
+ temperature: this.temparature
25409
25419
  });
25420
+ return { text: result.text };
25410
25421
  } catch (error) {
25411
25422
  if (error instanceof import_ai2.ToolExecutionError) {
25412
25423
  if (error.cause instanceof Error && error.cause.name === "WorkflowAbort") {
@@ -25421,6 +25432,11 @@ var Agent = class {
25421
25432
  }
25422
25433
  }
25423
25434
  }
25435
+ /**
25436
+ * Convert the agent to a tool which can be used by other agents.
25437
+ *
25438
+ * @returns the agent as a tool
25439
+ */
25424
25440
  asTool() {
25425
25441
  const toolDescriptions = Object.values(this.tools).map((tool3) => tool3.description).join("\n");
25426
25442
  return (0, import_ai2.tool)({
@@ -25432,18 +25448,25 @@ var Agent = class {
25432
25448
  });
25433
25449
  }
25434
25450
  };
25435
- var MANAGER_AGENT_PROMPT = `You are an AI agent who orchestrates other AI Agents.
25436
- These other agents have tools available to them.
25437
- Given a prompt, utilize these agents to address requests.
25438
- Don't always call all the agents provided to you at the same time. You can call one and use it's response to call another.
25439
- `;
25440
25451
  var ManagerAgent = class extends Agent {
25441
25452
  agents;
25453
+ /**
25454
+ * A manager agent which coordinates agents available to it to achieve a
25455
+ * given task
25456
+ *
25457
+ * @param name Name of the agent
25458
+ * @param background Background of the agent. If not passed, default will be used.
25459
+ * @param model LLM model to use
25460
+ * @param agents: List of agents available to the agent
25461
+ * @param maxSteps number of times the manager agent can call the LLM at most.
25462
+ * If the agent abruptly stops execution after calling other agents, you may
25463
+ * need to increase maxSteps
25464
+ */
25442
25465
  constructor({
25443
- maxSteps,
25444
- background = MANAGER_AGENT_PROMPT,
25445
25466
  agents,
25467
+ background = MANAGER_AGENT_PROMPT,
25446
25468
  model,
25469
+ maxSteps,
25447
25470
  name = "manager llm"
25448
25471
  }) {
25449
25472
  super({
@@ -25468,6 +25491,11 @@ var Task = class {
25468
25491
  this.context = context;
25469
25492
  this.taskParameters = taskParameters;
25470
25493
  }
25494
+ /**
25495
+ * Run the agents to complete the task
25496
+ *
25497
+ * @returns Result of the task as { text: string }
25498
+ */
25471
25499
  async run() {
25472
25500
  const { prompt, ...otherParams } = this.taskParameters;
25473
25501
  const safePrompt = await this.context.run("Get Prompt", () => prompt);
@@ -25498,6 +25526,29 @@ var WorkflowAgents = class {
25498
25526
  constructor({ context }) {
25499
25527
  this.context = context;
25500
25528
  }
25529
+ /**
25530
+ * Defines an agent
25531
+ *
25532
+ * ```ts
25533
+ * const researcherAgent = context.agents.agent({
25534
+ * model,
25535
+ * name: 'academic',
25536
+ * maxSteps: 2,
25537
+ * tools: {
25538
+ * wikiTool: new WikipediaQueryRun({
25539
+ * topKResults: 1,
25540
+ * maxDocContentLength: 500,
25541
+ * })
25542
+ * },
25543
+ * background:
25544
+ * 'You are researcher agent with access to Wikipedia. ' +
25545
+ * 'Utilize Wikipedia as much as possible for correct information',
25546
+ * });
25547
+ * ```
25548
+ *
25549
+ * @param params agent parameters
25550
+ * @returns
25551
+ */
25501
25552
  agent(params) {
25502
25553
  const wrappedTools = wrapTools({ context: this.context, tools: params.tools });
25503
25554
  return new Agent({
@@ -25508,6 +25559,9 @@ var WorkflowAgents = class {
25508
25559
  task(taskParameters) {
25509
25560
  return new Task({ context: this.context, taskParameters });
25510
25561
  }
25562
+ /**
25563
+ * creates an openai model for agents
25564
+ */
25511
25565
  openai(...params) {
25512
25566
  const openai2 = createWorkflowOpenAI(this.context);
25513
25567
  return openai2(...params);
@@ -26108,7 +26162,6 @@ var checkIfLastOneIsDuplicate = async (steps, debug) => {
26108
26162
  if (step.stepId === lastStepId && step.targetStep === lastTargetStepId) {
26109
26163
  const message = `Upstash Workflow: The step '${step.stepName}' with id '${step.stepId}' has run twice during workflow execution. Rest of the workflow will continue running as usual.`;
26110
26164
  await debug?.log("WARN", "RESPONSE_DEFAULT", message);
26111
- console.log(steps);
26112
26165
  console.warn(message);
26113
26166
  return true;
26114
26167
  }
package/express.mjs CHANGED
@@ -1,12 +1,10 @@
1
1
  import {
2
2
  SDK_TELEMETRY,
3
- serveBase
4
- } from "./chunk-RFX5YRRT.mjs";
5
- import {
6
3
  __commonJS,
7
4
  __require,
8
- __toESM
9
- } from "./chunk-PU5J4TNC.mjs";
5
+ __toESM,
6
+ serveBase
7
+ } from "./chunk-VOM3CFYZ.mjs";
10
8
 
11
9
  // node_modules/depd/index.js
12
10
  var require_depd = __commonJS({
package/h3.d.mts CHANGED
@@ -1,9 +1,8 @@
1
1
  import * as h3 from 'h3';
2
- import { R as RouteFunction, j as PublicServeOptions } from './types-BEyIoCRe.mjs';
2
+ import { R as RouteFunction, j as PublicServeOptions } from './types-D9gwTj2n.mjs';
3
3
  import '@upstash/qstash';
4
4
  import 'ai';
5
5
  import '@ai-sdk/openai';
6
- import 'langchain/tools';
7
6
 
8
7
  declare const serve: <TInitialPayload = unknown>(routeFunction: RouteFunction<TInitialPayload>, options?: PublicServeOptions<TInitialPayload>) => {
9
8
  handler: h3.EventHandler<h3.EventHandlerRequest, Promise<Response | {
package/h3.d.ts CHANGED
@@ -1,9 +1,8 @@
1
1
  import * as h3 from 'h3';
2
- import { R as RouteFunction, j as PublicServeOptions } from './types-BEyIoCRe.js';
2
+ import { R as RouteFunction, j as PublicServeOptions } from './types-D9gwTj2n.js';
3
3
  import '@upstash/qstash';
4
4
  import 'ai';
5
5
  import '@ai-sdk/openai';
6
- import 'langchain/tools';
7
6
 
8
7
  declare const serve: <TInitialPayload = unknown>(routeFunction: RouteFunction<TInitialPayload>, options?: PublicServeOptions<TInitialPayload>) => {
9
8
  handler: h3.EventHandler<h3.EventHandlerRequest, Promise<Response | {