promptlayer 1.2.1 → 1.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/claude-agents.js +1 -1
- package/dist/claude-agents.js.map +1 -1
- package/dist/esm/{chunk-UKSCOWKT.js → chunk-KDVM55TG.js} +2 -2
- package/dist/esm/{chunk-UKSCOWKT.js.map → chunk-KDVM55TG.js.map} +1 -1
- package/dist/esm/claude-agents.js +1 -1
- package/dist/esm/claude-agents.js.map +1 -1
- package/dist/esm/index.js +2 -2
- package/dist/esm/index.js.map +1 -1
- package/dist/esm/openai-agents.js +1 -1
- package/dist/esm/openai-agents.js.map +1 -1
- package/dist/index.d.mts +98 -1
- package/dist/index.d.ts +98 -1
- package/dist/index.js +2 -2
- package/dist/index.js.map +1 -1
- package/dist/openai-agents.js +1 -1
- package/dist/openai-agents.js.map +1 -1
- package/package.json +1 -1
- package/vendor/claude-agents/trace/hooks/lib.sh +4 -522
- package/vendor/claude-agents/trace/hooks/post_tool_use.sh +2 -27
- package/vendor/claude-agents/trace/hooks/py/__init__.py +1 -0
- package/vendor/claude-agents/trace/hooks/py/cli.py +81 -0
- package/vendor/claude-agents/trace/hooks/py/context.py +63 -0
- package/vendor/claude-agents/trace/hooks/py/handlers.py +244 -0
- package/vendor/claude-agents/trace/hooks/py/otlp.py +278 -0
- package/vendor/claude-agents/trace/hooks/py/settings.py +33 -0
- package/vendor/claude-agents/trace/hooks/py/state.py +135 -0
- package/vendor/claude-agents/trace/hooks/{parse_stop_transcript.py → py/stop_parser.py} +69 -31
- package/vendor/claude-agents/trace/hooks/py/traceparent.py +31 -0
- package/vendor/claude-agents/trace/hooks/session_end.sh +1 -23
- package/vendor/claude-agents/trace/hooks/session_start.sh +5 -41
- package/vendor/claude-agents/trace/hooks/stop_hook.sh +3 -106
- package/vendor/claude-agents/trace/hooks/user_prompt_submit.sh +1 -11
- package/vendor/claude-agents/trace/setup.sh +170 -0
- package/vendor/claude-agents/vendor_metadata.json +2 -2
- package/vendor/claude-agents/trace/hooks/hook_utils.py +0 -38
package/dist/claude-agents.js
CHANGED
|
@@ -1,2 +1,2 @@
|
|
|
1
|
-
"use strict";var h=Object.create;var r=Object.defineProperty;var x=Object.getOwnPropertyDescriptor;var b=Object.getOwnPropertyNames;var R=Object.getPrototypeOf,A=Object.prototype.hasOwnProperty;var P=(t,e)=>{for(var n in e)r(t,n,{get:e[n],enumerable:!0})},d=(t,e,n,s)=>{if(e&&typeof e=="object"||typeof e=="function")for(let o of b(e))!A.call(t,o)&&o!==n&&r(t,o,{get:()=>e[o],enumerable:!(s=x(e,o))||s.enumerable});return t};var l=(t,e,n)=>(n=t!=null?h(R(t)):{},d(e||!t||!t.__esModule?r(n,"default",{value:t,enumerable:!0}):n,t)),w=t=>d(r({},"__esModule",{value:!0}),t);var K={};P(K,{getClaudeConfig:()=>y});module.exports=w(K);var c=l(require("fs")),a=l(require("path")),g=require("url");var I=l(require("ably")),O=require("centrifuge"),S=l(require("p-retry"));var N="1.
|
|
1
|
+
"use strict";var h=Object.create;var r=Object.defineProperty;var x=Object.getOwnPropertyDescriptor;var b=Object.getOwnPropertyNames;var R=Object.getPrototypeOf,A=Object.prototype.hasOwnProperty;var P=(t,e)=>{for(var n in e)r(t,n,{get:e[n],enumerable:!0})},d=(t,e,n,s)=>{if(e&&typeof e=="object"||typeof e=="function")for(let o of b(e))!A.call(t,o)&&o!==n&&r(t,o,{get:()=>e[o],enumerable:!(s=x(e,o))||s.enumerable});return t};var l=(t,e,n)=>(n=t!=null?h(R(t)):{},d(e||!t||!t.__esModule?r(n,"default",{value:t,enumerable:!0}):n,t)),w=t=>d(r({},"__esModule",{value:!0}),t);var K={};P(K,{getClaudeConfig:()=>y});module.exports=w(K);var c=l(require("fs")),a=l(require("path")),g=require("url");var I=l(require("ably")),O=require("centrifuge"),S=l(require("p-retry"));var N="1.3.0",q=()=>{var t;if(typeof process!="undefined"&&((t=process.versions)!=null&&t.node)){let e=process.versions.node.split(".");return`${e[0]}.${e[1]}`}return"unknown"},L=q(),it=`promptlayer-js/${N} (node ${L})`;var M=t=>{var e,n,s,o,i,u;if(typeof globalThis.process!="undefined")return(s=(n=(e=globalThis.process.env)==null?void 0:e[t])==null?void 0:n.trim())!=null?s:void 0;if(typeof globalThis.Deno!="undefined")return(u=(i=(o=globalThis.Deno.env)==null?void 0:o.get)==null?void 0:i.call(o,t))==null?void 0:u.trim()},f=t=>{if(t!==void 0){let n=t.trim();if(!n)throw new Error("PromptLayer API key not provided. Please set PROMPTLAYER_API_KEY or pass apiKey.");return n}let e=M("PROMPTLAYER_API_KEY");if(!e)throw new Error("PromptLayer API key not provided. Please set PROMPTLAYER_API_KEY or pass apiKey.");return e};var W=[".claude-plugin/plugin.json","setup.sh","hooks/hooks.json","hooks/lib.sh","hooks/session_start.sh","hooks/user_prompt_submit.sh","hooks/post_tool_use.sh","hooks/stop_hook.sh","hooks/session_end.sh","hooks/py/__init__.py","hooks/py/cli.py","hooks/py/context.py","hooks/py/handlers.py","hooks/py/otlp.py","hooks/py/settings.py","hooks/py/state.py","hooks/py/stop_parser.py","hooks/py/traceparent.py"],p=a.default.join("vendor","claude-agents","trace");function j(){let t=Error.prepareStackTrace;try{Error.prepareStackTrace=(s,o)=>o;let e=new Error().stack,n=e==null?void 0:e.map(s=>s.getFileName()).find(s=>s&&!s.startsWith("node:"));if(!n)throw new Error("PromptLayer Claude Agents could not determine its current module path.");return a.default.dirname(n.startsWith("file://")?(0,g.fileURLToPath)(n):n)}finally{Error.prepareStackTrace=t}}function $(){if(process.platform==="win32")throw new Error("PromptLayer Claude Agents integration does not support Windows. Use Linux or macOS.")}function G(){let t=j();return[a.default.resolve(t,"..","..","..",p),a.default.resolve(t,"..",p),a.default.resolve(t,"..","..",p)]}function U(t){let e=W.filter(n=>!c.default.existsSync(a.default.join(t,n)));if(e.length>0)throw new Error(`PromptLayer Claude Agents vendored plugin is incomplete. Missing: ${e.join(", ")}`)}function F(){for(let t of G())if(!(!c.default.existsSync(t)||!c.default.statSync(t).isDirectory()))return U(t),t;throw new Error("PromptLayer Claude Agents vendored plugin was not found in the installed package.")}function y(t={}){var i;$();let e=F(),n=f(t.apiKey),s=(i=t.traceparent)==null?void 0:i.trim(),o={TRACE_TO_PROMPTLAYER:"true",PROMPTLAYER_API_KEY:n};return s&&(o.PROMPTLAYER_TRACEPARENT=s),{plugin:{type:"local",path:e},env:o}}0&&(module.exports={getClaudeConfig});
|
|
2
2
|
//# sourceMappingURL=claude-agents.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/claude-agents.ts","../src/integrations/claude-agents/config.ts","../src/utils/utils.ts"],"sourcesContent":["export * from \"@/integrations/claude-agents\";\n","import fs from \"fs\";\nimport path from \"path\";\nimport { fileURLToPath } from \"url\";\nimport { requirePromptLayerApiKey } from \"@/utils/utils\";\n\nexport interface GetClaudeConfigOptions {\n apiKey?: string;\n traceparent?: string;\n}\n\nexport interface PromptLayerClaudeAgentsPlugin {\n type: \"local\";\n path: string;\n}\n\nexport interface PromptLayerClaudeAgentsEnv {\n TRACE_TO_PROMPTLAYER: \"true\";\n PROMPTLAYER_API_KEY: string;\n PROMPTLAYER_TRACEPARENT?: string;\n}\n\nexport interface PromptLayerClaudeAgentsConfig {\n plugin: PromptLayerClaudeAgentsPlugin;\n env: PromptLayerClaudeAgentsEnv;\n}\n\nconst REQUIRED_PLUGIN_FILES = [\n \".claude-plugin/plugin.json\",\n \"hooks/hooks.json\",\n \"hooks/lib.sh\",\n \"hooks/session_start.sh\",\n \"hooks/user_prompt_submit.sh\",\n \"hooks/post_tool_use.sh\",\n \"hooks/stop_hook.sh\",\n \"hooks/session_end.sh\",\n \"hooks/hook_utils.py\",\n \"hooks/parse_stop_transcript.py\",\n] as const;\n\nconst VENDORED_PLUGIN_RELATIVE_PATH = path.join(\"vendor\", \"claude-agents\", \"trace\");\n\nfunction currentModuleDir(): string {\n const originalPrepareStackTrace = Error.prepareStackTrace;\n\n try {\n Error.prepareStackTrace = (_error, stack) => stack;\n const callsites = new Error().stack as unknown as NodeJS.CallSite[] | undefined;\n const currentFile = callsites\n ?.map((callsite) => callsite.getFileName())\n .find((fileName) => fileName && !fileName.startsWith(\"node:\"));\n\n if (!currentFile) {\n throw new Error(\n \"PromptLayer Claude Agents could not determine its current module path.\"\n );\n }\n\n return path.dirname(\n currentFile.startsWith(\"file://\") ? fileURLToPath(currentFile) : currentFile\n );\n } finally {\n Error.prepareStackTrace = originalPrepareStackTrace;\n }\n}\n\nfunction assertSupportedPlatform(): void {\n if (process.platform === \"win32\") {\n throw new Error(\n \"PromptLayer Claude Agents integration does not support Windows. Use Linux or macOS.\"\n );\n }\n}\n\nfunction candidatePluginRoots(): string[] {\n const moduleDir = currentModuleDir();\n\n // Support all package layouts we exercise in development and publishing:\n // - source tests run from src/integrations/claude-agents/config.ts\n // - CJS consumers load dist/claude-agents.js\n // - ESM consumers load dist/esm/claude-agents.js\n // Each build places the current module at a different depth relative to\n // vendor/claude-agents/trace, so we probe the known stable offsets.\n return [\n path.resolve(moduleDir, \"..\", \"..\", \"..\", VENDORED_PLUGIN_RELATIVE_PATH),\n path.resolve(moduleDir, \"..\", VENDORED_PLUGIN_RELATIVE_PATH),\n path.resolve(moduleDir, \"..\", \"..\", VENDORED_PLUGIN_RELATIVE_PATH),\n ];\n}\n\nfunction assertRequiredVendoredFiles(pluginRoot: string): void {\n const missingFiles = REQUIRED_PLUGIN_FILES.filter(\n (relativePath) => !fs.existsSync(path.join(pluginRoot, relativePath))\n );\n\n if (missingFiles.length > 0) {\n throw new Error(\n `PromptLayer Claude Agents vendored plugin is incomplete. Missing: ${missingFiles.join(\", \")}`\n );\n }\n}\n\nfunction resolvePluginRoot(): string {\n for (const candidate of candidatePluginRoots()) {\n if (!fs.existsSync(candidate)) {\n continue;\n }\n\n const stats = fs.statSync(candidate);\n if (!stats.isDirectory()) {\n continue;\n }\n\n assertRequiredVendoredFiles(candidate);\n return candidate;\n }\n\n throw new Error(\n \"PromptLayer Claude Agents vendored plugin was not found in the installed package.\"\n );\n}\n\nexport function getClaudeConfig(\n options: GetClaudeConfigOptions = {}\n): PromptLayerClaudeAgentsConfig {\n assertSupportedPlatform();\n\n const pluginPath = resolvePluginRoot();\n const apiKey = requirePromptLayerApiKey(options.apiKey);\n const traceparent = options.traceparent?.trim();\n\n const env: PromptLayerClaudeAgentsEnv = {\n TRACE_TO_PROMPTLAYER: \"true\",\n PROMPTLAYER_API_KEY: apiKey,\n };\n\n if (traceparent) {\n env.PROMPTLAYER_TRACEPARENT = traceparent;\n }\n\n return {\n plugin: {\n type: \"local\",\n path: pluginPath,\n },\n env,\n };\n}\n","import {\n GetPromptTemplateParams,\n GetPromptTemplateResponse,\n ListPromptTemplatesResponse,\n LogRequest,\n Pagination,\n PublishPromptTemplate,\n PublishPromptTemplateResponse,\n RequestLog,\n RunWorkflowRequestParams,\n TrackGroup,\n TrackMetadata,\n TrackPrompt,\n TrackRequest,\n TrackScore,\n WorkflowResponse,\n} from \"@/types\";\nimport type { AnthropicBedrock } from \"@anthropic-ai/bedrock-sdk\";\nimport type TypeAnthropic from \"@anthropic-ai/sdk\";\nimport type { AnthropicVertex } from \"@anthropic-ai/vertex-sdk\";\nimport Ably from \"ably\";\nimport { Centrifuge } from \"centrifuge\";\nimport type TypeOpenAI from \"openai\";\nimport pRetry from \"p-retry\";\nimport {\n MAP_PROVIDER_TO_FUNCTION_NAME,\n cleaned_result,\n STREAMING_PROVIDERS_WITH_USAGE,\n} from \"./streaming\";\n\n// SDK version - injected at build time from package.json\ndeclare const __SDK_VERSION__: string;\nexport const SDK_VERSION = __SDK_VERSION__;\n\n// Get Node.js version (major.minor format)\nconst getNodeVersion = (): string => {\n if (typeof process !== \"undefined\" && process.versions?.node) {\n const parts = process.versions.node.split(\".\");\n return `${parts[0]}.${parts[1]}`;\n }\n return \"unknown\";\n};\n\nconst _NODE_VERSION = getNodeVersion();\nconst _PROMPTLAYER_USER_AGENT = `promptlayer-js/${SDK_VERSION} (node ${_NODE_VERSION})`;\n\n/**\n * Returns common headers to be included in all PromptLayer API requests.\n * Includes the SDK version and user agent for tracking and debugging purposes.\n */\nexport const getCommonHeaders = (): Record<string, string> => ({\n \"User-Agent\": _PROMPTLAYER_USER_AGENT,\n \"X-SDK-Version\": SDK_VERSION,\n});\n\nexport const SET_WORKFLOW_COMPLETE_MESSAGE = \"SET_WORKFLOW_COMPLETE\";\n\nexport enum FinalOutputCode {\n OK = \"OK\",\n EXCEEDS_SIZE_LIMIT = \"EXCEEDS_SIZE_LIMIT\",\n}\n\nasync function getFinalOutput(\n baseURL: string,\n executionId: number,\n returnAllOutputs: boolean,\n headers: Record<string, string>\n): Promise<any> {\n const response = await fetchWithRetry(\n `${baseURL}/workflow-version-execution-results?workflow_version_execution_id=${executionId}&return_all_outputs=${returnAllOutputs}`,\n { headers, ...getCommonHeaders() }\n );\n if (!response.ok) {\n throw new Error(\"Failed to fetch final output\");\n }\n return response.json();\n}\n\nfunction makeMessageListener(\n baseURL: string,\n resultsPromise: { resolve: (data: any) => void; reject: (err: any) => void },\n executionId: number,\n returnAllOutputs: boolean,\n headers: Record<string, string>\n) {\n return async function (message: any) {\n if (message.name !== SET_WORKFLOW_COMPLETE_MESSAGE) return;\n\n try {\n const data = JSON.parse(message.data);\n const resultCode = data.result_code;\n let results;\n\n if (resultCode === FinalOutputCode.OK || resultCode == null) {\n results = data.final_output;\n } else if (resultCode === FinalOutputCode.EXCEEDS_SIZE_LIMIT) {\n results = await getFinalOutput(\n baseURL,\n executionId,\n returnAllOutputs,\n headers\n );\n resultsPromise.resolve(results);\n } else {\n throw new Error(`Unsupported final output code: ${resultCode}`);\n }\n\n resultsPromise.resolve(results);\n } catch (err) {\n resultsPromise.reject(err);\n }\n };\n}\n\ninterface WaitForWorkflowCompletionParams {\n token: string;\n channelName: string;\n executionId: number;\n returnAllOutputs: boolean;\n headers: Record<string, string>;\n timeout: number;\n baseURL: string;\n}\n\nasync function waitForWorkflowCompletion({\n token,\n channelName,\n executionId,\n returnAllOutputs,\n headers,\n timeout,\n baseURL,\n}: WaitForWorkflowCompletionParams): Promise<any> {\n const client = new Ably.Realtime(token);\n const channel = client.channels.get(channelName);\n\n const resultsPromise = {} as {\n resolve: (value: any) => void;\n reject: (reason?: any) => void;\n };\n\n const promise = new Promise<any>((resolve, reject) => {\n resultsPromise.resolve = resolve;\n resultsPromise.reject = reject;\n });\n\n const listener = makeMessageListener(\n baseURL,\n resultsPromise,\n executionId,\n returnAllOutputs,\n headers\n );\n await channel.subscribe(SET_WORKFLOW_COMPLETE_MESSAGE, listener);\n\n try {\n return await new Promise((resolve, reject) => {\n const timer = setTimeout(() => {\n reject(\n new Error(\"Workflow execution did not complete properly (timeout)\")\n );\n }, timeout);\n\n promise\n .then((result) => {\n clearTimeout(timer);\n resolve(result);\n })\n .catch((err) => {\n clearTimeout(timer);\n reject(err);\n });\n });\n } finally {\n console.log(\"Closing client\");\n channel.unsubscribe(SET_WORKFLOW_COMPLETE_MESSAGE, listener);\n client.close();\n console.log(\"Closed client\");\n }\n}\n\n/**\n * Wrapper around fetch that retries on 5xx server errors with exponential backoff.\n * Uses p-retry for industry-standard retry logic with exponential backoff.\n *\n * @param input - The URL or Request object to fetch\n * @param init - The request initialization options\n * @returns Promise<Response> - The fetch response\n */\nexport const fetchWithRetry = async (\n input: RequestInfo | URL,\n init?: RequestInit\n): Promise<Response> => {\n return pRetry(\n async () => {\n const response = await fetch(input, init);\n\n if ((response.status >= 500 && response.status < 600) || (response.status === 429)) {\n throw new Error(\n `Server error: ${response.status} ${response.statusText}`\n );\n }\n\n return response;\n },\n {\n retries: 3, // Retry up to 3 times (4 total attempts)\n factor: 2, // Exponential backoff factor\n minTimeout: 2000, // First retry after 2 seconds\n maxTimeout: 15000, // Cap at 15 seconds (gives us ~2s, ~4s, ~8s progression with randomization)\n randomize: true, // Add jitter to avoid thundering herd\n onFailedAttempt: (error) => {\n console.info(\n `PromptLayer API request attempt ${error.attemptNumber} failed. ${error.retriesLeft} retries left.`\n );\n },\n }\n );\n};\n\nconst promptlayerApiHandler = async <Item>(\n apiKey: string,\n baseURL: string,\n body: TrackRequest & {\n request_response: AsyncIterable<Item> | any;\n },\n throwOnError: boolean = true\n) => {\n const isGenerator = body.request_response[Symbol.asyncIterator] !== undefined;\n if (isGenerator) {\n return proxyGenerator(\n apiKey,\n baseURL,\n body.request_response,\n body,\n throwOnError\n );\n }\n return await promptLayerApiRequest(apiKey, baseURL, body, throwOnError);\n};\n\nconst promptLayerApiRequest = async (\n apiKey: string,\n baseURL: string,\n body: TrackRequest,\n throwOnError: boolean = true\n) => {\n try {\n const response = await fetchWithRetry(`${baseURL}/track-request`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify(body),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to log request\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging your request, PromptLayer experienced the following error:\"\n );\n }\n }\n if (data && body.return_pl_id) {\n return [body.request_response, data.request_id];\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While logging your request PromptLayer had the following error: ${e}`\n );\n }\n return body.request_response;\n};\n\nconst promptLayerTrackMetadata = async (\n apiKey: string,\n baseURL: string,\n body: TrackMetadata,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/library-track-metadata`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track metadata\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging metadata to your request, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While logging metadata to your request, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerTrackScore = async (\n apiKey: string,\n baseURL: string,\n body: TrackScore,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/library-track-score`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track score\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While scoring your request, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While scoring your request, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerTrackPrompt = async (\n apiKey: string,\n baseURL: string,\n body: TrackPrompt,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/library-track-prompt`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track prompt\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While associating your request with a prompt template, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While associating your request with a prompt template, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerTrackGroup = async (\n apiKey: string,\n baseURL: string,\n body: TrackGroup,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/track-group`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track group\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While associating your request with a group, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While associating your request with a group, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerCreateGroup = async (\n apiKey: string,\n baseURL: string,\n throwOnError: boolean = true\n): Promise<number | boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/create-group`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to create group\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While creating a group PromptLayer had the following error\"\n );\n return false;\n }\n }\n return data.id;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While creating a group PromptLayer had the following error: ${e}`\n );\n return false;\n }\n};\n\nconst getPromptTemplate = async (\n apiKey: string,\n baseURL: string,\n promptName: string,\n params?: Partial<GetPromptTemplateParams>,\n throwOnError: boolean = true\n): Promise<GetPromptTemplateResponse | null> => {\n try {\n const url = new URL(`${baseURL}/prompt-templates/${promptName}`);\n const response = await fetchWithRetry(url, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n \"X-API-KEY\": apiKey,\n ...getCommonHeaders(),\n },\n body: JSON.stringify(params),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to fetch prompt template\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n console.warn(\n `WARNING: While fetching a prompt template PromptLayer had the following error: ${errorMessage}`\n );\n return null;\n }\n }\n if (data.warning) {\n console.warn(\n `WARNING: While fetching your prompt PromptLayer had the following error: ${data.warning}`\n );\n }\n return data as GetPromptTemplateResponse;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While fetching a prompt template PromptLayer had the following error: ${e}`\n );\n return null;\n }\n};\n\nconst publishPromptTemplate = async (\n apiKey: string,\n baseURL: string,\n body: PublishPromptTemplate,\n throwOnError: boolean = true\n): Promise<PublishPromptTemplateResponse> => {\n const response = await fetchWithRetry(`${baseURL}/rest/prompt-templates`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", \"X-API-KEY\": apiKey, ...getCommonHeaders() },\n body: JSON.stringify({\n prompt_template: { ...body },\n prompt_version: { ...body },\n release_labels: body.release_labels ? body.release_labels : undefined,\n }),\n });\n const data = await response.json();\n if (response.status !== 200 && response.status !== 201) {\n const errorMessage =\n data.message || data.error || \"Failed to publish prompt template\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While publishing a prompt template PromptLayer had the following error\"\n );\n }\n }\n return data as PublishPromptTemplateResponse;\n};\n\nconst getAllPromptTemplates = async (\n apiKey: string,\n baseURL: string,\n params?: Partial<Pagination>,\n throwOnError: boolean = true\n): Promise<Array<ListPromptTemplatesResponse>> => {\n const url = new URL(`${baseURL}/prompt-templates`);\n Object.entries(params || {}).forEach(([key, value]) =>\n url.searchParams.append(key, value.toString())\n );\n const response = await fetchWithRetry(url, {\n headers: { \"Content-Type\": \"application/json\", \"X-API-KEY\": apiKey, ...getCommonHeaders() },\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to fetch prompt templates\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While fetching all prompt templates PromptLayer had the following error\"\n );\n return [];\n }\n }\n return (data.items ?? []) as Array<ListPromptTemplatesResponse>;\n};\n\nconst waitForWorkflowCompletionCentrifugo = async (\n params: WaitForWorkflowCompletionParams\n): Promise<any> => {\n const url = new URL(`${params.baseURL}/connection/websocket`);\n url.protocol = url.protocol === \"https:\" ? \"wss:\" : \"ws:\";\n\n const client = new Centrifuge(url.toString(), { token: params.token });\n const sub = client.newSubscription(params.channelName);\n\n return new Promise((resolve, reject) => {\n const cleanupWithResolve = (data: any) => {\n cleanup();\n resolve(data);\n };\n\n const listener = makeMessageListener(\n params.baseURL,\n { resolve: cleanupWithResolve, reject },\n params.executionId,\n params.returnAllOutputs,\n params.headers\n );\n\n sub.on(\"publication\", (message) => {\n listener({\n name: message.data.message_name,\n data: message.data.data,\n });\n });\n\n const timeout = setTimeout(() => {\n reject(\n new Error(\"Workflow execution did not complete properly (timeout)\")\n );\n }, params.timeout);\n\n const cleanup = () => {\n clearTimeout(timeout);\n sub.unsubscribe();\n client.disconnect();\n };\n\n sub.on(\"error\", (err) => {\n cleanup();\n reject(`Centrifugo subscription error: ${err}`);\n });\n\n client.on(\"error\", (err) => {\n cleanup();\n reject(`Centrifugo client error: ${err}`);\n });\n\n sub.subscribe();\n client.connect();\n });\n};\n\nexport const runWorkflowRequest = async ({\n workflow_name,\n input_variables,\n metadata = {},\n workflow_label_name = null,\n workflow_version_number = null,\n return_all_outputs = false,\n api_key,\n timeout = 3600000, // Default timeout is 1 hour in milliseconds\n baseURL,\n}: RunWorkflowRequestParams): Promise<WorkflowResponse> => {\n const payload = {\n input_variables,\n metadata,\n workflow_label_name,\n workflow_version_number,\n return_all_outputs,\n };\n\n const headers = { \"Content-Type\": \"application/json\", \"X-API-KEY\": api_key, ...getCommonHeaders() };\n\n try {\n const response = await fetchWithRetry(\n `${baseURL}/workflows/${encodeURIComponent(workflow_name)}/run`,\n {\n method: \"POST\",\n headers: headers,\n body: JSON.stringify(payload),\n }\n );\n\n if (response.status !== 201) {\n const errorData = await response.json().catch(() => ({}));\n return {\n success: false,\n message: `Failed to run workflow: ${\n errorData.error || response.statusText\n }`,\n };\n }\n\n const result = await response.json();\n if (result.warning) {\n console.warn(`WARNING: ${result.warning}`);\n }\n const execution_id = result.workflow_version_execution_id;\n if (!execution_id) {\n console.log(\"No execution ID returned from workflow run\");\n return { success: false, message: \"Failed to run workflow\" };\n }\n\n const channel_name = `workflow_updates:${execution_id}`;\n const ws_response = await fetchWithRetry(\n `${baseURL}/ws-token-request-library?capability=${channel_name}`,\n {\n method: \"POST\",\n headers: headers,\n }\n );\n\n const ws_token_response = await ws_response.json();\n const token = ws_token_response.token_details.token;\n\n const params: WaitForWorkflowCompletionParams = {\n token,\n channelName: channel_name,\n executionId: execution_id,\n returnAllOutputs: return_all_outputs,\n headers: headers,\n timeout: timeout,\n baseURL: baseURL,\n };\n if (ws_token_response.messaging_backend === \"centrifugo\")\n return waitForWorkflowCompletionCentrifugo(params);\n return await waitForWorkflowCompletion(params);\n } catch (error) {\n console.error(\n `Failed to run workflow: ${\n error instanceof Error ? error.message : error\n }`\n );\n throw error;\n }\n};\n\nasync function* proxyGenerator<Item>(\n apiKey: string,\n baseURL: string,\n generator: AsyncIterable<Item>,\n body: TrackRequest,\n throwOnError: boolean = true\n) {\n const results = [];\n for await (const value of generator) {\n yield body.return_pl_id ? [value, null] : value;\n results.push(value);\n }\n const request_response = cleaned_result(results, body.function_name);\n const response = await promptLayerApiRequest(\n apiKey,\n baseURL,\n {\n ...body,\n request_response,\n request_end_time: new Date().toISOString(),\n },\n throwOnError\n );\n if (response) {\n if (body.return_pl_id) {\n const request_id = (response as any)[1];\n const lastResult = results.at(-1);\n yield [lastResult, request_id];\n }\n }\n}\n\nconst warnOnBadResponse = (request_response: any, main_message: string) => {\n try {\n console.warn(`${main_message}: ${request_response.message}`);\n } catch (e) {\n console.warn(`${main_message}: ${request_response}`);\n }\n};\n\nconst trackRequest = async (\n baseURL: string,\n body: TrackRequest,\n throwOnError: boolean = true\n) => {\n try {\n const response = await fetchWithRetry(`${baseURL}/track-request`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify(body),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track request\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging your request, PromptLayer experienced the following error:\"\n );\n }\n }\n return data;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While logging your request PromptLayer had the following error: ${e}`\n );\n }\n return {};\n};\n\nconst openaiChatRequest = async (client: TypeOpenAI, kwargs: any) => {\n return await client.chat.completions.create(kwargs);\n};\n\nconst openaiCompletionsRequest = async (client: TypeOpenAI, kwargs: any) => {\n return await client.completions.create(kwargs);\n};\n\nconst MAP_TYPE_TO_OPENAI_FUNCTION = {\n chat: openaiChatRequest,\n completion: openaiCompletionsRequest,\n};\n\nconst openaiRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const OpenAI = require(\"openai\").default;\n const client = new OpenAI({\n baseURL: kwargs.baseURL,\n apiKey: kwargs.apiKey,\n });\n\n delete kwargs?.apiKey;\n delete kwargs?.baseURL;\n\n const api_type = promptBlueprint.metadata?.model?.api_type;\n if (api_type === \"chat-completions\") {\n const requestToMake =\n MAP_TYPE_TO_OPENAI_FUNCTION[promptBlueprint.prompt_template.type];\n return await requestToMake(client, kwargs);\n } else if (api_type === \"images\") {\n return await client.images.generate(kwargs);\n } else {\n return await client.responses.create(kwargs);\n }\n};\n\nconst azureOpenAIRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { AzureOpenAI } = require(\"openai\");\n const client = new AzureOpenAI({\n endpoint: process.env.AZURE_OPENAI_ENDPOINT || kwargs.baseURL,\n apiVersion: process.env.OPENAI_API_VERSION || kwargs.apiVersion,\n apiKey: process.env.AZURE_OPENAI_API_KEY || kwargs.apiKey,\n });\n delete kwargs?.baseURL;\n delete kwargs?.apiVersion;\n delete kwargs?.apiKey;\n\n const api_type = promptBlueprint.metadata?.model?.api_type;\n\n if (api_type === \"chat-completions\") {\n const requestToMake = MAP_TYPE_TO_OPENAI_FUNCTION[promptBlueprint.prompt_template.type];\n return await requestToMake(client, kwargs);\n } else if (api_type === \"images\") {\n return await client.images.generate(kwargs);\n } else {\n return await client.responses.create(kwargs);\n }\n};\n\nconst anthropicChatRequest = async (\n client: TypeAnthropic | AnthropicVertex | AnthropicBedrock,\n kwargs: any\n) => {\n return client.messages.create(kwargs);\n};\n\nconst anthropicCompletionsRequest = async (\n client: TypeAnthropic | AnthropicBedrock,\n kwargs: any\n) => {\n return client.completions.create(kwargs);\n};\n\nconst MAP_TYPE_TO_ANTHROPIC_FUNCTION = {\n chat: anthropicChatRequest,\n completion: anthropicCompletionsRequest,\n};\n\nconst anthropicRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const Anthropic = require(\"@anthropic-ai/sdk\").default;\n const client = new Anthropic({\n baseURL: kwargs.baseURL,\n apiKey: kwargs.apiKey,\n });\n const requestToMake =\n MAP_TYPE_TO_ANTHROPIC_FUNCTION[promptBlueprint.prompt_template.type];\n return requestToMake(client, kwargs);\n};\n\nconst utilLogRequest = async (\n apiKey: string,\n baseURL: string,\n body: LogRequest,\n throwOnError: boolean = true\n): Promise<RequestLog | null> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/log-request`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", \"X-API-KEY\": apiKey, ...getCommonHeaders() },\n body: JSON.stringify(body),\n });\n const data = await response.json();\n if (response.status !== 201) {\n const errorMessage =\n data.message || data.error || \"Failed to log request\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging your request PromptLayer had the following error\"\n );\n return null;\n }\n }\n return data;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While tracking your prompt PromptLayer had the following error: ${e}`\n );\n return null;\n }\n};\n\nconst googleChatRequest = async (model_client: any, kwargs: any) => {\n const history = kwargs?.history;\n const generationConfig = kwargs?.generationConfig;\n const lastMessage =\n history.length > 0 ? history[history.length - 1]?.parts : \"\";\n const chat = model_client.chats.create({\n model: kwargs?.model,\n history: history.slice(0, -1) ?? [],\n config: generationConfig,\n });\n\n if (kwargs?.stream)\n return await chat.sendMessageStream({ message: lastMessage });\n return await chat.sendMessage({ message: lastMessage });\n};\n\nconst googleCompletionsRequest = async (\n model_client: any,\n { stream, ...kwargs }: any\n) => {\n if (stream) return await model_client.generateContentStream({ ...kwargs });\n return await model_client.generateContent({ ...kwargs });\n};\n\nconst MAP_TYPE_TO_GOOGLE_FUNCTION = {\n chat: googleChatRequest,\n completion: googleCompletionsRequest,\n};\n\nconst googleRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { GoogleGenAI } = await import(\"@google/genai\");\n\n const geminiAPI = process.env.GOOGLE_API_KEY || process.env.GEMINI_API_KEY;\n const project =\n process.env.VERTEX_AI_PROJECT_ID ||\n process.env.GOOGLE_PROJECT_ID ||\n process.env.GOOGLE_CLOUD_PROJECT;\n const location =\n process.env.VERTEX_AI_PROJECT_LOCATION ||\n process.env.GOOGLE_PROJECT_LOCATION ||\n process.env.GOOGLE_CLOUD_PROJECT_LOCATION;\n const googleAuthOptions = {\n keyFilename: process.env.GOOGLE_APPLICATION_CREDENTIALS,\n projectId: project,\n scopes: [\"https://www.googleapis.com/auth/cloud-platform\"],\n };\n\n const genAI = geminiAPI\n ? new GoogleGenAI({ apiKey: geminiAPI })\n : new GoogleGenAI({\n vertexai: true,\n project: project,\n location: location,\n googleAuthOptions,\n });\n const requestToMake =\n MAP_TYPE_TO_GOOGLE_FUNCTION[promptBlueprint.prompt_template.type];\n\n return await requestToMake(genAI, kwargs);\n};\n\nconst snakeToCamel = (str: string): string =>\n str.replace(/_([a-z])/g, (_, letter) => letter.toUpperCase());\n\nconst convertKeysToCamelCase = <T>(\n obj: T,\n ignoreValuesWithKeys: Set<string> = new Set()\n): T => {\n if (!obj || typeof obj !== \"object\") return obj;\n if (Array.isArray(obj))\n return obj.map((item) =>\n convertKeysToCamelCase(item, ignoreValuesWithKeys)\n ) as T;\n\n return Object.fromEntries(\n Object.entries(obj).map(([key, value]) => {\n if (ignoreValuesWithKeys.has(key)) {\n return [snakeToCamel(key), value];\n }\n return [\n snakeToCamel(key),\n convertKeysToCamelCase(value, ignoreValuesWithKeys),\n ];\n })\n ) as T;\n};\n\nconst configureProviderSettings = (\n promptBlueprint: any,\n customProvider: any,\n modelParameterOverrides: any = {},\n stream: boolean = false\n) => {\n const provider_type =\n customProvider?.client ?? promptBlueprint.metadata?.model?.provider;\n const api_type = promptBlueprint.metadata?.model?.api_type;\n\n if (!provider_type) {\n throw new Error(\n \"Provider type not found in prompt blueprint or custom provider\"\n );\n }\n\n let kwargs = {\n ...(promptBlueprint.llm_kwargs || {}),\n stream,\n };\n\n if (\n [\"google\", \"vertexai\"].includes(provider_type) &&\n promptBlueprint.metadata?.model?.name.startsWith(\"gemini\")\n )\n kwargs = convertKeysToCamelCase(\n kwargs,\n new Set([\"function_declarations\", \"properties\"])\n );\n\n const providerConfig = {\n baseURL: customProvider?.base_url ?? promptBlueprint.provider_base_url?.url,\n apiKey: customProvider?.api_key,\n };\n\n Object.entries(providerConfig).forEach(([key, value]) => {\n if (value !== undefined) {\n kwargs[key] = value;\n }\n });\n\n if (stream && STREAMING_PROVIDERS_WITH_USAGE.includes(provider_type as any) && api_type === \"chat-completions\") {\n kwargs.stream_options = { include_usage: true };\n }\n\n return { provider_type, kwargs };\n};\n\nconst getProviderConfig = (provider_type: string, promptTemplate: any) => {\n const providerMap =\n MAP_PROVIDER_TO_FUNCTION_NAME[\n provider_type as keyof typeof MAP_PROVIDER_TO_FUNCTION_NAME\n ];\n\n if (!providerMap) {\n throw new Error(`Unsupported provider type: ${provider_type}`);\n }\n\n const templateType = promptTemplate.type as keyof typeof providerMap;\n const config = providerMap[templateType];\n\n if (!config) {\n throw new Error(\n `Unsupported template type '${promptTemplate.type}' for provider '${provider_type}'`\n );\n }\n\n return config;\n};\n\nconst vertexaiRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const model = promptBlueprint.metadata?.model;\n if (!model) throw new Error(\"Model metadata not found in prompt blueprint\");\n if (model.name.startsWith(\"gemini\"))\n return googleRequest(promptBlueprint, kwargs);\n if (model.name.startsWith(\"claude\")) {\n const { AnthropicVertex } = await import(\"@anthropic-ai/vertex-sdk\");\n const client = new AnthropicVertex({ baseURL: kwargs.baseURL });\n if (promptBlueprint.prompt_template.type === \"chat\")\n return anthropicChatRequest(client, kwargs);\n throw new Error(\n `Unsupported prompt template type '${promptBlueprint.prompt_template.type}' for Anthropic Vertex AI`\n );\n }\n throw new Error(\n `Unsupported model name '${model.name}' for Vertex AI request`\n );\n};\n\nconst amazonBedrockRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { BedrockRuntimeClient, ConverseCommand, ConverseStreamCommand } =\n await import(\"@aws-sdk/client-bedrock-runtime\");\n const client = new BedrockRuntimeClient({\n credentials: {\n accessKeyId: kwargs?.aws_access_key || process.env.AWS_ACCESS_KEY_ID,\n secretAccessKey:\n kwargs?.aws_secret_key || process.env.AWS_SECRET_ACCESS_KEY,\n sessionToken: kwargs?.aws_session_token || process.env.AWS_SESSION_TOKEN,\n },\n region:\n kwargs?.aws_region ||\n process.env.AWS_REGION ||\n process.env.AWS_DEFAULT_REGION ||\n \"us-east-1\",\n });\n\n if (kwargs?.stream) {\n delete kwargs.stream;\n const command = new ConverseStreamCommand({\n ...kwargs,\n });\n return await client.send(command);\n } else {\n delete kwargs?.stream;\n const command = new ConverseCommand({\n ...kwargs,\n });\n return await client.send(command);\n }\n};\n\nconst anthropicBedrockRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { AnthropicBedrock } = await import(\"@anthropic-ai/bedrock-sdk\");\n const client = new AnthropicBedrock({\n awsAccessKey: kwargs.aws_access_key,\n awsSecretKey: kwargs.aws_secret_key,\n awsRegion: kwargs.aws_region,\n awsSessionToken: kwargs.aws_session_token,\n baseURL: kwargs.base_url,\n });\n\n const requestToMake =\n MAP_TYPE_TO_ANTHROPIC_FUNCTION[promptBlueprint.prompt_template.type];\n return requestToMake(client, kwargs);\n};\n\nconst mistralRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { Mistral } = await import(\"@mistralai/mistralai\");\n const client = new Mistral({ apiKey: process.env.MISTRAL_API_KEY });\n kwargs = convertKeysToCamelCase(kwargs, new Set());\n if (kwargs?.stream) {\n delete kwargs.stream;\n return await client.chat.stream(kwargs);\n }\n delete kwargs.stream;\n return await client.chat.complete(kwargs);\n};\n\nexport const readEnv = (env: string): string | undefined => {\n if (typeof (globalThis as any).process !== \"undefined\")\n return (globalThis as any).process.env?.[env]?.trim() ?? undefined;\n\n if (typeof (globalThis as any).Deno !== \"undefined\")\n return (globalThis as any).Deno.env?.get?.(env)?.trim();\n return undefined;\n};\n\nexport const requirePromptLayerApiKey = (apiKey?: string): string => {\n if (apiKey !== undefined) {\n const explicitApiKey = apiKey.trim();\n if (!explicitApiKey) {\n throw new Error(\n \"PromptLayer API key not provided. Please set PROMPTLAYER_API_KEY or pass apiKey.\"\n );\n }\n return explicitApiKey;\n }\n\n const envApiKey = readEnv(\"PROMPTLAYER_API_KEY\");\n if (!envApiKey) {\n throw new Error(\n \"PromptLayer API key not provided. Please set PROMPTLAYER_API_KEY or pass apiKey.\"\n );\n }\n\n return envApiKey;\n};\n\nexport {\n amazonBedrockRequest,\n anthropicBedrockRequest,\n anthropicRequest,\n azureOpenAIRequest,\n configureProviderSettings,\n getAllPromptTemplates,\n getPromptTemplate,\n getProviderConfig,\n googleRequest,\n mistralRequest,\n openaiRequest,\n promptlayerApiHandler,\n promptLayerApiRequest,\n promptLayerCreateGroup,\n promptLayerTrackGroup,\n promptLayerTrackMetadata,\n promptLayerTrackPrompt,\n promptLayerTrackScore,\n publishPromptTemplate,\n trackRequest,\n utilLogRequest,\n vertexaiRequest,\n};\n"],"mappings":"0jBAAA,IAAAA,EAAA,GAAAC,EAAAD,EAAA,qBAAAE,IAAA,eAAAC,EAAAH,GCAA,IAAAI,EAAe,iBACfC,EAAiB,mBACjBC,EAA8B,eCkB9B,IAAAC,EAAiB,mBACjBC,EAA2B,sBAE3BC,EAAmB,sBASZ,IAAMC,EAAc,QAGrBC,EAAiB,IAAc,CAnCrC,IAAAC,EAoCE,GAAI,OAAO,SAAY,eAAeA,EAAA,QAAQ,WAAR,MAAAA,EAAkB,MAAM,CAC5D,IAAMC,EAAQ,QAAQ,SAAS,KAAK,MAAM,GAAG,EAC7C,MAAO,GAAGA,EAAM,CAAC,CAAC,IAAIA,EAAM,CAAC,CAAC,EAChC,CACA,MAAO,SACT,EAEMC,EAAgBH,EAAe,EAC/BI,GAA0B,kBAAkBL,CAAW,UAAUI,CAAa,IA+nC7E,IAAME,EAAWC,GAAoC,CA3qC5D,IAAAC,EAAAC,EAAAC,EAAAC,EAAAC,EAAAC,EA4qCE,GAAI,OAAQ,WAAmB,SAAY,YACzC,OAAQH,GAAAD,GAAAD,EAAA,WAAmB,QAAQ,MAA3B,YAAAA,EAAiCD,KAAjC,YAAAE,EAAuC,SAAvC,KAAAC,EAAiD,OAE3D,GAAI,OAAQ,WAAmB,MAAS,YACtC,OAAQG,GAAAD,GAAAD,EAAA,WAAmB,KAAK,MAAxB,YAAAA,EAA6B,MAA7B,YAAAC,EAAA,KAAAD,EAAmCJ,KAAnC,YAAAM,EAAyC,MAErD,EAEaC,EAA4BC,GAA4B,CACnE,GAAIA,IAAW,OAAW,CACxB,IAAMC,EAAiBD,EAAO,KAAK,EACnC,GAAI,CAACC,EACH,MAAM,IAAI,MACR,kFACF,EAEF,OAAOA,CACT,CAEA,IAAMC,EAAYX,EAAQ,qBAAqB,EAC/C,GAAI,CAACW,EACH,MAAM,IAAI,MACR,kFACF,EAGF,OAAOA,CACT,ED7qCA,IAAMC,EAAwB,CAC5B,6BACA,mBACA,eACA,yBACA,8BACA,yBACA,qBACA,uBACA,sBACA,gCACF,EAEMC,EAAgC,EAAAC,QAAK,KAAK,SAAU,gBAAiB,OAAO,EAElF,SAASC,GAA2B,CAClC,IAAMC,EAA4B,MAAM,kBAExC,GAAI,CACF,MAAM,kBAAoB,CAACC,EAAQC,IAAUA,EAC7C,IAAMC,EAAY,IAAI,MAAM,EAAE,MACxBC,EAAcD,GAAA,YAAAA,EAChB,IAAKE,GAAaA,EAAS,YAAY,GACxC,KAAMC,GAAaA,GAAY,CAACA,EAAS,WAAW,OAAO,GAE9D,GAAI,CAACF,EACH,MAAM,IAAI,MACR,wEACF,EAGF,OAAO,EAAAN,QAAK,QACVM,EAAY,WAAW,SAAS,KAAI,iBAAcA,CAAW,EAAIA,CACnE,CACF,QAAE,CACA,MAAM,kBAAoBJ,CAC5B,CACF,CAEA,SAASO,GAAgC,CACvC,GAAI,QAAQ,WAAa,QACvB,MAAM,IAAI,MACR,qFACF,CAEJ,CAEA,SAASC,GAAiC,CACxC,IAAMC,EAAYV,EAAiB,EAQnC,MAAO,CACL,EAAAD,QAAK,QAAQW,EAAW,KAAM,KAAM,KAAMZ,CAA6B,EACvE,EAAAC,QAAK,QAAQW,EAAW,KAAMZ,CAA6B,EAC3D,EAAAC,QAAK,QAAQW,EAAW,KAAM,KAAMZ,CAA6B,CACnE,CACF,CAEA,SAASa,EAA4BC,EAA0B,CAC7D,IAAMC,EAAehB,EAAsB,OACxCiB,GAAiB,CAAC,EAAAC,QAAG,WAAW,EAAAhB,QAAK,KAAKa,EAAYE,CAAY,CAAC,CACtE,EAEA,GAAID,EAAa,OAAS,EACxB,MAAM,IAAI,MACR,qEAAqEA,EAAa,KAAK,IAAI,CAAC,EAC9F,CAEJ,CAEA,SAASG,GAA4B,CACnC,QAAWC,KAAaR,EAAqB,EAM3C,GALI,GAAC,EAAAM,QAAG,WAAWE,CAAS,GAKxB,CADU,EAAAF,QAAG,SAASE,CAAS,EACxB,YAAY,GAIvB,OAAAN,EAA4BM,CAAS,EAC9BA,EAGT,MAAM,IAAI,MACR,mFACF,CACF,CAEO,SAASC,EACdC,EAAkC,CAAC,EACJ,CA3HjC,IAAAC,EA4HEZ,EAAwB,EAExB,IAAMa,EAAaL,EAAkB,EAC/BM,EAASC,EAAyBJ,EAAQ,MAAM,EAChDK,GAAcJ,EAAAD,EAAQ,cAAR,YAAAC,EAAqB,OAEnCK,EAAkC,CACtC,qBAAsB,OACtB,oBAAqBH,CACvB,EAEA,OAAIE,IACFC,EAAI,wBAA0BD,GAGzB,CACL,OAAQ,CACN,KAAM,QACN,KAAMH,CACR,EACA,IAAAI,CACF,CACF","names":["claude_agents_exports","__export","getClaudeConfig","__toCommonJS","import_fs","import_path","import_url","import_ably","import_centrifuge","import_p_retry","SDK_VERSION","getNodeVersion","_a","parts","_NODE_VERSION","_PROMPTLAYER_USER_AGENT","readEnv","env","_a","_b","_c","_d","_e","_f","requirePromptLayerApiKey","apiKey","explicitApiKey","envApiKey","REQUIRED_PLUGIN_FILES","VENDORED_PLUGIN_RELATIVE_PATH","path","currentModuleDir","originalPrepareStackTrace","_error","stack","callsites","currentFile","callsite","fileName","assertSupportedPlatform","candidatePluginRoots","moduleDir","assertRequiredVendoredFiles","pluginRoot","missingFiles","relativePath","fs","resolvePluginRoot","candidate","getClaudeConfig","options","_a","pluginPath","apiKey","requirePromptLayerApiKey","traceparent","env"]}
|
|
1
|
+
{"version":3,"sources":["../src/claude-agents.ts","../src/integrations/claude-agents/config.ts","../src/utils/utils.ts"],"sourcesContent":["export * from \"@/integrations/claude-agents\";\n","import fs from \"fs\";\nimport path from \"path\";\nimport { fileURLToPath } from \"url\";\nimport { requirePromptLayerApiKey } from \"@/utils/utils\";\n\nexport interface GetClaudeConfigOptions {\n apiKey?: string;\n traceparent?: string;\n}\n\nexport interface PromptLayerClaudeAgentsPlugin {\n type: \"local\";\n path: string;\n}\n\nexport interface PromptLayerClaudeAgentsEnv {\n TRACE_TO_PROMPTLAYER: \"true\";\n PROMPTLAYER_API_KEY: string;\n PROMPTLAYER_TRACEPARENT?: string;\n}\n\nexport interface PromptLayerClaudeAgentsConfig {\n plugin: PromptLayerClaudeAgentsPlugin;\n env: PromptLayerClaudeAgentsEnv;\n}\n\nconst REQUIRED_PLUGIN_FILES = [\n \".claude-plugin/plugin.json\",\n \"setup.sh\",\n \"hooks/hooks.json\",\n \"hooks/lib.sh\",\n \"hooks/session_start.sh\",\n \"hooks/user_prompt_submit.sh\",\n \"hooks/post_tool_use.sh\",\n \"hooks/stop_hook.sh\",\n \"hooks/session_end.sh\",\n \"hooks/py/__init__.py\",\n \"hooks/py/cli.py\",\n \"hooks/py/context.py\",\n \"hooks/py/handlers.py\",\n \"hooks/py/otlp.py\",\n \"hooks/py/settings.py\",\n \"hooks/py/state.py\",\n \"hooks/py/stop_parser.py\",\n \"hooks/py/traceparent.py\",\n] as const;\n\nconst VENDORED_PLUGIN_RELATIVE_PATH = path.join(\"vendor\", \"claude-agents\", \"trace\");\n\nfunction currentModuleDir(): string {\n const originalPrepareStackTrace = Error.prepareStackTrace;\n\n try {\n Error.prepareStackTrace = (_error, stack) => stack;\n const callsites = new Error().stack as unknown as NodeJS.CallSite[] | undefined;\n const currentFile = callsites\n ?.map((callsite) => callsite.getFileName())\n .find((fileName) => fileName && !fileName.startsWith(\"node:\"));\n\n if (!currentFile) {\n throw new Error(\n \"PromptLayer Claude Agents could not determine its current module path.\"\n );\n }\n\n return path.dirname(\n currentFile.startsWith(\"file://\") ? fileURLToPath(currentFile) : currentFile\n );\n } finally {\n Error.prepareStackTrace = originalPrepareStackTrace;\n }\n}\n\nfunction assertSupportedPlatform(): void {\n if (process.platform === \"win32\") {\n throw new Error(\n \"PromptLayer Claude Agents integration does not support Windows. Use Linux or macOS.\"\n );\n }\n}\n\nfunction candidatePluginRoots(): string[] {\n const moduleDir = currentModuleDir();\n\n // Support all package layouts we exercise in development and publishing:\n // - source tests run from src/integrations/claude-agents/config.ts\n // - CJS consumers load dist/claude-agents.js\n // - ESM consumers load dist/esm/claude-agents.js\n // Each build places the current module at a different depth relative to\n // vendor/claude-agents/trace, so we probe the known stable offsets.\n return [\n path.resolve(moduleDir, \"..\", \"..\", \"..\", VENDORED_PLUGIN_RELATIVE_PATH),\n path.resolve(moduleDir, \"..\", VENDORED_PLUGIN_RELATIVE_PATH),\n path.resolve(moduleDir, \"..\", \"..\", VENDORED_PLUGIN_RELATIVE_PATH),\n ];\n}\n\nfunction assertRequiredVendoredFiles(pluginRoot: string): void {\n const missingFiles = REQUIRED_PLUGIN_FILES.filter(\n (relativePath) => !fs.existsSync(path.join(pluginRoot, relativePath))\n );\n\n if (missingFiles.length > 0) {\n throw new Error(\n `PromptLayer Claude Agents vendored plugin is incomplete. Missing: ${missingFiles.join(\", \")}`\n );\n }\n}\n\nfunction resolvePluginRoot(): string {\n for (const candidate of candidatePluginRoots()) {\n if (!fs.existsSync(candidate)) {\n continue;\n }\n\n const stats = fs.statSync(candidate);\n if (!stats.isDirectory()) {\n continue;\n }\n\n assertRequiredVendoredFiles(candidate);\n return candidate;\n }\n\n throw new Error(\n \"PromptLayer Claude Agents vendored plugin was not found in the installed package.\"\n );\n}\n\nexport function getClaudeConfig(\n options: GetClaudeConfigOptions = {}\n): PromptLayerClaudeAgentsConfig {\n assertSupportedPlatform();\n\n const pluginPath = resolvePluginRoot();\n const apiKey = requirePromptLayerApiKey(options.apiKey);\n const traceparent = options.traceparent?.trim();\n\n const env: PromptLayerClaudeAgentsEnv = {\n TRACE_TO_PROMPTLAYER: \"true\",\n PROMPTLAYER_API_KEY: apiKey,\n };\n\n if (traceparent) {\n env.PROMPTLAYER_TRACEPARENT = traceparent;\n }\n\n return {\n plugin: {\n type: \"local\",\n path: pluginPath,\n },\n env,\n };\n}\n","import {\n GetPromptTemplateParams,\n GetPromptTemplateResponse,\n ListPromptTemplatesResponse,\n LogRequest,\n Pagination,\n PublishPromptTemplate,\n PublishPromptTemplateResponse,\n RequestLog,\n RunWorkflowRequestParams,\n TrackGroup,\n TrackMetadata,\n TrackPrompt,\n TrackRequest,\n TrackScore,\n WorkflowResponse,\n} from \"@/types\";\nimport type { AnthropicBedrock } from \"@anthropic-ai/bedrock-sdk\";\nimport type TypeAnthropic from \"@anthropic-ai/sdk\";\nimport type { AnthropicVertex } from \"@anthropic-ai/vertex-sdk\";\nimport Ably from \"ably\";\nimport { Centrifuge } from \"centrifuge\";\nimport type TypeOpenAI from \"openai\";\nimport pRetry from \"p-retry\";\nimport {\n MAP_PROVIDER_TO_FUNCTION_NAME,\n cleaned_result,\n STREAMING_PROVIDERS_WITH_USAGE,\n} from \"./streaming\";\n\n// SDK version - injected at build time from package.json\ndeclare const __SDK_VERSION__: string;\nexport const SDK_VERSION = __SDK_VERSION__;\n\n// Get Node.js version (major.minor format)\nconst getNodeVersion = (): string => {\n if (typeof process !== \"undefined\" && process.versions?.node) {\n const parts = process.versions.node.split(\".\");\n return `${parts[0]}.${parts[1]}`;\n }\n return \"unknown\";\n};\n\nconst _NODE_VERSION = getNodeVersion();\nconst _PROMPTLAYER_USER_AGENT = `promptlayer-js/${SDK_VERSION} (node ${_NODE_VERSION})`;\n\n/**\n * Returns common headers to be included in all PromptLayer API requests.\n * Includes the SDK version and user agent for tracking and debugging purposes.\n */\nexport const getCommonHeaders = (): Record<string, string> => ({\n \"User-Agent\": _PROMPTLAYER_USER_AGENT,\n \"X-SDK-Version\": SDK_VERSION,\n});\n\nexport const SET_WORKFLOW_COMPLETE_MESSAGE = \"SET_WORKFLOW_COMPLETE\";\n\nexport enum FinalOutputCode {\n OK = \"OK\",\n EXCEEDS_SIZE_LIMIT = \"EXCEEDS_SIZE_LIMIT\",\n}\n\nasync function getFinalOutput(\n baseURL: string,\n executionId: number,\n returnAllOutputs: boolean,\n headers: Record<string, string>\n): Promise<any> {\n const response = await fetchWithRetry(\n `${baseURL}/workflow-version-execution-results?workflow_version_execution_id=${executionId}&return_all_outputs=${returnAllOutputs}`,\n { headers, ...getCommonHeaders() }\n );\n if (!response.ok) {\n throw new Error(\"Failed to fetch final output\");\n }\n return response.json();\n}\n\nfunction makeMessageListener(\n baseURL: string,\n resultsPromise: { resolve: (data: any) => void; reject: (err: any) => void },\n executionId: number,\n returnAllOutputs: boolean,\n headers: Record<string, string>\n) {\n return async function (message: any) {\n if (message.name !== SET_WORKFLOW_COMPLETE_MESSAGE) return;\n\n try {\n const data = JSON.parse(message.data);\n const resultCode = data.result_code;\n let results;\n\n if (resultCode === FinalOutputCode.OK || resultCode == null) {\n results = data.final_output;\n } else if (resultCode === FinalOutputCode.EXCEEDS_SIZE_LIMIT) {\n results = await getFinalOutput(\n baseURL,\n executionId,\n returnAllOutputs,\n headers\n );\n resultsPromise.resolve(results);\n } else {\n throw new Error(`Unsupported final output code: ${resultCode}`);\n }\n\n resultsPromise.resolve(results);\n } catch (err) {\n resultsPromise.reject(err);\n }\n };\n}\n\ninterface WaitForWorkflowCompletionParams {\n token: string;\n channelName: string;\n executionId: number;\n returnAllOutputs: boolean;\n headers: Record<string, string>;\n timeout: number;\n baseURL: string;\n}\n\nasync function waitForWorkflowCompletion({\n token,\n channelName,\n executionId,\n returnAllOutputs,\n headers,\n timeout,\n baseURL,\n}: WaitForWorkflowCompletionParams): Promise<any> {\n const client = new Ably.Realtime(token);\n const channel = client.channels.get(channelName);\n\n const resultsPromise = {} as {\n resolve: (value: any) => void;\n reject: (reason?: any) => void;\n };\n\n const promise = new Promise<any>((resolve, reject) => {\n resultsPromise.resolve = resolve;\n resultsPromise.reject = reject;\n });\n\n const listener = makeMessageListener(\n baseURL,\n resultsPromise,\n executionId,\n returnAllOutputs,\n headers\n );\n await channel.subscribe(SET_WORKFLOW_COMPLETE_MESSAGE, listener);\n\n try {\n return await new Promise((resolve, reject) => {\n const timer = setTimeout(() => {\n reject(\n new Error(\"Workflow execution did not complete properly (timeout)\")\n );\n }, timeout);\n\n promise\n .then((result) => {\n clearTimeout(timer);\n resolve(result);\n })\n .catch((err) => {\n clearTimeout(timer);\n reject(err);\n });\n });\n } finally {\n console.log(\"Closing client\");\n channel.unsubscribe(SET_WORKFLOW_COMPLETE_MESSAGE, listener);\n client.close();\n console.log(\"Closed client\");\n }\n}\n\n/**\n * Wrapper around fetch that retries on 5xx server errors with exponential backoff.\n * Uses p-retry for industry-standard retry logic with exponential backoff.\n *\n * @param input - The URL or Request object to fetch\n * @param init - The request initialization options\n * @returns Promise<Response> - The fetch response\n */\nexport const fetchWithRetry = async (\n input: RequestInfo | URL,\n init?: RequestInit\n): Promise<Response> => {\n return pRetry(\n async () => {\n const response = await fetch(input, init);\n\n if ((response.status >= 500 && response.status < 600) || (response.status === 429)) {\n throw new Error(\n `Server error: ${response.status} ${response.statusText}`\n );\n }\n\n return response;\n },\n {\n retries: 3, // Retry up to 3 times (4 total attempts)\n factor: 2, // Exponential backoff factor\n minTimeout: 2000, // First retry after 2 seconds\n maxTimeout: 15000, // Cap at 15 seconds (gives us ~2s, ~4s, ~8s progression with randomization)\n randomize: true, // Add jitter to avoid thundering herd\n onFailedAttempt: (error) => {\n console.info(\n `PromptLayer API request attempt ${error.attemptNumber} failed. ${error.retriesLeft} retries left.`\n );\n },\n }\n );\n};\n\nconst promptlayerApiHandler = async <Item>(\n apiKey: string,\n baseURL: string,\n body: TrackRequest & {\n request_response: AsyncIterable<Item> | any;\n },\n throwOnError: boolean = true\n) => {\n const isGenerator = body.request_response[Symbol.asyncIterator] !== undefined;\n if (isGenerator) {\n return proxyGenerator(\n apiKey,\n baseURL,\n body.request_response,\n body,\n throwOnError\n );\n }\n return await promptLayerApiRequest(apiKey, baseURL, body, throwOnError);\n};\n\nconst promptLayerApiRequest = async (\n apiKey: string,\n baseURL: string,\n body: TrackRequest,\n throwOnError: boolean = true\n) => {\n try {\n const response = await fetchWithRetry(`${baseURL}/track-request`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify(body),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to log request\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging your request, PromptLayer experienced the following error:\"\n );\n }\n }\n if (data && body.return_pl_id) {\n return [body.request_response, data.request_id];\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While logging your request PromptLayer had the following error: ${e}`\n );\n }\n return body.request_response;\n};\n\nconst promptLayerTrackMetadata = async (\n apiKey: string,\n baseURL: string,\n body: TrackMetadata,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/library-track-metadata`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track metadata\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging metadata to your request, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While logging metadata to your request, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerTrackScore = async (\n apiKey: string,\n baseURL: string,\n body: TrackScore,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/library-track-score`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track score\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While scoring your request, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While scoring your request, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerTrackPrompt = async (\n apiKey: string,\n baseURL: string,\n body: TrackPrompt,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/library-track-prompt`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track prompt\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While associating your request with a prompt template, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While associating your request with a prompt template, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerTrackGroup = async (\n apiKey: string,\n baseURL: string,\n body: TrackGroup,\n throwOnError: boolean = true\n): Promise<boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/track-group`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n ...body,\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track group\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While associating your request with a group, PromptLayer experienced the following error\"\n );\n return false;\n }\n }\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While associating your request with a group, PromptLayer experienced the following error: ${e}`\n );\n return false;\n }\n return true;\n};\n\nconst promptLayerCreateGroup = async (\n apiKey: string,\n baseURL: string,\n throwOnError: boolean = true\n): Promise<number | boolean> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/create-group`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify({\n api_key: apiKey,\n }),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to create group\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While creating a group PromptLayer had the following error\"\n );\n return false;\n }\n }\n return data.id;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While creating a group PromptLayer had the following error: ${e}`\n );\n return false;\n }\n};\n\nconst getPromptTemplate = async (\n apiKey: string,\n baseURL: string,\n promptName: string,\n params?: Partial<GetPromptTemplateParams>,\n throwOnError: boolean = true\n): Promise<GetPromptTemplateResponse | null> => {\n try {\n const url = new URL(`${baseURL}/prompt-templates/${promptName}`);\n const response = await fetchWithRetry(url, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n \"X-API-KEY\": apiKey,\n ...getCommonHeaders(),\n },\n body: JSON.stringify(params),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to fetch prompt template\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n console.warn(\n `WARNING: While fetching a prompt template PromptLayer had the following error: ${errorMessage}`\n );\n return null;\n }\n }\n if (data.warning) {\n console.warn(\n `WARNING: While fetching your prompt PromptLayer had the following error: ${data.warning}`\n );\n }\n return data as GetPromptTemplateResponse;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While fetching a prompt template PromptLayer had the following error: ${e}`\n );\n return null;\n }\n};\n\nconst publishPromptTemplate = async (\n apiKey: string,\n baseURL: string,\n body: PublishPromptTemplate,\n throwOnError: boolean = true\n): Promise<PublishPromptTemplateResponse> => {\n const response = await fetchWithRetry(`${baseURL}/rest/prompt-templates`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", \"X-API-KEY\": apiKey, ...getCommonHeaders() },\n body: JSON.stringify({\n prompt_template: { ...body },\n prompt_version: { ...body },\n release_labels: body.release_labels ? body.release_labels : undefined,\n }),\n });\n const data = await response.json();\n if (response.status !== 200 && response.status !== 201) {\n const errorMessage =\n data.message || data.error || \"Failed to publish prompt template\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While publishing a prompt template PromptLayer had the following error\"\n );\n }\n }\n return data as PublishPromptTemplateResponse;\n};\n\nconst getAllPromptTemplates = async (\n apiKey: string,\n baseURL: string,\n params?: Partial<Pagination>,\n throwOnError: boolean = true\n): Promise<Array<ListPromptTemplatesResponse>> => {\n const url = new URL(`${baseURL}/prompt-templates`);\n Object.entries(params || {}).forEach(([key, value]) =>\n url.searchParams.append(key, value.toString())\n );\n const response = await fetchWithRetry(url, {\n headers: { \"Content-Type\": \"application/json\", \"X-API-KEY\": apiKey, ...getCommonHeaders() },\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to fetch prompt templates\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While fetching all prompt templates PromptLayer had the following error\"\n );\n return [];\n }\n }\n return (data.items ?? []) as Array<ListPromptTemplatesResponse>;\n};\n\nconst waitForWorkflowCompletionCentrifugo = async (\n params: WaitForWorkflowCompletionParams\n): Promise<any> => {\n const url = new URL(`${params.baseURL}/connection/websocket`);\n url.protocol = url.protocol === \"https:\" ? \"wss:\" : \"ws:\";\n\n const client = new Centrifuge(url.toString(), { token: params.token });\n const sub = client.newSubscription(params.channelName);\n\n return new Promise((resolve, reject) => {\n const cleanupWithResolve = (data: any) => {\n cleanup();\n resolve(data);\n };\n\n const listener = makeMessageListener(\n params.baseURL,\n { resolve: cleanupWithResolve, reject },\n params.executionId,\n params.returnAllOutputs,\n params.headers\n );\n\n sub.on(\"publication\", (message) => {\n listener({\n name: message.data.message_name,\n data: message.data.data,\n });\n });\n\n const timeout = setTimeout(() => {\n reject(\n new Error(\"Workflow execution did not complete properly (timeout)\")\n );\n }, params.timeout);\n\n const cleanup = () => {\n clearTimeout(timeout);\n sub.unsubscribe();\n client.disconnect();\n };\n\n sub.on(\"error\", (err) => {\n cleanup();\n reject(`Centrifugo subscription error: ${err}`);\n });\n\n client.on(\"error\", (err) => {\n cleanup();\n reject(`Centrifugo client error: ${err}`);\n });\n\n sub.subscribe();\n client.connect();\n });\n};\n\nexport const runWorkflowRequest = async ({\n workflow_name,\n input_variables,\n metadata = {},\n workflow_label_name = null,\n workflow_version_number = null,\n return_all_outputs = false,\n api_key,\n timeout = 3600000, // Default timeout is 1 hour in milliseconds\n baseURL,\n}: RunWorkflowRequestParams): Promise<WorkflowResponse> => {\n const payload = {\n input_variables,\n metadata,\n workflow_label_name,\n workflow_version_number,\n return_all_outputs,\n };\n\n const headers = { \"Content-Type\": \"application/json\", \"X-API-KEY\": api_key, ...getCommonHeaders() };\n\n try {\n const response = await fetchWithRetry(\n `${baseURL}/workflows/${encodeURIComponent(workflow_name)}/run`,\n {\n method: \"POST\",\n headers: headers,\n body: JSON.stringify(payload),\n }\n );\n\n if (response.status !== 201) {\n const errorData = await response.json().catch(() => ({}));\n return {\n success: false,\n message: `Failed to run workflow: ${\n errorData.error || response.statusText\n }`,\n };\n }\n\n const result = await response.json();\n if (result.warning) {\n console.warn(`WARNING: ${result.warning}`);\n }\n const execution_id = result.workflow_version_execution_id;\n if (!execution_id) {\n console.log(\"No execution ID returned from workflow run\");\n return { success: false, message: \"Failed to run workflow\" };\n }\n\n const channel_name = `workflow_updates:${execution_id}`;\n const ws_response = await fetchWithRetry(\n `${baseURL}/ws-token-request-library?capability=${channel_name}`,\n {\n method: \"POST\",\n headers: headers,\n }\n );\n\n const ws_token_response = await ws_response.json();\n const token = ws_token_response.token_details.token;\n\n const params: WaitForWorkflowCompletionParams = {\n token,\n channelName: channel_name,\n executionId: execution_id,\n returnAllOutputs: return_all_outputs,\n headers: headers,\n timeout: timeout,\n baseURL: baseURL,\n };\n if (ws_token_response.messaging_backend === \"centrifugo\")\n return waitForWorkflowCompletionCentrifugo(params);\n return await waitForWorkflowCompletion(params);\n } catch (error) {\n console.error(\n `Failed to run workflow: ${\n error instanceof Error ? error.message : error\n }`\n );\n throw error;\n }\n};\n\nasync function* proxyGenerator<Item>(\n apiKey: string,\n baseURL: string,\n generator: AsyncIterable<Item>,\n body: TrackRequest,\n throwOnError: boolean = true\n) {\n const results = [];\n for await (const value of generator) {\n yield body.return_pl_id ? [value, null] : value;\n results.push(value);\n }\n const request_response = cleaned_result(results, body.function_name);\n const response = await promptLayerApiRequest(\n apiKey,\n baseURL,\n {\n ...body,\n request_response,\n request_end_time: new Date().toISOString(),\n },\n throwOnError\n );\n if (response) {\n if (body.return_pl_id) {\n const request_id = (response as any)[1];\n const lastResult = results.at(-1);\n yield [lastResult, request_id];\n }\n }\n}\n\nconst warnOnBadResponse = (request_response: any, main_message: string) => {\n try {\n console.warn(`${main_message}: ${request_response.message}`);\n } catch (e) {\n console.warn(`${main_message}: ${request_response}`);\n }\n};\n\nconst trackRequest = async (\n baseURL: string,\n body: TrackRequest,\n throwOnError: boolean = true\n) => {\n try {\n const response = await fetchWithRetry(`${baseURL}/track-request`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", ...getCommonHeaders() },\n body: JSON.stringify(body),\n });\n const data = await response.json();\n if (response.status !== 200) {\n const errorMessage =\n data.message || data.error || \"Failed to track request\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging your request, PromptLayer experienced the following error:\"\n );\n }\n }\n return data;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While logging your request PromptLayer had the following error: ${e}`\n );\n }\n return {};\n};\n\nconst openaiChatRequest = async (client: TypeOpenAI, kwargs: any) => {\n return await client.chat.completions.create(kwargs);\n};\n\nconst openaiCompletionsRequest = async (client: TypeOpenAI, kwargs: any) => {\n return await client.completions.create(kwargs);\n};\n\nconst MAP_TYPE_TO_OPENAI_FUNCTION = {\n chat: openaiChatRequest,\n completion: openaiCompletionsRequest,\n};\n\nconst openaiRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const OpenAI = require(\"openai\").default;\n const client = new OpenAI({\n baseURL: kwargs.baseURL,\n apiKey: kwargs.apiKey,\n });\n\n delete kwargs?.apiKey;\n delete kwargs?.baseURL;\n\n const api_type = promptBlueprint.metadata?.model?.api_type;\n if (api_type === \"chat-completions\") {\n const requestToMake =\n MAP_TYPE_TO_OPENAI_FUNCTION[promptBlueprint.prompt_template.type];\n return await requestToMake(client, kwargs);\n } else if (api_type === \"images\") {\n return await client.images.generate(kwargs);\n } else {\n return await client.responses.create(kwargs);\n }\n};\n\nconst azureOpenAIRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { AzureOpenAI } = require(\"openai\");\n const client = new AzureOpenAI({\n endpoint: process.env.AZURE_OPENAI_ENDPOINT || kwargs.baseURL,\n apiVersion: process.env.OPENAI_API_VERSION || kwargs.apiVersion,\n apiKey: process.env.AZURE_OPENAI_API_KEY || kwargs.apiKey,\n });\n delete kwargs?.baseURL;\n delete kwargs?.apiVersion;\n delete kwargs?.apiKey;\n\n const api_type = promptBlueprint.metadata?.model?.api_type;\n\n if (api_type === \"chat-completions\") {\n const requestToMake = MAP_TYPE_TO_OPENAI_FUNCTION[promptBlueprint.prompt_template.type];\n return await requestToMake(client, kwargs);\n } else if (api_type === \"images\") {\n return await client.images.generate(kwargs);\n } else {\n return await client.responses.create(kwargs);\n }\n};\n\nconst anthropicChatRequest = async (\n client: TypeAnthropic | AnthropicVertex | AnthropicBedrock,\n kwargs: any\n) => {\n return client.messages.create(kwargs);\n};\n\nconst anthropicCompletionsRequest = async (\n client: TypeAnthropic | AnthropicBedrock,\n kwargs: any\n) => {\n return client.completions.create(kwargs);\n};\n\nconst MAP_TYPE_TO_ANTHROPIC_FUNCTION = {\n chat: anthropicChatRequest,\n completion: anthropicCompletionsRequest,\n};\n\nconst anthropicRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const Anthropic = require(\"@anthropic-ai/sdk\").default;\n const client = new Anthropic({\n baseURL: kwargs.baseURL,\n apiKey: kwargs.apiKey,\n });\n const requestToMake =\n MAP_TYPE_TO_ANTHROPIC_FUNCTION[promptBlueprint.prompt_template.type];\n return requestToMake(client, kwargs);\n};\n\nconst utilLogRequest = async (\n apiKey: string,\n baseURL: string,\n body: LogRequest,\n throwOnError: boolean = true\n): Promise<RequestLog | null> => {\n try {\n const response = await fetchWithRetry(`${baseURL}/log-request`, {\n method: \"POST\",\n headers: { \"Content-Type\": \"application/json\", \"X-API-KEY\": apiKey, ...getCommonHeaders() },\n body: JSON.stringify(body),\n });\n const data = await response.json();\n if (response.status !== 201) {\n const errorMessage =\n data.message || data.error || \"Failed to log request\";\n if (throwOnError) {\n throw new Error(errorMessage);\n } else {\n warnOnBadResponse(\n data,\n \"WARNING: While logging your request PromptLayer had the following error\"\n );\n return null;\n }\n }\n return data;\n } catch (e) {\n if (throwOnError) {\n throw e;\n }\n console.warn(\n `WARNING: While tracking your prompt PromptLayer had the following error: ${e}`\n );\n return null;\n }\n};\n\nconst googleChatRequest = async (model_client: any, kwargs: any) => {\n const history = kwargs?.history;\n const generationConfig = kwargs?.generationConfig;\n const lastMessage =\n history.length > 0 ? history[history.length - 1]?.parts : \"\";\n const chat = model_client.chats.create({\n model: kwargs?.model,\n history: history.slice(0, -1) ?? [],\n config: generationConfig,\n });\n\n if (kwargs?.stream)\n return await chat.sendMessageStream({ message: lastMessage });\n return await chat.sendMessage({ message: lastMessage });\n};\n\nconst googleCompletionsRequest = async (\n model_client: any,\n { stream, ...kwargs }: any\n) => {\n if (stream) return await model_client.generateContentStream({ ...kwargs });\n return await model_client.generateContent({ ...kwargs });\n};\n\nconst MAP_TYPE_TO_GOOGLE_FUNCTION = {\n chat: googleChatRequest,\n completion: googleCompletionsRequest,\n};\n\nconst googleRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { GoogleGenAI } = await import(\"@google/genai\");\n\n const geminiAPI = process.env.GOOGLE_API_KEY || process.env.GEMINI_API_KEY;\n const project =\n process.env.VERTEX_AI_PROJECT_ID ||\n process.env.GOOGLE_PROJECT_ID ||\n process.env.GOOGLE_CLOUD_PROJECT;\n const location =\n process.env.VERTEX_AI_PROJECT_LOCATION ||\n process.env.GOOGLE_PROJECT_LOCATION ||\n process.env.GOOGLE_CLOUD_PROJECT_LOCATION;\n const googleAuthOptions = {\n keyFilename: process.env.GOOGLE_APPLICATION_CREDENTIALS,\n projectId: project,\n scopes: [\"https://www.googleapis.com/auth/cloud-platform\"],\n };\n\n const genAI = geminiAPI\n ? new GoogleGenAI({ apiKey: geminiAPI })\n : new GoogleGenAI({\n vertexai: true,\n project: project,\n location: location,\n googleAuthOptions,\n });\n const requestToMake =\n MAP_TYPE_TO_GOOGLE_FUNCTION[promptBlueprint.prompt_template.type];\n\n return await requestToMake(genAI, kwargs);\n};\n\nconst snakeToCamel = (str: string): string =>\n str.replace(/_([a-z])/g, (_, letter) => letter.toUpperCase());\n\nconst convertKeysToCamelCase = <T>(\n obj: T,\n ignoreValuesWithKeys: Set<string> = new Set()\n): T => {\n if (!obj || typeof obj !== \"object\") return obj;\n if (Array.isArray(obj))\n return obj.map((item) =>\n convertKeysToCamelCase(item, ignoreValuesWithKeys)\n ) as T;\n\n return Object.fromEntries(\n Object.entries(obj).map(([key, value]) => {\n if (ignoreValuesWithKeys.has(key)) {\n return [snakeToCamel(key), value];\n }\n return [\n snakeToCamel(key),\n convertKeysToCamelCase(value, ignoreValuesWithKeys),\n ];\n })\n ) as T;\n};\n\nconst configureProviderSettings = (\n promptBlueprint: any,\n customProvider: any,\n modelParameterOverrides: any = {},\n stream: boolean = false\n) => {\n const provider_type =\n customProvider?.client ?? promptBlueprint.metadata?.model?.provider;\n const api_type = promptBlueprint.metadata?.model?.api_type;\n\n if (!provider_type) {\n throw new Error(\n \"Provider type not found in prompt blueprint or custom provider\"\n );\n }\n\n let kwargs = {\n ...(promptBlueprint.llm_kwargs || {}),\n stream,\n };\n\n if (\n [\"google\", \"vertexai\"].includes(provider_type) &&\n promptBlueprint.metadata?.model?.name.startsWith(\"gemini\")\n )\n kwargs = convertKeysToCamelCase(\n kwargs,\n new Set([\"function_declarations\", \"properties\"])\n );\n\n const providerConfig = {\n baseURL: customProvider?.base_url ?? promptBlueprint.provider_base_url?.url,\n apiKey: customProvider?.api_key,\n };\n\n Object.entries(providerConfig).forEach(([key, value]) => {\n if (value !== undefined) {\n kwargs[key] = value;\n }\n });\n\n if (stream && STREAMING_PROVIDERS_WITH_USAGE.includes(provider_type as any) && api_type === \"chat-completions\") {\n kwargs.stream_options = { include_usage: true };\n }\n\n return { provider_type, kwargs };\n};\n\nconst getProviderConfig = (provider_type: string, promptTemplate: any) => {\n const providerMap =\n MAP_PROVIDER_TO_FUNCTION_NAME[\n provider_type as keyof typeof MAP_PROVIDER_TO_FUNCTION_NAME\n ];\n\n if (!providerMap) {\n throw new Error(`Unsupported provider type: ${provider_type}`);\n }\n\n const templateType = promptTemplate.type as keyof typeof providerMap;\n const config = providerMap[templateType];\n\n if (!config) {\n throw new Error(\n `Unsupported template type '${promptTemplate.type}' for provider '${provider_type}'`\n );\n }\n\n return config;\n};\n\nconst vertexaiRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const model = promptBlueprint.metadata?.model;\n if (!model) throw new Error(\"Model metadata not found in prompt blueprint\");\n if (model.name.startsWith(\"gemini\"))\n return googleRequest(promptBlueprint, kwargs);\n if (model.name.startsWith(\"claude\")) {\n const { AnthropicVertex } = await import(\"@anthropic-ai/vertex-sdk\");\n const client = new AnthropicVertex({ baseURL: kwargs.baseURL });\n if (promptBlueprint.prompt_template.type === \"chat\")\n return anthropicChatRequest(client, kwargs);\n throw new Error(\n `Unsupported prompt template type '${promptBlueprint.prompt_template.type}' for Anthropic Vertex AI`\n );\n }\n throw new Error(\n `Unsupported model name '${model.name}' for Vertex AI request`\n );\n};\n\nconst amazonBedrockRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { BedrockRuntimeClient, ConverseCommand, ConverseStreamCommand } =\n await import(\"@aws-sdk/client-bedrock-runtime\");\n const client = new BedrockRuntimeClient({\n credentials: {\n accessKeyId: kwargs?.aws_access_key || process.env.AWS_ACCESS_KEY_ID,\n secretAccessKey:\n kwargs?.aws_secret_key || process.env.AWS_SECRET_ACCESS_KEY,\n sessionToken: kwargs?.aws_session_token || process.env.AWS_SESSION_TOKEN,\n },\n region:\n kwargs?.aws_region ||\n process.env.AWS_REGION ||\n process.env.AWS_DEFAULT_REGION ||\n \"us-east-1\",\n });\n\n if (kwargs?.stream) {\n delete kwargs.stream;\n const command = new ConverseStreamCommand({\n ...kwargs,\n });\n return await client.send(command);\n } else {\n delete kwargs?.stream;\n const command = new ConverseCommand({\n ...kwargs,\n });\n return await client.send(command);\n }\n};\n\nconst anthropicBedrockRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { AnthropicBedrock } = await import(\"@anthropic-ai/bedrock-sdk\");\n const client = new AnthropicBedrock({\n awsAccessKey: kwargs.aws_access_key,\n awsSecretKey: kwargs.aws_secret_key,\n awsRegion: kwargs.aws_region,\n awsSessionToken: kwargs.aws_session_token,\n baseURL: kwargs.base_url,\n });\n\n const requestToMake =\n MAP_TYPE_TO_ANTHROPIC_FUNCTION[promptBlueprint.prompt_template.type];\n return requestToMake(client, kwargs);\n};\n\nconst mistralRequest = async (\n promptBlueprint: GetPromptTemplateResponse,\n kwargs: any\n) => {\n const { Mistral } = await import(\"@mistralai/mistralai\");\n const client = new Mistral({ apiKey: process.env.MISTRAL_API_KEY });\n kwargs = convertKeysToCamelCase(kwargs, new Set());\n if (kwargs?.stream) {\n delete kwargs.stream;\n return await client.chat.stream(kwargs);\n }\n delete kwargs.stream;\n return await client.chat.complete(kwargs);\n};\n\nexport const readEnv = (env: string): string | undefined => {\n if (typeof (globalThis as any).process !== \"undefined\")\n return (globalThis as any).process.env?.[env]?.trim() ?? undefined;\n\n if (typeof (globalThis as any).Deno !== \"undefined\")\n return (globalThis as any).Deno.env?.get?.(env)?.trim();\n return undefined;\n};\n\nexport const requirePromptLayerApiKey = (apiKey?: string): string => {\n if (apiKey !== undefined) {\n const explicitApiKey = apiKey.trim();\n if (!explicitApiKey) {\n throw new Error(\n \"PromptLayer API key not provided. Please set PROMPTLAYER_API_KEY or pass apiKey.\"\n );\n }\n return explicitApiKey;\n }\n\n const envApiKey = readEnv(\"PROMPTLAYER_API_KEY\");\n if (!envApiKey) {\n throw new Error(\n \"PromptLayer API key not provided. Please set PROMPTLAYER_API_KEY or pass apiKey.\"\n );\n }\n\n return envApiKey;\n};\n\nexport {\n amazonBedrockRequest,\n anthropicBedrockRequest,\n anthropicRequest,\n azureOpenAIRequest,\n configureProviderSettings,\n getAllPromptTemplates,\n getPromptTemplate,\n getProviderConfig,\n googleRequest,\n mistralRequest,\n openaiRequest,\n promptlayerApiHandler,\n promptLayerApiRequest,\n promptLayerCreateGroup,\n promptLayerTrackGroup,\n promptLayerTrackMetadata,\n promptLayerTrackPrompt,\n promptLayerTrackScore,\n publishPromptTemplate,\n trackRequest,\n utilLogRequest,\n vertexaiRequest,\n warnOnBadResponse,\n};\n"],"mappings":"0jBAAA,IAAAA,EAAA,GAAAC,EAAAD,EAAA,qBAAAE,IAAA,eAAAC,EAAAH,GCAA,IAAAI,EAAe,iBACfC,EAAiB,mBACjBC,EAA8B,eCkB9B,IAAAC,EAAiB,mBACjBC,EAA2B,sBAE3BC,EAAmB,sBASZ,IAAMC,EAAc,QAGrBC,EAAiB,IAAc,CAnCrC,IAAAC,EAoCE,GAAI,OAAO,SAAY,eAAeA,EAAA,QAAQ,WAAR,MAAAA,EAAkB,MAAM,CAC5D,IAAMC,EAAQ,QAAQ,SAAS,KAAK,MAAM,GAAG,EAC7C,MAAO,GAAGA,EAAM,CAAC,CAAC,IAAIA,EAAM,CAAC,CAAC,EAChC,CACA,MAAO,SACT,EAEMC,EAAgBH,EAAe,EAC/BI,GAA0B,kBAAkBL,CAAW,UAAUI,CAAa,IA+nC7E,IAAME,EAAWC,GAAoC,CA3qC5D,IAAAC,EAAAC,EAAAC,EAAAC,EAAAC,EAAAC,EA4qCE,GAAI,OAAQ,WAAmB,SAAY,YACzC,OAAQH,GAAAD,GAAAD,EAAA,WAAmB,QAAQ,MAA3B,YAAAA,EAAiCD,KAAjC,YAAAE,EAAuC,SAAvC,KAAAC,EAAiD,OAE3D,GAAI,OAAQ,WAAmB,MAAS,YACtC,OAAQG,GAAAD,GAAAD,EAAA,WAAmB,KAAK,MAAxB,YAAAA,EAA6B,MAA7B,YAAAC,EAAA,KAAAD,EAAmCJ,KAAnC,YAAAM,EAAyC,MAErD,EAEaC,EAA4BC,GAA4B,CACnE,GAAIA,IAAW,OAAW,CACxB,IAAMC,EAAiBD,EAAO,KAAK,EACnC,GAAI,CAACC,EACH,MAAM,IAAI,MACR,kFACF,EAEF,OAAOA,CACT,CAEA,IAAMC,EAAYX,EAAQ,qBAAqB,EAC/C,GAAI,CAACW,EACH,MAAM,IAAI,MACR,kFACF,EAGF,OAAOA,CACT,ED7qCA,IAAMC,EAAwB,CAC5B,6BACA,WACA,mBACA,eACA,yBACA,8BACA,yBACA,qBACA,uBACA,uBACA,kBACA,sBACA,uBACA,mBACA,uBACA,oBACA,0BACA,yBACF,EAEMC,EAAgC,EAAAC,QAAK,KAAK,SAAU,gBAAiB,OAAO,EAElF,SAASC,GAA2B,CAClC,IAAMC,EAA4B,MAAM,kBAExC,GAAI,CACF,MAAM,kBAAoB,CAACC,EAAQC,IAAUA,EAC7C,IAAMC,EAAY,IAAI,MAAM,EAAE,MACxBC,EAAcD,GAAA,YAAAA,EAChB,IAAKE,GAAaA,EAAS,YAAY,GACxC,KAAMC,GAAaA,GAAY,CAACA,EAAS,WAAW,OAAO,GAE9D,GAAI,CAACF,EACH,MAAM,IAAI,MACR,wEACF,EAGF,OAAO,EAAAN,QAAK,QACVM,EAAY,WAAW,SAAS,KAAI,iBAAcA,CAAW,EAAIA,CACnE,CACF,QAAE,CACA,MAAM,kBAAoBJ,CAC5B,CACF,CAEA,SAASO,GAAgC,CACvC,GAAI,QAAQ,WAAa,QACvB,MAAM,IAAI,MACR,qFACF,CAEJ,CAEA,SAASC,GAAiC,CACxC,IAAMC,EAAYV,EAAiB,EAQnC,MAAO,CACL,EAAAD,QAAK,QAAQW,EAAW,KAAM,KAAM,KAAMZ,CAA6B,EACvE,EAAAC,QAAK,QAAQW,EAAW,KAAMZ,CAA6B,EAC3D,EAAAC,QAAK,QAAQW,EAAW,KAAM,KAAMZ,CAA6B,CACnE,CACF,CAEA,SAASa,EAA4BC,EAA0B,CAC7D,IAAMC,EAAehB,EAAsB,OACxCiB,GAAiB,CAAC,EAAAC,QAAG,WAAW,EAAAhB,QAAK,KAAKa,EAAYE,CAAY,CAAC,CACtE,EAEA,GAAID,EAAa,OAAS,EACxB,MAAM,IAAI,MACR,qEAAqEA,EAAa,KAAK,IAAI,CAAC,EAC9F,CAEJ,CAEA,SAASG,GAA4B,CACnC,QAAWC,KAAaR,EAAqB,EAM3C,GALI,GAAC,EAAAM,QAAG,WAAWE,CAAS,GAKxB,CADU,EAAAF,QAAG,SAASE,CAAS,EACxB,YAAY,GAIvB,OAAAN,EAA4BM,CAAS,EAC9BA,EAGT,MAAM,IAAI,MACR,mFACF,CACF,CAEO,SAASC,EACdC,EAAkC,CAAC,EACJ,CAnIjC,IAAAC,EAoIEZ,EAAwB,EAExB,IAAMa,EAAaL,EAAkB,EAC/BM,EAASC,EAAyBJ,EAAQ,MAAM,EAChDK,GAAcJ,EAAAD,EAAQ,cAAR,YAAAC,EAAqB,OAEnCK,EAAkC,CACtC,qBAAsB,OACtB,oBAAqBH,CACvB,EAEA,OAAIE,IACFC,EAAI,wBAA0BD,GAGzB,CACL,OAAQ,CACN,KAAM,QACN,KAAMH,CACR,EACA,IAAAI,CACF,CACF","names":["claude_agents_exports","__export","getClaudeConfig","__toCommonJS","import_fs","import_path","import_url","import_ably","import_centrifuge","import_p_retry","SDK_VERSION","getNodeVersion","_a","parts","_NODE_VERSION","_PROMPTLAYER_USER_AGENT","readEnv","env","_a","_b","_c","_d","_e","_f","requirePromptLayerApiKey","apiKey","explicitApiKey","envApiKey","REQUIRED_PLUGIN_FILES","VENDORED_PLUGIN_RELATIVE_PATH","path","currentModuleDir","originalPrepareStackTrace","_error","stack","callsites","currentFile","callsite","fileName","assertSupportedPlatform","candidatePluginRoots","moduleDir","assertRequiredVendoredFiles","pluginRoot","missingFiles","relativePath","fs","resolvePluginRoot","candidate","getClaudeConfig","options","_a","pluginPath","apiKey","requirePromptLayerApiKey","traceparent","env"]}
|