@lelemondev/sdk 0.9.2 → 0.9.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +148 -15
- package/dist/anthropic.d.mts +2 -2
- package/dist/anthropic.d.ts +2 -2
- package/dist/anthropic.js +1 -1
- package/dist/anthropic.mjs +1 -1
- package/dist/bedrock.d.mts +2 -2
- package/dist/bedrock.d.ts +2 -2
- package/dist/bedrock.js +1 -1
- package/dist/bedrock.mjs +1 -1
- package/dist/capture-BJnZrkYl.d.mts +264 -0
- package/dist/capture-BJnZrkYl.d.ts +264 -0
- package/dist/gemini.d.mts +2 -2
- package/dist/gemini.d.ts +2 -2
- package/dist/gemini.js +2 -2
- package/dist/gemini.mjs +1 -1
- package/dist/index.d.mts +21 -245
- package/dist/index.d.ts +21 -245
- package/dist/index.js +2 -2
- package/dist/index.mjs +2 -2
- package/dist/openai.d.mts +2 -2
- package/dist/openai.d.ts +2 -2
- package/dist/openai.js +1 -1
- package/dist/openai.mjs +1 -1
- package/dist/openrouter.d.mts +2 -2
- package/dist/openrouter.d.ts +2 -2
- package/dist/openrouter.js +1 -1
- package/dist/openrouter.mjs +1 -1
- package/package.json +5 -2
- package/dist/anthropic.js.map +0 -1
- package/dist/anthropic.mjs.map +0 -1
- package/dist/bedrock.js.map +0 -1
- package/dist/bedrock.mjs.map +0 -1
- package/dist/express.js.map +0 -1
- package/dist/express.mjs.map +0 -1
- package/dist/gemini.js.map +0 -1
- package/dist/gemini.mjs.map +0 -1
- package/dist/hono.js.map +0 -1
- package/dist/hono.mjs.map +0 -1
- package/dist/index.js.map +0 -1
- package/dist/index.mjs.map +0 -1
- package/dist/integrations.js.map +0 -1
- package/dist/integrations.mjs.map +0 -1
- package/dist/lambda.js.map +0 -1
- package/dist/lambda.mjs.map +0 -1
- package/dist/next.js.map +0 -1
- package/dist/next.mjs.map +0 -1
- package/dist/openai.js.map +0 -1
- package/dist/openai.mjs.map +0 -1
- package/dist/openrouter.js.map +0 -1
- package/dist/openrouter.mjs.map +0 -1
package/dist/index.mjs.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../package.json","../src/core/logger.ts","../src/core/transport.ts","../src/core/telemetry.ts","../src/core/config.ts","../src/core/capture.ts","../src/core/context.ts"],"names":["require_package","__commonJSMin","exports","module","debugEnabled","setDebug","enabled","isDebugEnabled","getEnvVar","PREFIX","debug","message","data","logWithPrefix","info","warn","traceCaptureError","provider","err","batchSend","count","endpoint","batchSuccess","durationMs","batchError","requestDetails","method","url","bodySize","responseDetails","status","level","logFn","name","DEFAULT_BATCH_SIZE","DEFAULT_FLUSH_INTERVAL_MS","DEFAULT_REQUEST_TIMEOUT_MS","Transport","config","__publicField","trace","items","startTime","error","path","body","controller","bodyStr","timeoutId","response","errorText","text","SDK_NAME","SDK_LANGUAGE","detectRuntime","detectOS","platform","ua","getSDKVersion","__require","cachedTelemetry","buildTelemetry","service","runtime","os","telemetry","globalConfig","globalTransport","globalTelemetry","DEFAULT_ENDPOINT","init","createTransport","getTelemetry","isEnabled","getTransport","flush","apiKey","globalContext","getGlobalContext","captureSpan","options","transport","traceContext","getTraceContext","metadataTraceId","metadataParentSpanId","cleanMetadata","request","sanitizeInput","sanitize","generateId","MAX_STRING_LENGTH","SENSITIVE_KEYS","SENSITIVE_TOKEN_PATTERNS","SAFE_TOKEN_KEYS","isSensitiveKey","key","lowerKey","k","input","value","depth","item","sanitized","val","traceStorage","AsyncLocalStorage","getToolParentSpanId","toolCallId","ctx","clearToolCall","nameOrOptions","fn","parentContext","traceId","rootSpanId","context","result","e","sendRootSpan","output","rootSpan","span","parentSpanId"],"mappings":";ieAAA,IAAAA,CAAAA,CAAAC,CAAAA,CAAA,CAAAC,EAAAA,CAAAC,EAAAA,GAAA,CAAAA,EAAAA,CAAA,OAAA,CAAA,CACE,IAAA,CAAQ,iBAAA,CACR,OAAA,CAAW,OAAA,CACX,WAAA,CAAe,sEAAA,CACf,MAAA,CAAU,4BAAA,CACV,OAAA,CAAW,KAAA,CACX,UAAA,CAAc,CACZ,IAAA,CAAQ,KAAA,CACR,GAAA,CAAO,sDACT,CAAA,CACA,QAAA,CAAY,qBAAA,CACZ,IAAA,CAAQ,CACN,IAAO,qDACT,CAAA,CACA,QAAA,CAAY,CACV,KAAA,CACA,eAAA,CACA,SAAA,CACA,QAAA,CACA,WAAA,CACA,QAAA,CACA,QAAA,CACA,SAAA,CACA,MAAA,CACA,QAAA,CACA,KAAA,CACA,IAAA,CACA,YAAA,CACA,YACF,CAAA,CACA,IAAA,CAAQ,iBAAA,CACR,MAAA,CAAU,kBAAA,CACV,KAAA,CAAS,mBAAA,CACT,OAAA,CAAW,CACT,GAAA,CAAK,CACH,KAAA,CAAS,oBACT,MAAA,CAAU,kBAAA,CACV,OAAA,CAAW,iBACb,CAAA,CACA,UAAA,CAAY,CACV,KAAA,CAAS,oBAAA,CACT,MAAA,CAAU,mBAAA,CACV,OAAA,CAAW,kBACb,CAAA,CACA,aAAA,CAAe,CACb,KAAA,CAAS,uBAAA,CACT,MAAA,CAAU,sBAAA,CACV,OAAA,CAAW,qBACb,CAAA,CACA,WAAA,CAAa,CACX,KAAA,CAAS,qBAAA,CACT,MAAA,CAAU,oBAAA,CACV,OAAA,CAAW,mBACb,CAAA,CACA,UAAA,CAAY,CACV,KAAA,CAAS,oBAAA,CACT,MAAA,CAAU,mBAAA,CACV,OAAA,CAAW,kBACb,CAAA,CACA,cAAA,CAAgB,CACd,KAAA,CAAS,wBAAA,CACT,MAAA,CAAU,uBAAA,CACV,OAAA,CAAW,sBACb,CAAA,CACA,QAAA,CAAU,CACR,KAAA,CAAS,kBAAA,CACT,MAAA,CAAU,iBAAA,CACV,OAAA,CAAW,gBACb,CAAA,CACA,UAAA,CAAY,CACV,KAAA,CAAS,qBACT,MAAA,CAAU,mBAAA,CACV,OAAA,CAAW,kBACb,CAAA,CACA,WAAA,CAAa,CACX,KAAA,CAAS,qBAAA,CACT,MAAA,CAAU,oBAAA,CACV,OAAA,CAAW,mBACb,CAAA,CACA,QAAA,CAAU,CACR,KAAA,CAAS,kBAAA,CACT,MAAA,CAAU,iBAAA,CACV,OAAA,CAAW,gBACb,CAAA,CACA,gBAAA,CAAkB,CAChB,KAAA,CAAS,0BAAA,CACT,MAAA,CAAU,yBAAA,CACV,OAAA,CAAW,wBACb,CAAA,CACA,gBAAA,CAAkB,gBACpB,CAAA,CACA,aAAA,CAAiB,CACf,GAAA,CAAK,CACH,MAAA,CAAU,CACR,oBACF,CAAA,CACA,SAAA,CAAa,CACX,uBACF,CAAA,CACA,OAAA,CAAW,CACT,qBACF,CAAA,CACA,MAAA,CAAU,CACR,oBACF,CAAA,CACA,UAAA,CAAc,CACZ,wBACF,CAAA,CACA,IAAA,CAAQ,CACN,kBACF,CAAA,CACA,MAAA,CAAU,CACR,oBACF,CAAA,CACA,OAAA,CAAW,CACT,qBACF,CAAA,CACA,IAAA,CAAQ,CACN,kBACF,CAAA,CACA,YAAA,CAAgB,CACd,0BACF,CAAA,CACA,GAAA,CAAK,CACH,mBACF,CACF,CACF,CAAA,CACA,KAAA,CAAS,CACP,MAAA,CACA,WACF,CAAA,CACA,WAAA,CAAe,KAAA,CACf,OAAA,CAAW,CACT,IAAA,CAAQ,UACV,CAAA,CACA,OAAA,CAAW,CACT,KAAA,CAAS,MAAA,CACT,GAAA,CAAO,cAAA,CACP,IAAA,CAAQ,+CAAA,CACR,cAAA,CAAkB,eAAA,CAClB,IAAA,CAAQ,aAAA,CACR,IAAA,CAAQ,QAAA,CACR,UAAA,CAAY,YAAA,CACZ,eAAA,CAAiB,uBAAA,CACjB,UAAA,CAAY,sBAAA,CACZ,SAAA,CAAa,cACf,CAAA,CACA,eAAA,CAAmB,CACjB,iCAAA,CAAmC,UAAA,CACnC,wBAAyB,SAAA,CACzB,aAAA,CAAe,SAAA,CACf,qBAAA,CAAuB,QAAA,CACvB,MAAA,CAAU,SAAA,CACV,MAAA,CAAU,SAAA,CACV,IAAA,CAAQ,QAAA,CACR,OAAA,CAAW,UAAA,CACX,UAAA,CAAc,QAAA,CACd,MAAA,CAAU,QACZ,CACF,EAAA,CAAA,CAAA,CCrJA,IAAIC,CAAAA,CAAe,KAAA,CASZ,SAASC,CAAAA,CAASC,CAAAA,CAAwB,CAC/CF,CAAAA,CAAeE,EACjB,CAKO,SAASC,GAA0B,CAExC,OAAIH,CAAAA,CAAqB,IAAA,CAClBI,CAAAA,CAAU,eAAe,CAAA,GAAM,MACxC,CAMA,IAAMC,CAAAA,CAAS,WAAA,CAKR,SAASC,CAAAA,CAAMC,CAAAA,CAAiBC,CAAAA,CAAsB,CACtDL,CAAAA,EAAe,EACpBM,CAAAA,CAAc,OAAA,CAASF,CAAAA,CAASC,CAAI,EACtC,CAKO,SAASE,CAAAA,CAAKH,CAAAA,CAAiBC,CAAAA,CAAsB,CACrDL,GAAe,EACpBM,CAAAA,CAAc,MAAA,CAAQF,CAAAA,CAASC,CAAI,EACrC,CAKO,SAASG,CAAAA,CAAKJ,CAAAA,CAAiBC,CAAAA,CAAsB,CAC1DC,CAAAA,CAAc,MAAA,CAAQF,CAAAA,CAASC,CAAI,EACrC,CA0BO,SAASI,CAAAA,CAAkBC,CAAAA,CAAkBC,CAAAA,CAAkB,CACpE,OAAA,CAAQ,KAAA,CAAM,CAAA,EAAGT,CAAM,CAAA,mCAAA,EAAsCQ,CAAQ,CAAA,OAAA,EAAUC,CAAAA,CAAI,OAAO,CAAA,CAAE,EAC9F,CAyBO,SAASC,CAAAA,CAAUC,CAAAA,CAAeC,CAAAA,CAAwB,CAC1Dd,CAAAA,EAAe,EACpB,OAAA,CAAQ,GAAA,CAAI,CAAA,EAAGE,CAAM,CAAA,sBAAA,EAAyBW,CAAK,CAAA,UAAA,EAAaC,CAAQ,CAAA,CAAE,EAC5E,CAKO,SAASC,CAAAA,CAAaF,CAAAA,CAAeG,CAAAA,CAA0B,CAC/DhB,CAAAA,EAAe,EACpB,OAAA,CAAQ,GAAA,CAAI,GAAGE,CAAM,CAAA,gCAAA,EAAmCW,CAAK,CAAA,UAAA,EAAaG,CAAU,CAAA,EAAA,CAAI,EAC1F,CAKO,SAASC,CAAAA,CAAWJ,CAAAA,CAAeF,CAAAA,CAAoB,CAC5D,IAAMP,CAAAA,CAAUO,CAAAA,YAAe,KAAA,CAAQA,CAAAA,CAAI,OAAA,CAAU,MAAA,CAAOA,CAAG,CAAA,CAC/D,OAAA,CAAQ,KAAA,CAAM,CAAA,EAAGT,CAAM,CAAA,0BAAA,EAA6BW,CAAK,CAAA,OAAA,EAAUT,CAAO,CAAA,CAAE,EAC9E,CAKO,SAASc,CAAAA,CAAeC,CAAAA,CAAgBC,CAAAA,CAAaC,CAAAA,CAAwB,CAC7ErB,CAAAA,EAAe,EACpB,OAAA,CAAQ,GAAA,CAAI,CAAA,EAAGE,CAAM,CAAA,UAAA,EAAaiB,CAAM,CAAA,CAAA,EAAIC,CAAG,CAAA,EAAA,EAAKC,CAAQ,CAAA,OAAA,CAAS,EACvE,CAKO,SAASC,CAAAA,CAAgBC,CAAAA,CAAgBP,CAAAA,CAA0B,CACnEhB,CAAAA,EAAe,EACpB,OAAA,CAAQ,IAAI,CAAA,EAAGE,CAAM,CAAA,kBAAA,EAAqBqB,CAAM,CAAA,UAAA,EAAaP,CAAU,CAAA,EAAA,CAAI,EAC7E,CAMA,SAASV,CAAAA,CAAckB,CAAAA,CAA4CpB,CAAAA,CAAiBC,CAAAA,CAAsB,CACxG,IAAMoB,CAAAA,CAAQD,CAAAA,GAAU,OAAA,CAAU,OAAA,CAAQ,KAAA,CAAQA,CAAAA,GAAU,MAAA,CAAS,OAAA,CAAQ,IAAA,CAAO,OAAA,CAAQ,GAAA,CAExFnB,CAAAA,GAAS,MAAA,CACXoB,CAAAA,CAAM,CAAA,EAAGvB,CAAM,CAAA,CAAA,EAAIE,CAAO,CAAA,CAAA,CAAIC,CAAI,CAAA,CAElCoB,CAAAA,CAAM,CAAA,EAAGvB,CAAM,CAAA,CAAA,EAAIE,CAAO,CAAA,CAAE,EAEhC,CAEA,SAASH,CAAAA,CAAUyB,CAAAA,CAAkC,CACnD,GAAI,OAAO,OAAA,CAAY,GAAA,EAAe,OAAA,CAAQ,GAAA,CAC5C,OAAO,OAAA,CAAQ,GAAA,CAAIA,CAAI,CAG3B,CC7IA,IAAMC,EAAqB,EAAA,CACrBC,CAAAA,CAA4B,GAAA,CAC5BC,CAAAA,CAA6B,GAAA,CAMtBC,CAAAA,CAAN,KAAgB,CAMrB,WAAA,CAAYC,CAAAA,CAAyB,CALrCC,CAAAA,CAAA,IAAA,CAAiB,QAAA,CAAA,CACjBA,CAAAA,CAAA,IAAA,CAAQ,OAAA,CAA8B,EAAC,CAAA,CACvCA,CAAAA,CAAA,IAAA,CAAQ,cAAA,CAAqC,IAAA,CAAA,CAC7CA,CAAAA,CAAA,IAAA,CAAQ,YAAA,CAAmD,IAAA,CAAA,CAGzD,IAAA,CAAK,MAAA,CAAS,CACZ,OAAQD,CAAAA,CAAO,MAAA,CACf,QAAA,CAAUA,CAAAA,CAAO,QAAA,CACjB,KAAA,CAAOA,CAAAA,CAAO,KAAA,CACd,QAAA,CAAUA,CAAAA,CAAO,QAAA,CACjB,SAAA,CAAWA,CAAAA,CAAO,SAAA,EAAaJ,CAAAA,CAC/B,eAAA,CAAiBI,CAAAA,CAAO,eAAA,EAAmBH,CAAAA,CAC3C,gBAAA,CAAkBG,CAAAA,CAAO,gBAAA,EAAoBF,CAC/C,EACF,CAKA,SAAA,EAAqB,CACnB,OAAO,CAAC,IAAA,CAAK,OAAO,QAAA,EAAY,CAAC,CAAC,IAAA,CAAK,MAAA,CAAO,MAChD,CAMA,OAAA,CAAQI,CAAAA,CAAiC,CACnC,IAAA,CAAK,MAAA,CAAO,QAAA,GAEhB,IAAA,CAAK,KAAA,CAAM,IAAA,CAAKA,CAAK,CAAA,CAEjB,IAAA,CAAK,KAAA,CAAM,MAAA,EAAU,IAAA,CAAK,MAAA,CAAO,SAAA,CACnC,IAAA,CAAK,KAAA,EAAM,CAEX,IAAA,CAAK,aAAA,EAAc,EAEvB,CAMA,MAAM,KAAA,EAAuB,CAC3B,GAAI,IAAA,CAAK,YAAA,CACP,OAAO,IAAA,CAAK,YAAA,CAGd,GAAI,IAAA,CAAK,KAAA,CAAM,MAAA,GAAW,CAAA,CACxB,OAGF,IAAA,CAAK,oBAAA,EAAqB,CAE1B,IAAMC,CAAAA,CAAQ,IAAA,CAAK,KAAA,CACnB,OAAA,IAAA,CAAK,KAAA,CAAQ,EAAC,CAEd,IAAA,CAAK,YAAA,CAAe,IAAA,CAAK,SAAA,CAAUA,CAAK,EAAE,OAAA,CAAQ,IAAM,CACtD,IAAA,CAAK,YAAA,CAAe,KACtB,CAAC,CAAA,CAEM,IAAA,CAAK,YACd,CAKA,eAAA,EAA0B,CACxB,OAAO,IAAA,CAAK,KAAA,CAAM,MACpB,CAMQ,aAAA,EAAsB,CACxB,IAAA,CAAK,UAAA,GAAe,IAAA,GAExB,IAAA,CAAK,UAAA,CAAa,UAAA,CAAW,IAAM,CACjC,IAAA,CAAK,UAAA,CAAa,IAAA,CAClB,IAAA,CAAK,KAAA,GACP,CAAA,CAAG,IAAA,CAAK,MAAA,CAAO,eAAe,CAAA,EAChC,CAEQ,oBAAA,EAA6B,CAC/B,IAAA,CAAK,UAAA,GAAe,IAAA,GACtB,YAAA,CAAa,IAAA,CAAK,UAAU,CAAA,CAC5B,IAAA,CAAK,UAAA,CAAa,IAAA,EAEtB,CAEA,MAAc,SAAA,CAAUA,CAAAA,CAA4C,CAClE,GAAIA,CAAAA,CAAM,MAAA,GAAW,CAAA,CAAG,OAExB,IAAMC,CAAAA,CAAY,IAAA,CAAK,GAAA,EAAI,CAC3BvB,CAAAA,CAAUsB,CAAAA,CAAM,MAAA,CAAQ,CAAA,EAAG,IAAA,CAAK,MAAA,CAAO,QAAQ,CAAA,cAAA,CAAgB,CAAA,CAE/D,GAAI,CACF,MAAM,IAAA,CAAK,OAAA,CAAQ,MAAA,CAAQ,gBAAA,CAAkB,CAAE,MAAA,CAAQA,CAAM,CAAC,CAAA,CAC9DnB,CAAAA,CAAamB,CAAAA,CAAM,MAAA,CAAQ,IAAA,CAAK,GAAA,EAAI,CAAIC,CAAS,EACnD,CAAA,MAASC,CAAAA,CAAO,CACdnB,CAAAA,CAAWiB,CAAAA,CAAM,MAAA,CAAQE,CAAK,EAEhC,CACF,CAEA,MAAc,OAAA,CAAQjB,CAAAA,CAAgBkB,CAAAA,CAAcC,CAAAA,CAAkC,CACpF,IAAMlB,CAAAA,CAAM,CAAA,EAAG,IAAA,CAAK,MAAA,CAAO,QAAQ,CAAA,EAAGiB,CAAI,CAAA,CAAA,CACpCE,CAAAA,CAAa,IAAI,eAAA,CACjBC,CAAAA,CAAUF,EAAO,IAAA,CAAK,SAAA,CAAUA,CAAI,CAAA,CAAI,MAAA,CAE9CpB,CAAAA,CAAeC,CAAAA,CAAQC,CAAAA,CAAKoB,CAAAA,EAAS,MAAA,EAAU,CAAC,CAAA,CAEhD,IAAMC,CAAAA,CAAY,UAAA,CAAW,IAAM,CACjCF,CAAAA,CAAW,KAAA,GACb,CAAA,CAAG,IAAA,CAAK,MAAA,CAAO,gBAAgB,CAAA,CAEzBJ,CAAAA,CAAY,IAAA,CAAK,GAAA,EAAI,CAE3B,GAAI,CACF,IAAMO,CAAAA,CAAW,MAAM,KAAA,CAAMtB,CAAAA,CAAK,CAChC,MAAA,CAAAD,CAAAA,CACA,OAAA,CAAS,CACP,cAAA,CAAgB,kBAAA,CAChB,aAAA,CAAiB,CAAA,OAAA,EAAU,IAAA,CAAK,MAAA,CAAO,MAAM,CAAA,CAC/C,CAAA,CACA,IAAA,CAAMqB,CAAAA,CACN,MAAA,CAAQD,CAAAA,CAAW,MACrB,CAAC,CAAA,CAKD,GAHA,YAAA,CAAaE,CAAS,CAAA,CACtBnB,EAAgBoB,CAAAA,CAAS,MAAA,CAAQ,IAAA,CAAK,GAAA,EAAI,CAAIP,CAAS,CAAA,CAEnD,CAACO,CAAAA,CAAS,EAAA,CAAI,CAChB,IAAMC,CAAAA,CAAY,MAAMD,CAAAA,CAAS,IAAA,EAAK,CAAE,KAAA,CAAM,IAAM,eAAe,CAAA,CACnE,MAAM,IAAI,KAAA,CAAM,CAAA,KAAA,EAAQA,CAAAA,CAAS,MAAM,CAAA,EAAA,EAAKC,CAAS,CAAA,CAAE,CACzD,CAEA,IAAMC,CAAAA,CAAO,MAAMF,CAAAA,CAAS,IAAA,EAAK,CACjC,OAAOE,CAAAA,CAAO,IAAA,CAAK,KAAA,CAAMA,CAAI,CAAA,CAAI,EACnC,CAAA,MAASR,CAAAA,CAAO,CAGd,MAFA,YAAA,CAAaK,CAAS,CAAA,CAElBL,CAAAA,YAAiB,KAAA,EAASA,CAAAA,CAAM,IAAA,GAAS,YAAA,CACrC,IAAI,KAAA,CAAM,CAAA,sBAAA,EAAyB,IAAA,CAAK,OAAO,gBAAgB,CAAA,EAAA,CAAI,CAAA,CAGrEA,CACR,CACF,CACF,CAAA,CClLA,IAAMS,EAAAA,CAAW,iBAAA,CAEXC,EAAAA,CAAe,SAWrB,SAASC,EAAAA,EAAoC,CAE3C,OAAI,OAAO,OAAA,CAAY,GAAA,EAAe,OAAA,CAAQ,QAAA,EAAU,IAAA,CAC/C,CACL,IAAA,CAAM,QAAA,CACN,OAAA,CAAS,OAAA,CAAQ,QAAA,CAAS,IAC5B,CAAA,CAIE,OAAO,IAAA,CAAS,GAAA,CACX,CACL,IAAA,CAAM,MAAA,CACN,OAAA,CAAU,IAAA,CAAyC,OAAA,EAAS,IAAA,EAAQ,SACtE,CAAA,CAIE,OAAO,GAAA,CAAQ,GAAA,CACV,CACL,IAAA,CAAM,KAAA,CACN,OAAA,CAAU,GAAA,CAA6B,OAAA,EAAW,SACpD,CAAA,CAIE,OAAO,MAAA,CAAW,GAAA,EAAe,OAAO,SAAA,CAAc,GAAA,CACjD,CACL,KAAM,SAAA,CACN,OAAA,CAAS,SAAA,CAAU,SACrB,CAAA,CAGK,IACT,CAEA,SAASC,EAAAA,EAA0B,CAEjC,GAAI,OAAO,OAAA,CAAY,GAAA,EAAe,OAAA,CAAQ,QAAA,CAAU,CACtD,IAAMC,CAAAA,CAAW,OAAA,CAAQ,QAAA,CACzB,OAAQA,CAAAA,EACN,KAAK,QAAA,CACH,OAAO,QAAA,CACT,KAAK,OAAA,CACH,OAAO,SAAA,CACT,KAAK,OAAA,CACH,OAAO,OAAA,CACT,QACE,OAAOA,CACX,CACF,CAGA,GAAI,OAAO,SAAA,CAAc,GAAA,CAAa,CACpC,IAAMC,CAAAA,CAAK,SAAA,CAAU,SAAA,CAAU,WAAA,EAAY,CAC3C,GAAIA,CAAAA,CAAG,QAAA,CAAS,KAAK,CAAA,CAAG,OAAO,QAAA,CAC/B,GAAIA,CAAAA,CAAG,QAAA,CAAS,KAAK,CAAA,CAAG,OAAO,SAAA,CAC/B,GAAIA,CAAAA,CAAG,QAAA,CAAS,OAAO,CAAA,CAAG,OAAO,OACnC,CAEA,OAAO,IACT,CAEA,SAASC,EAAAA,EAAwB,CAO/B,GAAI,CAEF,GAAI,OAAOC,CAAAA,CAAY,GAAA,CAGrB,OADY,CAAA,EAAA,CACD,SAAW,SAE1B,CAAA,KAAQ,CAER,CAEA,OAAO,SACT,CAMA,IAAIC,CAAAA,CAAuC,IAAA,CAKpC,SAASC,CAAAA,CAAeC,CAAAA,CAAuC,CAEpE,GAAI,CAACF,CAAAA,CAAiB,CACpB,IAAMG,CAAAA,CAAUT,EAAAA,EAAc,CACxBU,CAAAA,CAAKT,EAAAA,EAAS,CAEpBK,CAAAA,CAAkB,CAChB,oBAAA,CAAsBR,EAAAA,CACtB,uBAAA,CAAyBM,IAAc,CACvC,wBAAA,CAA0BL,EAC5B,CAAA,CAEIU,CAAAA,GACFH,CAAAA,CAAgB,sBAAsB,CAAA,CAAIG,CAAAA,CAAQ,IAAA,CAClDH,CAAAA,CAAgB,yBAAyB,CAAA,CAAIG,CAAAA,CAAQ,OAAA,CAAA,CAGnDC,CAAAA,GACFJ,CAAAA,CAAgB,SAAS,CAAA,CAAII,CAAAA,EAEjC,CAGA,IAAMC,CAAAA,CAA0B,CAAE,GAAGL,CAAgB,CAAA,CAErD,OAAIE,CAAAA,EAAS,IAAA,GACXG,CAAAA,CAAU,cAAc,CAAA,CAAIH,CAAAA,CAAQ,IAAA,CAAA,CAElCA,CAAAA,EAAS,OAAA,GACXG,CAAAA,CAAU,iBAAiB,CAAA,CAAIH,CAAAA,CAAQ,OAAA,CAAA,CAErCA,CAAAA,EAAS,WAAA,GACXG,CAAAA,CAAU,wBAAwB,CAAA,CAAIH,CAAAA,CAAQ,WAAA,CAAA,CAGzCG,CACT,CCzIA,IAAIC,CAAAA,CAA8B,EAAC,CAC/BC,CAAAA,CAAoC,IAAA,CACpCC,CAAAA,CAAuC,IAAA,CAOrCC,CAAAA,CAAmB,0BAMlB,SAASC,EAAAA,CAAKhC,CAAAA,CAAwB,EAAC,CAAS,CACrD4B,CAAAA,CAAe5B,CAAAA,CAGXA,CAAAA,CAAO,KAAA,EACTjC,CAAAA,CAAS,IAAI,CAAA,CAIf+D,CAAAA,CAAkBP,CAAAA,CAAevB,CAAAA,CAAO,OAAO,CAAA,CAE/CxB,CAAAA,CAAK,kBAAA,CAAoB,CACvB,QAAA,CAAUwB,CAAAA,CAAO,QAAA,EAAY+B,CAAAA,CAC7B,KAAA,CAAO/B,CAAAA,CAAO,KAAA,EAAS,KAAA,CACvB,QAAA,CAAUA,CAAAA,CAAO,QAAA,EAAY,KAAA,CAC7B,SAAA,CAAW8B,CACb,CAAC,CAAA,CAEDD,CAAAA,CAAkBI,CAAAA,CAAgBjC,CAAM,CAAA,CAIpC6B,CAAAA,CAAgB,SAAA,EAAU,CAC5BrD,CAAAA,CAAK,mCAAmC,CAAA,CAExCJ,CAAAA,CAAM,wEAAwE,EAElF,CAYO,SAAS8D,CAAAA,EAAoC,CAClD,OAAOJ,CACT,CAYO,SAASK,EAAAA,EAAqB,CACnC,OAAOC,CAAAA,EAAa,CAAE,SAAA,EACxB,CASO,SAASA,CAAAA,EAA0B,CACxC,OAAKP,CAAAA,GACHA,CAAAA,CAAkBI,CAAAA,CAAgBL,CAAY,CAAA,CAAA,CAEzCC,CACT,CAKA,eAAsBQ,EAAAA,EAAuB,CACvCR,CAAAA,EACF,MAAMA,CAAAA,CAAgB,KAAA,GAE1B,CAKA,SAASI,CAAAA,CAAgBjC,CAAAA,CAAkC,CACzD,IAAMsC,CAAAA,CAAStC,CAAAA,CAAO,MAAA,EAAU9B,EAAAA,CAAU,iBAAiB,CAAA,CAE3D,OAAI,CAACoE,CAAAA,EAAU,CAACtC,CAAAA,CAAO,QAAA,EACrBvB,CAAAA,CAAK,yFAAyF,CAAA,CAGzF,IAAIsB,CAAAA,CAAU,CACnB,MAAA,CAAQuC,CAAAA,EAAU,EAAA,CAClB,QAAA,CAAUtC,CAAAA,CAAO,QAAA,EAAY+B,CAAAA,CAC7B,KAAA,CAAO/B,CAAAA,CAAO,KAAA,EAAS,KAAA,CACvB,QAAA,CAAUA,CAAAA,CAAO,QAAA,EAAY,CAACsC,CAAAA,CAC9B,SAAA,CAAWtC,CAAAA,CAAO,SAAA,CAClB,eAAA,CAAiBA,CAAAA,CAAO,eAAA,CACxB,gBAAA,CAAkBA,CAAAA,CAAO,gBAC3B,CAAC,CACH,CAKA,SAAS9B,EAAAA,CAAUyB,CAAAA,CAAkC,CACnD,GAAI,OAAO,OAAA,CAAY,GAAA,EAAe,OAAA,CAAQ,GAAA,CAC5C,OAAO,OAAA,CAAQ,GAAA,CAAIA,CAAI,CAG3B,CC3HA,IAAI4C,EAAAA,CAAgC,EAAC,CAO9B,SAASC,CAAAA,EAAmC,CACjD,OAAOD,EACT,CAyKO,SAASE,CAAAA,CAAYC,CAAAA,CAAmC,CAC7D,GAAI,CACF,IAAMC,CAAAA,CAAYP,CAAAA,EAAa,CAC/B,GAAI,CAACO,CAAAA,CAAU,SAAA,EAAU,CAAG,CAC1BvE,CAAAA,CAAM,2CAA2C,EACjD,MACF,CAEA,IAAMmE,CAAAA,CAAgBC,CAAAA,EAAiB,CACjCI,CAAAA,CAAeC,CAAAA,EAAgB,CAG/BC,CAAAA,CAAmBJ,CAAAA,CAAQ,QAAA,EAAsC,QAAA,CACjEK,CAAAA,CAAwBL,CAAAA,CAAQ,QAAA,EAAsC,aAAA,CAGtEf,CAAAA,CAAYO,CAAAA,EAAa,CAGzBc,CAAAA,CAAyC,CAC7C,GAAGT,CAAAA,CAAc,QAAA,CACjB,GAAGG,CAAAA,CAAQ,QAAA,CACX,GAAIf,CAAAA,CAAY,CAAE,UAAA,CAAYA,CAAU,CAAA,CAAI,EAC9C,CAAA,CACA,OAAOqB,CAAAA,CAAc,QAAA,CACrB,OAAOA,CAAAA,CAAc,aAAA,CAErB,IAAMC,CAAAA,CAA8B,CAClC,QAAA,CAAUP,CAAAA,CAAQ,IAAA,CAClB,IAAA,CAAMA,CAAAA,CAAQ,IAAA,CACd,QAAA,CAAU,SAAA,CACV,KAAA,CAAOA,CAAAA,CAAQ,IAAA,CACf,KAAA,CAAOQ,EAAAA,CAAcR,CAAAA,CAAQ,KAAK,CAAA,CAElC,OAAQS,CAAAA,CAAST,CAAAA,CAAQ,MAAA,CAAQ,CAAC,CAAA,CAClC,UAAA,CAAYA,CAAAA,CAAQ,UAAA,CACpB,MAAA,CAAQA,CAAAA,CAAQ,MAAA,EAAU,SAAA,CAC1B,YAAA,CAAcA,CAAAA,CAAQ,YAAA,CACtB,SAAA,CAAW,CAAA,CAAA,CACX,SAAA,CAAWH,CAAAA,CAAc,SAAA,CACzB,MAAA,CAAQA,CAAAA,CAAc,MAAA,CACtB,OAAA,CAASO,CAAAA,EAAmBF,CAAAA,EAAc,OAAA,CAC1C,MAAA,CAAQQ,CAAAA,EAAW,CACnB,YAAA,CAAcL,CAAAA,EAAwBH,CAAAA,EAAc,aAAA,CACpD,UAAA,CAAYF,CAAAA,CAAQ,UAAA,CACpB,QAAA,CAAUM,CAAAA,CACV,IAAA,CAAMT,CAAAA,CAAc,IACtB,CAAA,CAEAnE,CAAAA,CAAM,CAAA,eAAA,EAAkBsE,CAAAA,CAAQ,IAAI,CAAA,CAAA,EAAIA,CAAAA,CAAQ,IAAI,CAAA,CAAA,CAAI,CAAE,UAAA,CAAYA,CAAAA,CAAQ,UAAW,CAAC,CAAA,CAC1FC,CAAAA,CAAU,OAAA,CAAQM,CAAO,EAC3B,CAAA,MAASrE,EAAK,CACZF,CAAAA,CAAkB,SAAA,CAAWE,CAAAA,YAAe,KAAA,CAAQA,CAAAA,CAAM,IAAI,KAAA,CAAM,MAAA,CAAOA,CAAG,CAAC,CAAC,EAClF,CACF,CAMA,IAAMyE,CAAAA,CAAoB,GAAA,CAIpBC,EAAAA,CAAiB,CAAC,SAAA,CAAW,QAAA,CAAU,UAAA,CAAY,QAAA,CAAU,eAAe,CAAA,CAC5EC,EAAAA,CAA2B,CAAC,cAAA,CAAgB,YAAA,CAAc,cAAA,CAAgB,eAAA,CAAiB,UAAA,CAAY,eAAe,CAAA,CAGtHC,EAAAA,CAAkB,CAAC,aAAA,CAAe,cAAA,CAAgB,aAAA,CAAe,cAAA,CAAgB,kBAAA,CAAoB,iBAAA,CAAmB,kBAAA,CAAoB,sBAAA,CAAwB,uBAAA,CAAyB,iBAAiB,CAAA,CAKpN,SAASC,EAAAA,CAAeC,CAAAA,CAAsB,CAC5C,IAAMC,CAAAA,CAAWD,CAAAA,CAAI,WAAA,EAAY,CAGjC,OAAIF,EAAAA,CAAgB,QAAA,CAASG,CAAQ,CAAA,CAC5B,MAIL,CAAA,EAAAL,EAAAA,CAAe,IAAA,CAAMM,CAAAA,EAAMD,CAAAA,CAAS,QAAA,CAASC,CAAC,CAAC,CAAA,EAK/CL,EAAAA,CAAyB,IAAA,CAAMK,CAAAA,EAAMD,CAAAA,CAAS,QAAA,CAASC,CAAC,CAAC,CAAA,CAK/D,CAOA,SAASV,EAAAA,CAAcW,CAAAA,CAAyB,CAC9C,OAAOV,CAAAA,CAASU,CAAAA,CAAO,CAAC,CAC1B,CASA,SAASV,CAAAA,CAASW,EAAgBC,CAAAA,CAAwB,CAExD,GAAIA,CAAAA,CAAQ,EAAA,CAAI,OAAO,sBAAA,CAEvB,GAAID,CAAAA,EAAU,IAAA,CAA6B,OAAOA,CAAAA,CAElD,GAAI,OAAOA,CAAAA,EAAU,QAAA,CACnB,OAAOA,CAAAA,CAAM,MAAA,CAAST,CAAAA,CAClBS,CAAAA,CAAM,KAAA,CAAM,CAAA,CAAGT,CAAiB,CAAA,CAAI,gBAAA,CACpCS,CAAAA,CAGN,GAAI,OAAOA,CAAAA,EAAU,UAAY,OAAOA,CAAAA,EAAU,SAAA,CAChD,OAAOA,CAAAA,CAGT,GAAI,KAAA,CAAM,OAAA,CAAQA,CAAK,CAAA,CACrB,OAAOA,CAAAA,CAAM,GAAA,CAAKE,CAAAA,EAASb,CAAAA,CAASa,CAAAA,CAAMD,CAAAA,CAAQ,CAAC,CAAC,CAAA,CAGtD,GAAI,OAAOD,CAAAA,EAAU,QAAA,CAAU,CAC7B,IAAMG,CAAAA,CAAqC,EAAC,CAE5C,IAAA,GAAW,CAACP,CAAAA,CAAKQ,CAAG,CAAA,GAAK,MAAA,CAAO,OAAA,CAAQJ,CAAK,CAAA,CAEvCL,EAAAA,CAAeC,CAAG,CAAA,CACpBO,CAAAA,CAAUP,CAAG,CAAA,CAAI,YAAA,CAEjBO,CAAAA,CAAUP,CAAG,CAAA,CAAIP,CAAAA,CAASe,CAAAA,CAAKH,CAAAA,CAAQ,CAAC,CAAA,CAI5C,OAAOE,CACT,CAGA,OAAO,MAAA,CAAOH,CAAK,CACrB,CC3PA,IAAMK,EAAe,IAAIC,iBAAAA,CAUlB,SAAShB,CAAAA,EAAqB,CACnC,OAAI,OAAO,MAAA,CAAW,GAAA,EAAe,MAAA,CAAO,UAAA,CACnC,MAAA,CAAO,UAAA,EAAW,CAGpB,CAAA,EAAG,IAAA,CAAK,GAAA,EAAI,CAAE,QAAA,CAAS,EAAE,CAAC,CAAA,CAAA,EAAI,IAAA,CAAK,MAAA,EAAO,CAAE,QAAA,CAAS,EAAE,CAAA,CAAE,KAAA,CAAM,CAAA,CAAG,EAAE,CAAC,CAAA,CAC9E,CASO,SAASP,CAAAA,EAA4C,CAC1D,OAAOsB,CAAAA,CAAa,QAAA,EACtB,CAkCO,SAASE,EAAAA,CAAoBC,CAAAA,CAAyC,CAC3E,IAAMC,CAAAA,CAAM1B,CAAAA,EAAgB,CAC5B,GAAK0B,CAAAA,CAEL,OAAID,CAAAA,EAAcC,CAAAA,CAAI,gBAAA,CAAiB,GAAA,CAAID,CAAU,CAAA,CAC5CC,CAAAA,CAAI,gBAAA,CAAiB,GAAA,CAAID,CAAU,CAAA,CAGrCC,CAAAA,CAAI,aACb,CAKO,SAASC,EAAAA,CAAcF,CAAAA,CAA0B,CACtD,IAAMC,CAAAA,CAAM1B,CAAAA,EAAgB,CACxB0B,CAAAA,EACFA,CAAAA,CAAI,gBAAA,CAAiB,MAAA,CAAOD,CAAU,EAE1C,CA4BA,eAAsBpE,EAAAA,CACpBuE,CAAAA,CACAC,CAAAA,CACY,CAEZ,IAAMhC,CAAAA,CACJ,OAAO+B,CAAAA,EAAkB,QAAA,CAAW,CAAE,IAAA,CAAMA,CAAc,CAAA,CAAIA,CAAAA,CAE1DE,CAAAA,CAAgB9B,CAAAA,EAAgB,CAChC+B,CAAAA,CAAUD,CAAAA,EAAe,OAAA,EAAWvB,CAAAA,EAAW,CAC/CyB,CAAAA,CAAazB,CAAAA,EAAW,CAExB0B,CAAAA,CAAwB,CAC5B,OAAA,CAAAF,CAAAA,CACA,UAAA,CAAAC,CAAAA,CACA,aAAA,CAAeA,CAAAA,CACf,YAAA,CAAcF,CAAAA,EAAe,aAAA,CAC7B,IAAA,CAAMjC,CAAAA,CAAQ,IAAA,CACd,SAAA,CAAW,IAAA,CAAK,GAAA,GAChB,KAAA,CAAOA,CAAAA,CAAQ,KAAA,CACf,QAAA,CAAUA,CAAAA,CAAQ,QAAA,CAClB,IAAA,CAAMA,CAAAA,CAAQ,IAAA,CACd,gBAAA,CAAkB,IAAI,GACxB,CAAA,CAGA,OAAOyB,CAAAA,CAAa,GAAA,CAAIW,CAAAA,CAAS,SAAY,CAC3C,IAAIC,CAAAA,CACA1E,CAAAA,CAEJ,GAAI,CACF,OAAA0E,CAAAA,CAAS,MAAML,CAAAA,EAAG,CACXK,CACT,OAASC,CAAAA,CAAG,CACV,MAAA3E,CAAAA,CAAQ2E,CAAAA,YAAa,KAAA,CAAQA,CAAAA,CAAI,IAAI,KAAA,CAAM,MAAA,CAAOA,CAAC,CAAC,CAAA,CAC9CA,CACR,CAAA,OAAE,CAEAC,EAAAA,CAAaH,CAAAA,CAASzE,CAAAA,CAAQ,MAAA,CAAY0E,CAAAA,CAAQ1E,CAAK,EACzD,CACF,CAAC,CACH,CAMA,SAAS4E,EAAAA,CAAaH,CAAAA,CAAuBC,EAAkB1E,CAAAA,CAAqB,CAClF,IAAMsC,CAAAA,CAAYP,CAAAA,EAAa,CAC/B,GAAI,CAACO,CAAAA,CAAU,SAAA,EAAU,CAAG,CAC1BvE,CAAAA,CAAM,wCAAwC,CAAA,CAC9C,MACF,CAEA,IAAMmE,CAAAA,CAAgBC,CAAAA,EAAiB,CACjCvD,CAAAA,CAAa,IAAA,CAAK,GAAA,EAAI,CAAI6F,CAAAA,CAAQ,SAAA,CAGlCI,CAAAA,CAAS7E,CAAAA,CAAQ,IAAA,CAAO0E,CAAAA,CAExBI,CAAAA,CAA+B,CACnC,QAAA,CAAU,OAAA,CACV,IAAA,CAAML,CAAAA,CAAQ,IAAA,CACd,QAAA,CAAU,OAAA,CACV,KAAA,CAAOA,CAAAA,CAAQ,IAAA,CACf,OAAA,CAASA,CAAAA,CAAQ,OAAA,CACjB,MAAA,CAAQA,CAAAA,CAAQ,UAAA,CAChB,YAAA,CAAcA,CAAAA,CAAQ,YAAA,CACtB,KAAA,CAAOA,CAAAA,CAAQ,KAAA,CACf,MAAA,CAAAI,CAAAA,CACA,WAAA,CAAa,CAAA,CACb,YAAA,CAAc,CAAA,CACd,UAAA,CAAAjG,EACA,MAAA,CAAQoB,CAAAA,CAAQ,OAAA,CAAU,SAAA,CAC1B,YAAA,CAAcA,CAAAA,EAAO,OAAA,CACrB,SAAA,CAAW,KAAA,CACX,SAAA,CAAWkC,CAAAA,CAAc,SAAA,CACzB,MAAA,CAAQA,CAAAA,CAAc,MAAA,CACtB,QAAA,CAAU,CACR,GAAGA,CAAAA,CAAc,QAAA,CACjB,GAAGuC,CAAAA,CAAQ,QACb,CAAA,CACA,IAAA,CAAMA,CAAAA,CAAQ,IAAA,EAAQvC,CAAAA,CAAc,IACtC,CAAA,CAEAnE,CAAAA,CAAM,CAAA,mBAAA,EAAsB0G,CAAAA,CAAQ,IAAI,CAAA,CAAA,CAAI,CAAE,UAAA,CAAA7F,CAAAA,CAAY,QAAA,CAAU,CAAC,CAACoB,CAAM,CAAC,CAAA,CAC7EsC,CAAAA,CAAU,OAAA,CAAQwC,CAAQ,EAC5B,CAoCO,SAASC,EAAAA,CAAK1C,CAAAA,CAA4B,CAC/C,IAAMoC,CAAAA,CAAUjC,CAAAA,EAAgB,CAEhC,GAAI,CAACiC,CAAAA,CAAS,CAER,OAAA,CAAQ,IAAI,QAAA,GAAa,YAAA,EAC3B,OAAA,CAAQ,IAAA,CAAK,wEAAwE,CAAA,CAEvF,MACF,CAGA,IAAMO,CAAAA,CAAehB,EAAAA,CAAoB3B,CAAAA,CAAQ,UAAU,CAAA,CAE3DD,CAAAA,CAAkB,CAChB,IAAA,CAAMC,CAAAA,CAAQ,IAAA,CACd,IAAA,CAAMA,CAAAA,CAAQ,IAAA,CACd,KAAA,CAAOA,CAAAA,CAAQ,KAAA,CACf,MAAA,CAAQA,CAAAA,CAAQ,MAAA,CAChB,UAAA,CAAYA,CAAAA,CAAQ,UAAA,EAAc,CAAA,CAClC,MAAA,CAAQA,CAAAA,CAAQ,MAAA,EAAU,SAAA,CAC1B,YAAA,CAAcA,CAAAA,CAAQ,YAAA,CACtB,UAAA,CAAYA,CAAAA,CAAQ,UAAA,CACpB,QAAA,CAAU,CACR,GAAGA,CAAAA,CAAQ,QAAA,CACX,QAAA,CAAUoC,CAAAA,CAAQ,OAAA,CAClB,aAAA,CAAeO,CACjB,CACF,CAAC,CAAA,CAGG3C,CAAAA,CAAQ,UAAA,EACV8B,EAAAA,CAAc9B,CAAAA,CAAQ,UAAU,EAEpC","file":"index.mjs","sourcesContent":["{\n \"name\": \"@lelemondev/sdk\",\n \"version\": \"0.9.2\",\n \"description\": \"Automatic LLM observability. Wrap your client, everything is traced.\",\n \"author\": \"Lelemon <info@lelemon.dev>\",\n \"license\": \"MIT\",\n \"repository\": {\n \"type\": \"git\",\n \"url\": \"git+https://github.com/lelemondev/lelemondev-sdk.git\"\n },\n \"homepage\": \"https://lelemon.dev\",\n \"bugs\": {\n \"url\": \"https://github.com/lelemondev/lelemondev-sdk/issues\"\n },\n \"keywords\": [\n \"llm\",\n \"observability\",\n \"tracing\",\n \"openai\",\n \"anthropic\",\n \"nextjs\",\n \"lambda\",\n \"express\",\n \"hono\",\n \"claude\",\n \"gpt\",\n \"ai\",\n \"monitoring\",\n \"serverless\"\n ],\n \"main\": \"./dist/index.js\",\n \"module\": \"./dist/index.mjs\",\n \"types\": \"./dist/index.d.ts\",\n \"exports\": {\n \".\": {\n \"types\": \"./dist/index.d.ts\",\n \"import\": \"./dist/index.mjs\",\n \"require\": \"./dist/index.js\"\n },\n \"./openai\": {\n \"types\": \"./dist/openai.d.ts\",\n \"import\": \"./dist/openai.mjs\",\n \"require\": \"./dist/openai.js\"\n },\n \"./anthropic\": {\n \"types\": \"./dist/anthropic.d.ts\",\n \"import\": \"./dist/anthropic.mjs\",\n \"require\": \"./dist/anthropic.js\"\n },\n \"./bedrock\": {\n \"types\": \"./dist/bedrock.d.ts\",\n \"import\": \"./dist/bedrock.mjs\",\n \"require\": \"./dist/bedrock.js\"\n },\n \"./gemini\": {\n \"types\": \"./dist/gemini.d.ts\",\n \"import\": \"./dist/gemini.mjs\",\n \"require\": \"./dist/gemini.js\"\n },\n \"./openrouter\": {\n \"types\": \"./dist/openrouter.d.ts\",\n \"import\": \"./dist/openrouter.mjs\",\n \"require\": \"./dist/openrouter.js\"\n },\n \"./next\": {\n \"types\": \"./dist/next.d.ts\",\n \"import\": \"./dist/next.mjs\",\n \"require\": \"./dist/next.js\"\n },\n \"./lambda\": {\n \"types\": \"./dist/lambda.d.ts\",\n \"import\": \"./dist/lambda.mjs\",\n \"require\": \"./dist/lambda.js\"\n },\n \"./express\": {\n \"types\": \"./dist/express.d.ts\",\n \"import\": \"./dist/express.mjs\",\n \"require\": \"./dist/express.js\"\n },\n \"./hono\": {\n \"types\": \"./dist/hono.d.ts\",\n \"import\": \"./dist/hono.mjs\",\n \"require\": \"./dist/hono.js\"\n },\n \"./integrations\": {\n \"types\": \"./dist/integrations.d.ts\",\n \"import\": \"./dist/integrations.mjs\",\n \"require\": \"./dist/integrations.js\"\n },\n \"./package.json\": \"./package.json\"\n },\n \"typesVersions\": {\n \"*\": {\n \"openai\": [\n \"./dist/openai.d.ts\"\n ],\n \"anthropic\": [\n \"./dist/anthropic.d.ts\"\n ],\n \"bedrock\": [\n \"./dist/bedrock.d.ts\"\n ],\n \"gemini\": [\n \"./dist/gemini.d.ts\"\n ],\n \"openrouter\": [\n \"./dist/openrouter.d.ts\"\n ],\n \"next\": [\n \"./dist/next.d.ts\"\n ],\n \"lambda\": [\n \"./dist/lambda.d.ts\"\n ],\n \"express\": [\n \"./dist/express.d.ts\"\n ],\n \"hono\": [\n \"./dist/hono.d.ts\"\n ],\n \"integrations\": [\n \"./dist/integrations.d.ts\"\n ],\n \"*\": [\n \"./dist/index.d.ts\"\n ]\n }\n },\n \"files\": [\n \"dist\",\n \"README.md\"\n ],\n \"sideEffects\": false,\n \"engines\": {\n \"node\": \">=18.0.0\"\n },\n \"scripts\": {\n \"build\": \"tsup\",\n \"dev\": \"tsup --watch\",\n \"docs\": \"typedoc && node scripts/generate-llms-txt.mjs\",\n \"prepublishOnly\": \"npm run build\",\n \"lint\": \"eslint src/\",\n \"test\": \"vitest\",\n \"test:run\": \"vitest run\",\n \"test:coverage\": \"vitest run --coverage\",\n \"test:e2e\": \"vitest run tests/e2e\",\n \"typecheck\": \"tsc --noEmit\"\n },\n \"devDependencies\": {\n \"@aws-sdk/client-bedrock-runtime\": \"^3.962.0\",\n \"@google/generative-ai\": \"^0.24.1\",\n \"@types/node\": \"^20.0.0\",\n \"@vitest/coverage-v8\": \"^2.0.0\",\n \"dotenv\": \"^17.2.3\",\n \"openai\": \"^6.15.0\",\n \"tsup\": \"^8.5.1\",\n \"typedoc\": \"^0.28.15\",\n \"typescript\": \"^5.9.3\",\n \"vitest\": \"^2.0.0\"\n }\n}\n","/**\n * Centralized Logger\n *\n * Provides consistent debug logging across the SDK.\n * Enable via: init({ debug: true }) or LELEMON_DEBUG=true\n */\n\n// ─────────────────────────────────────────────────────────────\n// State\n// ─────────────────────────────────────────────────────────────\n\nlet debugEnabled = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Enable or disable debug mode\n */\nexport function setDebug(enabled: boolean): void {\n debugEnabled = enabled;\n}\n\n/**\n * Check if debug mode is enabled\n */\nexport function isDebugEnabled(): boolean {\n // Config takes precedence, then check env var\n if (debugEnabled) return true;\n return getEnvVar('LELEMON_DEBUG') === 'true';\n}\n\n// ─────────────────────────────────────────────────────────────\n// Logging Functions\n// ─────────────────────────────────────────────────────────────\n\nconst PREFIX = '[Lelemon]';\n\n/**\n * Log debug message (only when debug enabled)\n */\nexport function debug(message: string, data?: unknown): void {\n if (!isDebugEnabled()) return;\n logWithPrefix('debug', message, data);\n}\n\n/**\n * Log info message (only when debug enabled)\n */\nexport function info(message: string, data?: unknown): void {\n if (!isDebugEnabled()) return;\n logWithPrefix('info', message, data);\n}\n\n/**\n * Log warning message (always shown)\n */\nexport function warn(message: string, data?: unknown): void {\n logWithPrefix('warn', message, data);\n}\n\n/**\n * Log error message (always shown)\n */\nexport function error(message: string, data?: unknown): void {\n logWithPrefix('error', message, data);\n}\n\n// ─────────────────────────────────────────────────────────────\n// Trace-specific logging\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Log when a trace is captured\n */\nexport function traceCapture(provider: string, model: string, durationMs: number, status: string): void {\n if (!isDebugEnabled()) return;\n console.log(\n `${PREFIX} Captured trace: provider=${provider} model=${model} duration=${durationMs}ms status=${status}`\n );\n}\n\n/**\n * Log when a trace capture fails (always visible)\n */\nexport function traceCaptureError(provider: string, err: Error): void {\n console.error(`${PREFIX} Failed to capture trace: provider=${provider} error=${err.message}`);\n}\n\n/**\n * Log when a client is wrapped\n */\nexport function clientWrapped(provider: string): void {\n if (!isDebugEnabled()) return;\n console.log(`${PREFIX} Wrapped client: provider=${provider}`);\n}\n\n/**\n * Log transport events\n */\nexport function transportEvent(event: string, details?: Record<string, unknown>): void {\n if (!isDebugEnabled()) return;\n if (details) {\n console.log(`${PREFIX} Transport: ${event}`, details);\n } else {\n console.log(`${PREFIX} Transport: ${event}`);\n }\n}\n\n/**\n * Log batch send details\n */\nexport function batchSend(count: number, endpoint: string): void {\n if (!isDebugEnabled()) return;\n console.log(`${PREFIX} Sending batch: count=${count} endpoint=${endpoint}`);\n}\n\n/**\n * Log batch send success\n */\nexport function batchSuccess(count: number, durationMs: number): void {\n if (!isDebugEnabled()) return;\n console.log(`${PREFIX} Batch sent successfully: count=${count} duration=${durationMs}ms`);\n}\n\n/**\n * Log batch send failure (always visible - errors should never be silent)\n */\nexport function batchError(count: number, err: unknown): void {\n const message = err instanceof Error ? err.message : String(err);\n console.error(`${PREFIX} Batch send failed: count=${count} error=${message}`);\n}\n\n/**\n * Log request details (for deep debugging)\n */\nexport function requestDetails(method: string, url: string, bodySize: number): void {\n if (!isDebugEnabled()) return;\n console.log(`${PREFIX} Request: ${method} ${url} (${bodySize} bytes)`);\n}\n\n/**\n * Log response details\n */\nexport function responseDetails(status: number, durationMs: number): void {\n if (!isDebugEnabled()) return;\n console.log(`${PREFIX} Response: status=${status} duration=${durationMs}ms`);\n}\n\n// ─────────────────────────────────────────────────────────────\n// Helpers\n// ─────────────────────────────────────────────────────────────\n\nfunction logWithPrefix(level: 'debug' | 'info' | 'warn' | 'error', message: string, data?: unknown): void {\n const logFn = level === 'error' ? console.error : level === 'warn' ? console.warn : console.log;\n\n if (data !== undefined) {\n logFn(`${PREFIX} ${message}`, data);\n } else {\n logFn(`${PREFIX} ${message}`);\n }\n}\n\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * Transport Layer\n *\n * Handles batched HTTP requests to the Lelemon API.\n * Features:\n * - Queue-based batching\n * - Auto-flush on batch size or interval\n * - Request timeout protection\n * - Graceful error handling\n */\n\nimport type { CreateTraceRequest } from './types';\nimport { batchSend, batchSuccess, batchError, requestDetails, responseDetails, transportEvent } from './logger';\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\ninterface TransportConfig {\n apiKey: string;\n endpoint: string;\n debug: boolean;\n disabled: boolean;\n batchSize?: number;\n flushIntervalMs?: number;\n requestTimeoutMs?: number;\n}\n\nconst DEFAULT_BATCH_SIZE = 10;\nconst DEFAULT_FLUSH_INTERVAL_MS = 1000;\nconst DEFAULT_REQUEST_TIMEOUT_MS = 10000;\n\n// ─────────────────────────────────────────────────────────────\n// Transport Class\n// ─────────────────────────────────────────────────────────────\n\nexport class Transport {\n private readonly config: Required<TransportConfig>;\n private queue: CreateTraceRequest[] = [];\n private flushPromise: Promise<void> | null = null;\n private flushTimer: ReturnType<typeof setTimeout> | null = null;\n\n constructor(config: TransportConfig) {\n this.config = {\n apiKey: config.apiKey,\n endpoint: config.endpoint,\n debug: config.debug,\n disabled: config.disabled,\n batchSize: config.batchSize ?? DEFAULT_BATCH_SIZE,\n flushIntervalMs: config.flushIntervalMs ?? DEFAULT_FLUSH_INTERVAL_MS,\n requestTimeoutMs: config.requestTimeoutMs ?? DEFAULT_REQUEST_TIMEOUT_MS,\n };\n }\n\n /**\n * Check if transport is enabled\n */\n isEnabled(): boolean {\n return !this.config.disabled && !!this.config.apiKey;\n }\n\n /**\n * Enqueue a trace for sending\n * Fire-and-forget - never blocks\n */\n enqueue(trace: CreateTraceRequest): void {\n if (this.config.disabled) return;\n\n this.queue.push(trace);\n\n if (this.queue.length >= this.config.batchSize) {\n this.flush();\n } else {\n this.scheduleFlush();\n }\n }\n\n /**\n * Flush all pending traces\n * Safe to call multiple times\n */\n async flush(): Promise<void> {\n if (this.flushPromise) {\n return this.flushPromise;\n }\n\n if (this.queue.length === 0) {\n return;\n }\n\n this.cancelScheduledFlush();\n\n const items = this.queue;\n this.queue = [];\n\n this.flushPromise = this.sendBatch(items).finally(() => {\n this.flushPromise = null;\n });\n\n return this.flushPromise;\n }\n\n /**\n * Get pending count (for debugging)\n */\n getPendingCount(): number {\n return this.queue.length;\n }\n\n // ─────────────────────────────────────────────────────────────\n // Private Methods\n // ─────────────────────────────────────────────────────────────\n\n private scheduleFlush(): void {\n if (this.flushTimer !== null) return;\n\n this.flushTimer = setTimeout(() => {\n this.flushTimer = null;\n this.flush();\n }, this.config.flushIntervalMs);\n }\n\n private cancelScheduledFlush(): void {\n if (this.flushTimer !== null) {\n clearTimeout(this.flushTimer);\n this.flushTimer = null;\n }\n }\n\n private async sendBatch(items: CreateTraceRequest[]): Promise<void> {\n if (items.length === 0) return;\n\n const startTime = Date.now();\n batchSend(items.length, `${this.config.endpoint}/api/v1/ingest`);\n\n try {\n await this.request('POST', '/api/v1/ingest', { events: items });\n batchSuccess(items.length, Date.now() - startTime);\n } catch (error) {\n batchError(items.length, error);\n // Don't rethrow - observability should never crash the app\n }\n }\n\n private async request(method: string, path: string, body?: unknown): Promise<unknown> {\n const url = `${this.config.endpoint}${path}`;\n const controller = new AbortController();\n const bodyStr = body ? JSON.stringify(body) : undefined;\n\n requestDetails(method, url, bodyStr?.length ?? 0);\n\n const timeoutId = setTimeout(() => {\n controller.abort();\n }, this.config.requestTimeoutMs);\n\n const startTime = Date.now();\n\n try {\n const response = await fetch(url, {\n method,\n headers: {\n 'Content-Type': 'application/json',\n 'Authorization': `Bearer ${this.config.apiKey}`,\n },\n body: bodyStr,\n signal: controller.signal,\n });\n\n clearTimeout(timeoutId);\n responseDetails(response.status, Date.now() - startTime);\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => 'Unknown error');\n throw new Error(`HTTP ${response.status}: ${errorText}`);\n }\n\n const text = await response.text();\n return text ? JSON.parse(text) : {};\n } catch (error) {\n clearTimeout(timeoutId);\n\n if (error instanceof Error && error.name === 'AbortError') {\n throw new Error(`Request timeout after ${this.config.requestTimeoutMs}ms`);\n }\n\n throw error;\n }\n }\n}\n","/**\n * SDK Telemetry\n *\n * Auto-detects runtime environment and SDK metadata.\n * Follows OpenTelemetry semantic conventions.\n */\n\nimport type { SDKTelemetry, ServiceConfig } from './types';\n\n// SDK metadata (injected at build time or from package.json)\nconst SDK_NAME = '@lelemondev/sdk';\nconst SDK_VERSION = '__SDK_VERSION__'; // Replaced by build script or fallback\nconst SDK_LANGUAGE = 'nodejs';\n\n// ─────────────────────────────────────────────────────────────\n// Runtime Detection\n// ─────────────────────────────────────────────────────────────\n\ninterface RuntimeInfo {\n name: string;\n version: string;\n}\n\nfunction detectRuntime(): RuntimeInfo | null {\n // Node.js\n if (typeof process !== 'undefined' && process.versions?.node) {\n return {\n name: 'nodejs',\n version: process.versions.node,\n };\n }\n\n // Deno\n if (typeof Deno !== 'undefined') {\n return {\n name: 'deno',\n version: (Deno as { version?: { deno?: string } }).version?.deno ?? 'unknown',\n };\n }\n\n // Bun\n if (typeof Bun !== 'undefined') {\n return {\n name: 'bun',\n version: (Bun as { version?: string }).version ?? 'unknown',\n };\n }\n\n // Browser\n if (typeof window !== 'undefined' && typeof navigator !== 'undefined') {\n return {\n name: 'browser',\n version: navigator.userAgent,\n };\n }\n\n return null;\n}\n\nfunction detectOS(): string | null {\n // Node.js\n if (typeof process !== 'undefined' && process.platform) {\n const platform = process.platform;\n switch (platform) {\n case 'darwin':\n return 'darwin';\n case 'win32':\n return 'windows';\n case 'linux':\n return 'linux';\n default:\n return platform;\n }\n }\n\n // Browser - try to detect from userAgent\n if (typeof navigator !== 'undefined') {\n const ua = navigator.userAgent.toLowerCase();\n if (ua.includes('mac')) return 'darwin';\n if (ua.includes('win')) return 'windows';\n if (ua.includes('linux')) return 'linux';\n }\n\n return null;\n}\n\nfunction getSDKVersion(): string {\n // If version was injected at build time\n if (SDK_VERSION !== '__SDK_VERSION__') {\n return SDK_VERSION;\n }\n\n // Try to read from package.json in Node.js\n try {\n // Dynamic import to avoid bundler issues\n if (typeof require !== 'undefined') {\n // eslint-disable-next-line @typescript-eslint/no-require-imports\n const pkg = require('../../package.json');\n return pkg.version ?? 'unknown';\n }\n } catch {\n // Ignore - package.json not available\n }\n\n return 'unknown';\n}\n\n// ─────────────────────────────────────────────────────────────\n// Telemetry Builder\n// ─────────────────────────────────────────────────────────────\n\nlet cachedTelemetry: SDKTelemetry | null = null;\n\n/**\n * Build SDK telemetry object with auto-detected values\n */\nexport function buildTelemetry(service?: ServiceConfig): SDKTelemetry {\n // Cache the auto-detected values\n if (!cachedTelemetry) {\n const runtime = detectRuntime();\n const os = detectOS();\n\n cachedTelemetry = {\n 'telemetry.sdk.name': SDK_NAME,\n 'telemetry.sdk.version': getSDKVersion(),\n 'telemetry.sdk.language': SDK_LANGUAGE,\n };\n\n if (runtime) {\n cachedTelemetry['process.runtime.name'] = runtime.name;\n cachedTelemetry['process.runtime.version'] = runtime.version;\n }\n\n if (os) {\n cachedTelemetry['os.type'] = os;\n }\n }\n\n // Merge with service config\n const telemetry: SDKTelemetry = { ...cachedTelemetry };\n\n if (service?.name) {\n telemetry['service.name'] = service.name;\n }\n if (service?.version) {\n telemetry['service.version'] = service.version;\n }\n if (service?.environment) {\n telemetry['deployment.environment'] = service.environment;\n }\n\n return telemetry;\n}\n\n/**\n * Reset cached telemetry (for testing)\n */\nexport function resetTelemetryCache(): void {\n cachedTelemetry = null;\n}\n\n// Type declarations for runtime detection\ndeclare const Deno: unknown;\ndeclare const Bun: unknown;\ndeclare const window: unknown;\ndeclare const navigator: { userAgent: string } | undefined;\n","/**\n * Global Configuration\n *\n * Manages SDK configuration and transport instance.\n */\n\nimport type { LelemonConfig, SDKTelemetry } from './types';\nimport { Transport } from './transport';\nimport { setDebug, info, warn, debug } from './logger';\nimport { buildTelemetry } from './telemetry';\n\n// ─────────────────────────────────────────────────────────────\n// Global State\n// ─────────────────────────────────────────────────────────────\n\nlet globalConfig: LelemonConfig = {};\nlet globalTransport: Transport | null = null;\nlet globalTelemetry: SDKTelemetry | null = null;\nlet initialized = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\nconst DEFAULT_ENDPOINT = 'https://www.lelemon.dev';\n\n/**\n * Initialize the SDK\n * Call once at app startup\n */\nexport function init(config: LelemonConfig = {}): void {\n globalConfig = config;\n\n // Configure debug mode\n if (config.debug) {\n setDebug(true);\n }\n\n // Build telemetry with service config\n globalTelemetry = buildTelemetry(config.service);\n\n info('Initializing SDK', {\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? false,\n telemetry: globalTelemetry,\n });\n\n globalTransport = createTransport(config);\n initialized = true;\n\n // Log status after transport is created\n if (globalTransport.isEnabled()) {\n info('SDK initialized - tracing enabled');\n } else {\n debug('SDK initialized - tracing disabled (no API key or explicitly disabled)');\n }\n}\n\n/**\n * Get current config\n */\nexport function getConfig(): LelemonConfig {\n return globalConfig;\n}\n\n/**\n * Get SDK telemetry\n */\nexport function getTelemetry(): SDKTelemetry | null {\n return globalTelemetry;\n}\n\n/**\n * Check if SDK is initialized\n */\nexport function isInitialized(): boolean {\n return initialized;\n}\n\n/**\n * Check if SDK is enabled\n */\nexport function isEnabled(): boolean {\n return getTransport().isEnabled();\n}\n\n// ─────────────────────────────────────────────────────────────\n// Transport\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get or create transport instance\n */\nexport function getTransport(): Transport {\n if (!globalTransport) {\n globalTransport = createTransport(globalConfig);\n }\n return globalTransport;\n}\n\n/**\n * Flush all pending traces\n */\nexport async function flush(): Promise<void> {\n if (globalTransport) {\n await globalTransport.flush();\n }\n}\n\n/**\n * Create transport instance\n */\nfunction createTransport(config: LelemonConfig): Transport {\n const apiKey = config.apiKey ?? getEnvVar('LELEMON_API_KEY');\n\n if (!apiKey && !config.disabled) {\n warn('No API key provided. Set apiKey in init() or LELEMON_API_KEY env var. Tracing disabled.');\n }\n\n return new Transport({\n apiKey: apiKey ?? '',\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? !apiKey,\n batchSize: config.batchSize,\n flushIntervalMs: config.flushIntervalMs,\n requestTimeoutMs: config.requestTimeoutMs,\n });\n}\n\n/**\n * Get environment variable (works in Node and edge)\n */\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * Capture Module\n *\n * Handles trace capture and batching.\n * Called by providers to record LLM calls.\n */\n\nimport type { ProviderName, CreateTraceRequest, ObserveOptions, CaptureSpanOptions, SpanType } from './types';\nimport { getTransport, getTelemetry } from './config';\nimport { traceCapture, traceCaptureError, debug } from './logger';\nimport { getTraceContext, generateId } from './context';\n\n// ─────────────────────────────────────────────────────────────\n// Global context (set via observe options)\n// ─────────────────────────────────────────────────────────────\n\nlet globalContext: ObserveOptions = {};\n\nexport function setGlobalContext(options: ObserveOptions): void {\n globalContext = options;\n debug('Global context updated', options);\n}\n\nexport function getGlobalContext(): ObserveOptions {\n return globalContext;\n}\n\n// ─────────────────────────────────────────────────────────────\n// Capture Functions\n// ─────────────────────────────────────────────────────────────\n\nexport interface CaptureTraceParams {\n provider: ProviderName;\n model: string;\n input: unknown;\n durationMs: number;\n status: 'success' | 'error';\n streaming: boolean;\n metadata?: Record<string, unknown>;\n\n // Raw response (server extracts tokens, output, tools, etc.)\n rawResponse?: unknown;\n\n // Timing\n firstTokenMs?: number;\n\n // Manual span type\n spanType?: SpanType;\n name?: string;\n}\n\nexport interface CaptureErrorParams {\n provider: ProviderName;\n model: string;\n input: unknown;\n error: Error;\n durationMs: number;\n streaming: boolean;\n metadata?: Record<string, unknown>;\n}\n\n/**\n * Capture a successful trace (LLM call)\n * Fire-and-forget - never throws\n * @returns The span ID, for linking tool calls to this LLM span\n */\nexport function captureTrace(params: CaptureTraceParams): string | undefined {\n try {\n const transport = getTransport();\n if (!transport.isEnabled()) {\n debug('Transport disabled, skipping trace capture');\n return undefined;\n }\n\n const globalContext = getGlobalContext();\n const traceContext = getTraceContext();\n const spanId = generateId();\n\n // Include SDK telemetry in metadata\n const telemetry = getTelemetry();\n\n const request: CreateTraceRequest = {\n provider: params.provider,\n model: params.model,\n input: sanitizeInput(params.input),\n rawResponse: params.rawResponse ? sanitize(params.rawResponse, 0) : undefined,\n durationMs: params.durationMs,\n status: params.status,\n streaming: params.streaming,\n firstTokenMs: params.firstTokenMs,\n sessionId: globalContext.sessionId,\n userId: globalContext.userId,\n // Hierarchy fields\n traceId: traceContext?.traceId,\n spanId,\n parentSpanId: traceContext?.currentSpanId,\n metadata: {\n ...globalContext.metadata,\n ...params.metadata,\n ...(traceContext ? { _traceName: traceContext.name } : {}),\n ...(telemetry ? { _telemetry: telemetry } : {}),\n },\n tags: globalContext.tags,\n // Manual span fields\n spanType: params.spanType,\n name: params.name,\n };\n\n traceCapture(params.provider, params.model, params.durationMs, params.status);\n transport.enqueue(request);\n\n return spanId;\n } catch (err) {\n traceCaptureError(params.provider, err instanceof Error ? err : new Error(String(err)));\n return undefined;\n }\n}\n\n/**\n * Capture an error trace\n * Fire-and-forget - never throws\n */\nexport function captureError(params: CaptureErrorParams): void {\n try {\n const transport = getTransport();\n if (!transport.isEnabled()) {\n debug('Transport disabled, skipping error capture');\n return;\n }\n\n const globalContext = getGlobalContext();\n const traceContext = getTraceContext();\n\n // Include SDK telemetry in metadata\n const telemetry = getTelemetry();\n\n const request: CreateTraceRequest = {\n provider: params.provider,\n model: params.model,\n input: sanitizeInput(params.input),\n durationMs: params.durationMs,\n status: 'error',\n errorMessage: params.error.message,\n streaming: params.streaming,\n sessionId: globalContext.sessionId,\n userId: globalContext.userId,\n traceId: traceContext?.traceId,\n spanId: generateId(),\n parentSpanId: traceContext?.currentSpanId,\n metadata: {\n ...globalContext.metadata,\n ...params.metadata,\n ...(traceContext ? { _traceName: traceContext.name } : {}),\n ...(telemetry ? { _telemetry: telemetry } : {}),\n },\n tags: globalContext.tags,\n };\n\n traceCapture(params.provider, params.model, params.durationMs, 'error');\n debug('Error details', { message: params.error.message, stack: params.error.stack });\n transport.enqueue(request);\n } catch (err) {\n traceCaptureError(params.provider, err instanceof Error ? err : new Error(String(err)));\n }\n}\n\n// ─────────────────────────────────────────────────────────────\n// Manual Span Capture\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Manually capture a span (tool call, retrieval, custom)\n * Use this when auto-detection doesn't cover your use case\n *\n * @example\n * // Capture a tool call\n * captureSpan({\n * type: 'tool',\n * name: 'get_weather',\n * input: { location: 'San Francisco' },\n * output: { temperature: 72, conditions: 'sunny' },\n * durationMs: 150,\n * });\n *\n * @example\n * // Capture a retrieval/RAG operation\n * captureSpan({\n * type: 'retrieval',\n * name: 'vector_search',\n * input: { query: 'user question', k: 5 },\n * output: { documents: [...] },\n * durationMs: 50,\n * });\n */\nexport function captureSpan(options: CaptureSpanOptions): void {\n try {\n const transport = getTransport();\n if (!transport.isEnabled()) {\n debug('Transport disabled, skipping span capture');\n return;\n }\n\n const globalContext = getGlobalContext();\n const traceContext = getTraceContext();\n\n // Extract trace context from metadata if passed from span() in context.ts\n const metadataTraceId = (options.metadata as Record<string, unknown>)?._traceId as string | undefined;\n const metadataParentSpanId = (options.metadata as Record<string, unknown>)?._parentSpanId as string | undefined;\n\n // Include SDK telemetry in metadata\n const telemetry = getTelemetry();\n\n // Clean up internal metadata keys and add telemetry\n const cleanMetadata: Record<string, unknown> = {\n ...globalContext.metadata,\n ...options.metadata,\n ...(telemetry ? { _telemetry: telemetry } : {}),\n };\n delete cleanMetadata._traceId;\n delete cleanMetadata._parentSpanId;\n\n const request: CreateTraceRequest = {\n spanType: options.type,\n name: options.name,\n provider: 'unknown',\n model: options.name,\n input: sanitizeInput(options.input),\n // Manual spans use output directly (not rawResponse)\n output: sanitize(options.output, 0),\n durationMs: options.durationMs,\n status: options.status || 'success',\n errorMessage: options.errorMessage,\n streaming: false,\n sessionId: globalContext.sessionId,\n userId: globalContext.userId,\n traceId: metadataTraceId ?? traceContext?.traceId,\n spanId: generateId(),\n parentSpanId: metadataParentSpanId ?? traceContext?.currentSpanId,\n toolCallId: options.toolCallId,\n metadata: cleanMetadata,\n tags: globalContext.tags,\n };\n\n debug(`Span captured: ${options.type}/${options.name}`, { durationMs: options.durationMs });\n transport.enqueue(request);\n } catch (err) {\n traceCaptureError('unknown', err instanceof Error ? err : new Error(String(err)));\n }\n}\n\n// ─────────────────────────────────────────────────────────────\n// Sanitization (security)\n// ─────────────────────────────────────────────────────────────\n\nconst MAX_STRING_LENGTH = 100_000; // 100KB per field\n\n// Keys that should be redacted (authentication-related)\n// Note: We use specific patterns to avoid false positives with LLM token counts\nconst SENSITIVE_KEYS = ['api_key', 'apikey', 'password', 'secret', 'authorization'];\nconst SENSITIVE_TOKEN_PATTERNS = ['access_token', 'auth_token', 'bearer_token', 'refresh_token', 'id_token', 'session_token'];\n\n// Keys that contain \"token\" but are safe (LLM token counts)\nconst SAFE_TOKEN_KEYS = ['inputtokens', 'outputtokens', 'totaltokens', 'prompttokens', 'completiontokens', 'cachereadtokens', 'cachewritetokens', 'cachereadinputtokens', 'cachewriteinputtokens', 'reasoningtokens'];\n\n/**\n * Check if a key should be redacted\n */\nfunction isSensitiveKey(key: string): boolean {\n const lowerKey = key.toLowerCase();\n\n // Check if it's a safe token key (LLM token counts)\n if (SAFE_TOKEN_KEYS.includes(lowerKey)) {\n return false;\n }\n\n // Check sensitive patterns\n if (SENSITIVE_KEYS.some((k) => lowerKey.includes(k))) {\n return true;\n }\n\n // Check specific token patterns (auth tokens)\n if (SENSITIVE_TOKEN_PATTERNS.some((k) => lowerKey.includes(k))) {\n return true;\n }\n\n return false;\n}\n\n/**\n * Sanitize input before sending\n * - Truncates large strings\n * - Removes sensitive data\n */\nfunction sanitizeInput(input: unknown): unknown {\n return sanitize(input, 0);\n}\n\n/**\n * Sanitize output before sending\n */\nfunction sanitizeOutput(output: unknown): unknown {\n return sanitize(output, 0);\n}\n\nfunction sanitize(value: unknown, depth: number): unknown {\n // Prevent infinite recursion\n if (depth > 10) return '[max depth exceeded]';\n\n if (value === null || value === undefined) return value;\n\n if (typeof value === 'string') {\n return value.length > MAX_STRING_LENGTH\n ? value.slice(0, MAX_STRING_LENGTH) + '...[truncated]'\n : value;\n }\n\n if (typeof value === 'number' || typeof value === 'boolean') {\n return value;\n }\n\n if (Array.isArray(value)) {\n return value.map((item) => sanitize(item, depth + 1));\n }\n\n if (typeof value === 'object') {\n const sanitized: Record<string, unknown> = {};\n\n for (const [key, val] of Object.entries(value)) {\n // Redact sensitive keys\n if (isSensitiveKey(key)) {\n sanitized[key] = '[REDACTED]';\n } else {\n sanitized[key] = sanitize(val, depth + 1);\n }\n }\n\n return sanitized;\n }\n\n // Functions, symbols, etc.\n return String(value);\n}\n","/**\n * Trace Context Module\n *\n * Provides AsyncLocalStorage-based context for grouping spans under a parent trace.\n * Supports hierarchical tracing where:\n * - trace() creates a root \"agent\" span\n * - LLM calls become children of the root\n * - Tool calls become children of the LLM that triggered them (via toolCallId linking)\n *\n * @example\n * ```typescript\n * import { trace, span } from '@lelemondev/sdk';\n *\n * await trace({ name: 'sales-agent', input: userMessage }, async () => {\n * const response = await client.send(new ConverseCommand({...}));\n * // Tools automatically linked to their parent LLM via toolCallId\n * return response;\n * });\n * ```\n */\n\nimport { AsyncLocalStorage } from 'async_hooks';\nimport { getTransport } from './config';\nimport { getGlobalContext } from './capture';\nimport { debug } from './logger';\nimport type { CreateTraceRequest, SpanType } from './types';\n\n// ─────────────────────────────────────────────────────────────\n// Types\n// ─────────────────────────────────────────────────────────────\n\nexport interface TraceContext {\n /** Unique trace ID (shared by all spans in this trace) */\n traceId: string;\n /** Root span ID (the agent/workflow span) */\n rootSpanId: string;\n /** Current span ID (for nesting - LLM calls become children of this) */\n currentSpanId: string;\n /** Parent span ID (for nested trace() calls) */\n parentSpanId?: string;\n /** Trace name */\n name: string;\n /** Start time in ms */\n startTime: number;\n /** Input data */\n input?: unknown;\n /** Trace metadata */\n metadata?: Record<string, unknown>;\n /** Trace tags */\n tags?: string[];\n /** Map of toolCallId → llmSpanId for linking tool spans to their parent LLM */\n pendingToolCalls: Map<string, string>;\n}\n\nexport interface TraceOptions {\n /** Name for the trace (e.g., 'sales-agent', 'rag-query') */\n name: string;\n /** Input data for the trace */\n input?: unknown;\n /** Custom metadata */\n metadata?: Record<string, unknown>;\n /** Tags for filtering */\n tags?: string[];\n}\n\nexport interface SpanOptions {\n /** Span type */\n type: 'retrieval' | 'embedding' | 'tool' | 'guardrail' | 'rerank' | 'custom';\n /** Span name (e.g., 'pinecone-search', 'cohere-rerank') */\n name: string;\n /** Input data */\n input?: unknown;\n /** Output data */\n output?: unknown;\n /** Duration in milliseconds (optional, will be set automatically if not provided) */\n durationMs?: number;\n /** Status */\n status?: 'success' | 'error';\n /** Error message if status is 'error' */\n errorMessage?: string;\n /** Tool call ID (links this tool span to the LLM that requested it) */\n toolCallId?: string;\n /** Custom metadata */\n metadata?: Record<string, unknown>;\n}\n\n// ─────────────────────────────────────────────────────────────\n// AsyncLocalStorage for trace context\n// ─────────────────────────────────────────────────────────────\n\nconst traceStorage = new AsyncLocalStorage<TraceContext>();\n\n// ─────────────────────────────────────────────────────────────\n// ID Generation\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Generate a unique trace/span ID\n * Uses crypto.randomUUID if available, falls back to timestamp-based ID\n */\nexport function generateId(): string {\n if (typeof crypto !== 'undefined' && crypto.randomUUID) {\n return crypto.randomUUID();\n }\n // Fallback for older Node.js versions\n return `${Date.now().toString(36)}-${Math.random().toString(36).slice(2, 11)}`;\n}\n\n// ─────────────────────────────────────────────────────────────\n// Context API\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get the current trace context, if any\n */\nexport function getTraceContext(): TraceContext | undefined {\n return traceStorage.getStore();\n}\n\n/**\n * Check if we're inside a trace() block\n */\nexport function hasTraceContext(): boolean {\n return traceStorage.getStore() !== undefined;\n}\n\n// ─────────────────────────────────────────────────────────────\n// Tool Call Linking\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Register tool calls from an LLM response.\n * When the LLM returns with tool_use, call this to link subsequent tool spans.\n * @param toolCallIds - Array of tool call IDs from the LLM response\n * @param llmSpanId - The span ID of the LLM call that requested these tools\n */\nexport function registerToolCalls(toolCallIds: string[], llmSpanId: string): void {\n const ctx = getTraceContext();\n if (!ctx) return;\n\n for (const id of toolCallIds) {\n ctx.pendingToolCalls.set(id, llmSpanId);\n debug(`Registered tool call ${id} → LLM span ${llmSpanId}`);\n }\n}\n\n/**\n * Get the correct parent span ID for a tool span.\n * If toolCallId matches a pending tool call, returns the LLM span that requested it.\n * Otherwise falls back to the current span ID.\n */\nexport function getToolParentSpanId(toolCallId?: string): string | undefined {\n const ctx = getTraceContext();\n if (!ctx) return undefined;\n\n if (toolCallId && ctx.pendingToolCalls.has(toolCallId)) {\n return ctx.pendingToolCalls.get(toolCallId);\n }\n\n return ctx.currentSpanId;\n}\n\n/**\n * Clear a tool call from the pending map after it's been processed.\n */\nexport function clearToolCall(toolCallId: string): void {\n const ctx = getTraceContext();\n if (ctx) {\n ctx.pendingToolCalls.delete(toolCallId);\n }\n}\n\n// ─────────────────────────────────────────────────────────────\n// trace() - Main API for grouping spans\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Execute a function within a trace context.\n * Creates a root \"agent\" span that contains all LLM calls and tool executions.\n * The result of the function becomes the output of the root span.\n *\n * @example Simple usage (just name)\n * ```typescript\n * await trace('sales-agent', async () => {\n * await client.send(new ConverseCommand({...}));\n * return finalResponse;\n * });\n * ```\n *\n * @example With options\n * ```typescript\n * await trace({ name: 'rag-query', input: question, tags: ['production'] }, async () => {\n * const docs = await search(question);\n * span({ type: 'retrieval', name: 'pinecone', output: { count: docs.length } });\n * return client.send(new ConverseCommand({...}));\n * });\n * ```\n */\nexport async function trace<T>(\n nameOrOptions: string | TraceOptions,\n fn: () => Promise<T>\n): Promise<T> {\n // Normalize options\n const options: TraceOptions =\n typeof nameOrOptions === 'string' ? { name: nameOrOptions } : nameOrOptions;\n\n const parentContext = getTraceContext();\n const traceId = parentContext?.traceId ?? generateId();\n const rootSpanId = generateId();\n\n const context: TraceContext = {\n traceId,\n rootSpanId,\n currentSpanId: rootSpanId,\n parentSpanId: parentContext?.currentSpanId,\n name: options.name,\n startTime: Date.now(),\n input: options.input,\n metadata: options.metadata,\n tags: options.tags,\n pendingToolCalls: new Map(),\n };\n\n // Run the function within the trace context\n return traceStorage.run(context, async () => {\n let result: T | undefined;\n let error: Error | undefined;\n\n try {\n result = await fn();\n return result;\n } catch (e) {\n error = e instanceof Error ? e : new Error(String(e));\n throw e;\n } finally {\n // Send the root span with the final result\n sendRootSpan(context, error ? undefined : result, error);\n }\n });\n}\n\n/**\n * Send the root \"agent\" span when the trace completes.\n * This span represents the entire agent execution with input/output.\n */\nfunction sendRootSpan(context: TraceContext, result?: unknown, error?: Error): void {\n const transport = getTransport();\n if (!transport.isEnabled()) {\n debug('Transport disabled, skipping root span');\n return;\n }\n\n const globalContext = getGlobalContext();\n const durationMs = Date.now() - context.startTime;\n\n // Sanitize result if it's a string (common case for agent responses)\n const output = error ? null : result;\n\n const rootSpan: CreateTraceRequest = {\n spanType: 'agent' as SpanType,\n name: context.name,\n provider: 'agent',\n model: context.name,\n traceId: context.traceId,\n spanId: context.rootSpanId,\n parentSpanId: context.parentSpanId,\n input: context.input,\n output,\n inputTokens: 0, // Will be aggregated from children\n outputTokens: 0,\n durationMs,\n status: error ? 'error' : 'success',\n errorMessage: error?.message,\n streaming: false,\n sessionId: globalContext.sessionId,\n userId: globalContext.userId,\n metadata: {\n ...globalContext.metadata,\n ...context.metadata,\n },\n tags: context.tags ?? globalContext.tags,\n };\n\n debug(`Sending root span: ${context.name}`, { durationMs, hasError: !!error });\n transport.enqueue(rootSpan);\n}\n\n// ─────────────────────────────────────────────────────────────\n// span() - Manual span capture for non-LLM operations\n// ─────────────────────────────────────────────────────────────\n\n// Import directly - circular dependency is handled by module system\nimport { captureSpan as captureSpanImport } from './capture';\n\n/**\n * Manually capture a span for non-LLM operations (retrieval, embedding, tool, etc.)\n * Must be called within a trace() block.\n *\n * @example Tool with toolCallId (links to parent LLM)\n * ```typescript\n * span({\n * type: 'tool',\n * name: 'query_database',\n * toolCallId: 'tooluse_abc123', // Links to LLM that requested this\n * input: { sql: 'SELECT ...' },\n * output: { rows: [...] },\n * durationMs: 15,\n * });\n * ```\n *\n * @example Retrieval without toolCallId\n * ```typescript\n * span({\n * type: 'retrieval',\n * name: 'pinecone-search',\n * input: { topK: 5 },\n * output: { count: 10 },\n * durationMs: 50,\n * });\n * ```\n */\nexport function span(options: SpanOptions): void {\n const context = getTraceContext();\n\n if (!context) {\n // Warn but don't throw - fire-and-forget principle\n if (process.env.NODE_ENV !== 'production') {\n console.warn('[Lelemon] span() called outside of trace() - span will not be captured');\n }\n return;\n }\n\n // Determine parent: if toolCallId provided and registered, use the LLM span\n const parentSpanId = getToolParentSpanId(options.toolCallId);\n\n captureSpanImport({\n type: options.type,\n name: options.name,\n input: options.input,\n output: options.output,\n durationMs: options.durationMs ?? 0,\n status: options.status ?? 'success',\n errorMessage: options.errorMessage,\n toolCallId: options.toolCallId,\n metadata: {\n ...options.metadata,\n _traceId: context.traceId,\n _parentSpanId: parentSpanId,\n },\n });\n\n // Clear the tool call after processing\n if (options.toolCallId) {\n clearToolCall(options.toolCallId);\n }\n}\n\n// ─────────────────────────────────────────────────────────────\n// Internal helpers (exported for capture.ts)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Run a function with a specific trace context (for internal use)\n */\nexport function runWithContext<T>(context: TraceContext, fn: () => T): T {\n return traceStorage.run(context, fn);\n}\n"]}
|
package/dist/integrations.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/integrations/next.ts","../src/core/config.ts","../src/integrations/lambda.ts","../src/integrations/express.ts","../src/integrations/hono.ts"],"names":["next_exports","__export","createWrapper","withObserve","flush","handler","options","request","context","defaultOptions","lambda_exports","event","express_exports","createMiddleware","_req","res","next","hono_exports","c"],"mappings":";AAAA,IAAA,CAAA,CAAA,MAAA,CAAA,cAAA,CAAA,IAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,GAAA,CAAA,IAAA,IAAA,CAAA,IAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,GAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,UAAA,CAAA,IAAA,CAAA,EAAA,CAAA,CAAA,IAAAA,EAAA,GAAAC,CAAAA,CAAAD,EAAA,CAAA,aAAA,CAAA,IAAAE,CAAAA,CAAA,gBAAAC,CAAAA,CAAAA,CAAAA,CCwGA,eAAsBC,GAAuB,CAI7C,CDhBO,SAASD,CAAAA,CACdE,CAAAA,CACAC,EAC4B,CAC5B,aAAcC,CAAAA,CAAkBC,CAAAA,GAA0C,CACxE,GAAI,CACF,OAAO,MAAMH,CAAAA,CAAQE,EAASC,CAAO,CACvC,QAAE,CAEIF,CAAAA,EAAS,MAEXA,CAAAA,CAAQ,KAAA,CAAM,IAAMF,CAAAA,EAAO,CAAA,CAClBE,CAAAA,EAAS,UAElBA,CAAAA,CAAQ,SAAA,CAAUF,GAAO,CAAA,CAGzB,MAAMA,CAAAA,GAEV,CACF,CACF,CAeO,SAASF,CAAAA,CAAcO,CAAAA,CAAoC,CAChE,OAAO,SACLJ,EACAC,CAAAA,CAC4B,CAC5B,OAAOH,CAAAA,CAAYE,CAAAA,CAAS,CAAE,GAAGI,CAAAA,CAAgB,GAAGH,CAAQ,CAAC,CAC/D,CACF,KEvIAI,CAAAA,CAAA,GAAAT,EAAAS,CAAAA,CAAA,CAAA,WAAA,CAAA,IAAAP,IAqFO,SAASA,CAAAA,CACdE,CAAAA,CACgC,CAChC,OAAO,MAAOM,CAAAA,CAAeH,IAA6C,CACxE,GAAI,CACF,OAAO,MAAMH,EAAQM,CAAAA,CAAOH,CAAO,CACrC,CAAA,OAAE,CAEA,MAAMJ,CAAAA,GACR,CACF,CACF,CChGA,IAAAQ,CAAAA,CAAA,GAAAX,CAAAA,CAAAW,CAAAA,CAAA,sBAAAC,CAAAA,CAAAA,CAAAA,CA0EO,SAASA,GAAsC,CACpD,OAAO,CAACC,CAAAA,CAAMC,CAAAA,CAAKC,IAAS,CAE1BD,CAAAA,CAAI,GAAG,QAAA,CAAU,IAAM,CACrBX,CAAAA,EAAM,CAAE,KAAA,CAAM,IAAM,CAEpB,CAAC,EACH,CAAC,CAAA,CAEDY,CAAAA,GACF,CACF,KCrFAC,CAAAA,CAAA,GAAAhB,EAAAgB,CAAAA,CAAA,CAAA,gBAAA,CAAA,IAAAJ,IAmFO,SAASA,CAAAA,EAAmC,CACjD,OAAO,MAAOK,EAAGF,CAAAA,GAAS,CACxB,MAAMA,CAAAA,EAAK,CAGPE,EAAE,YAAA,EAAc,SAAA,CAClBA,EAAE,YAAA,CAAa,SAAA,CAAUd,GAAO,CAAA,CAGhCA,GAAM,CAAE,KAAA,CAAM,IAAM,CAAC,CAAC,EAE1B,CACF","file":"integrations.js","sourcesContent":["/**\n * Next.js App Router Integration\n *\n * Wraps route handlers to automatically flush traces.\n * Supports Next.js 15+ `after()` and Vercel's `waitUntil()`.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/next';\n *\n * export const POST = withObserve(async (req) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return Response.json(result);\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Next.js App Router handler type\n *\n * @typeParam TContext - Optional context type for dynamic route parameters\n */\nexport type NextRouteHandler<TContext = unknown> = (\n request: Request,\n context?: TContext\n) => Response | Promise<Response>;\n\n/**\n * Options for the Next.js wrapper\n */\nexport interface NextObserveOptions {\n /**\n * Next.js 15+ after() function from 'next/server'\n * Preferred method - runs after response without blocking\n *\n * @example\n * import { after } from 'next/server';\n * export const POST = withObserve(handler, { after });\n */\n after?: (callback: () => void | Promise<void>) => void;\n\n /**\n * Vercel's waitUntil() from '@vercel/functions'\n * Alternative for Vercel deployments\n *\n * @example\n * import { waitUntil } from '@vercel/functions';\n * export const POST = withObserve(handler, { waitUntil });\n */\n waitUntil?: (promise: Promise<unknown>) => void;\n}\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap a Next.js App Router handler with automatic trace flushing\n *\n * @param handler - Your route handler function\n * @param options - Optional: pass `after` (Next.js 15+) or `waitUntil` (Vercel)\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // Basic usage (blocking flush)\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n *\n * @example\n * // Next.js 15+ with after() - non-blocking (recommended)\n * import { after } from 'next/server';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { after }\n * );\n *\n * @example\n * // Vercel with waitUntil() - non-blocking\n * import { waitUntil } from '@vercel/functions';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { waitUntil }\n * );\n */\nexport function withObserve<TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n): NextRouteHandler<TContext> {\n return async (request: Request, context?: TContext): Promise<Response> => {\n try {\n return await handler(request, context);\n } finally {\n // Priority: after() > waitUntil() > blocking flush\n if (options?.after) {\n // Next.js 15+ native - best option\n options.after(() => flush());\n } else if (options?.waitUntil) {\n // Vercel platform\n options.waitUntil(flush());\n } else {\n // Fallback: blocking flush\n await flush();\n }\n }\n };\n}\n\n/**\n * Create a pre-configured wrapper with default options\n *\n * @example\n * import { after } from 'next/server';\n * import { createWrapper } from '@lelemondev/sdk/next';\n *\n * const withObserve = createWrapper({ after });\n *\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n */\nexport function createWrapper(defaultOptions: NextObserveOptions) {\n return function <TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n ): NextRouteHandler<TContext> {\n return withObserve(handler, { ...defaultOptions, ...options });\n };\n}\n","/**\n * Global Configuration\n *\n * Manages SDK configuration and transport instance.\n */\n\nimport type { LelemonConfig, SDKTelemetry } from './types';\nimport { Transport } from './transport';\nimport { setDebug, info, warn, debug } from './logger';\nimport { buildTelemetry } from './telemetry';\n\n// ─────────────────────────────────────────────────────────────\n// Global State\n// ─────────────────────────────────────────────────────────────\n\nlet globalConfig: LelemonConfig = {};\nlet globalTransport: Transport | null = null;\nlet globalTelemetry: SDKTelemetry | null = null;\nlet initialized = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\nconst DEFAULT_ENDPOINT = 'https://www.lelemon.dev';\n\n/**\n * Initialize the SDK\n * Call once at app startup\n */\nexport function init(config: LelemonConfig = {}): void {\n globalConfig = config;\n\n // Configure debug mode\n if (config.debug) {\n setDebug(true);\n }\n\n // Build telemetry with service config\n globalTelemetry = buildTelemetry(config.service);\n\n info('Initializing SDK', {\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? false,\n telemetry: globalTelemetry,\n });\n\n globalTransport = createTransport(config);\n initialized = true;\n\n // Log status after transport is created\n if (globalTransport.isEnabled()) {\n info('SDK initialized - tracing enabled');\n } else {\n debug('SDK initialized - tracing disabled (no API key or explicitly disabled)');\n }\n}\n\n/**\n * Get current config\n */\nexport function getConfig(): LelemonConfig {\n return globalConfig;\n}\n\n/**\n * Get SDK telemetry\n */\nexport function getTelemetry(): SDKTelemetry | null {\n return globalTelemetry;\n}\n\n/**\n * Check if SDK is initialized\n */\nexport function isInitialized(): boolean {\n return initialized;\n}\n\n/**\n * Check if SDK is enabled\n */\nexport function isEnabled(): boolean {\n return getTransport().isEnabled();\n}\n\n// ─────────────────────────────────────────────────────────────\n// Transport\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get or create transport instance\n */\nexport function getTransport(): Transport {\n if (!globalTransport) {\n globalTransport = createTransport(globalConfig);\n }\n return globalTransport;\n}\n\n/**\n * Flush all pending traces\n */\nexport async function flush(): Promise<void> {\n if (globalTransport) {\n await globalTransport.flush();\n }\n}\n\n/**\n * Create transport instance\n */\nfunction createTransport(config: LelemonConfig): Transport {\n const apiKey = config.apiKey ?? getEnvVar('LELEMON_API_KEY');\n\n if (!apiKey && !config.disabled) {\n warn('No API key provided. Set apiKey in init() or LELEMON_API_KEY env var. Tracing disabled.');\n }\n\n return new Transport({\n apiKey: apiKey ?? '',\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? !apiKey,\n batchSize: config.batchSize,\n flushIntervalMs: config.flushIntervalMs,\n requestTimeoutMs: config.requestTimeoutMs,\n });\n}\n\n/**\n * Get environment variable (works in Node and edge)\n */\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * AWS Lambda Integration\n *\n * Wraps Lambda handlers to automatically flush traces before the function exits.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/lambda';\n *\n * export const handler = withObserve(async (event) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return { statusCode: 200, body: JSON.stringify(result) };\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring @types/aws-lambda)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * AWS Lambda Context object\n */\nexport interface LambdaContext {\n functionName: string;\n functionVersion: string;\n invokedFunctionArn: string;\n memoryLimitInMB: string;\n awsRequestId: string;\n logGroupName: string;\n logStreamName: string;\n getRemainingTimeInMillis(): number;\n [key: string]: unknown;\n}\n\n/**\n * Generic AWS Lambda handler type\n *\n * @typeParam TEvent - The event type (e.g., APIGatewayProxyEvent)\n * @typeParam TResult - The result type (e.g., APIGatewayProxyResult)\n */\nexport type LambdaHandler<TEvent = unknown, TResult = unknown> = (\n event: TEvent,\n context: LambdaContext\n) => Promise<TResult>;\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap an AWS Lambda handler with automatic trace flushing\n *\n * Always flushes before returning - Lambda freezes the container\n * immediately after the handler returns, so this is required.\n *\n * @param handler - Your Lambda handler function\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // API Gateway event\n * export const handler = withObserve(async (event) => {\n * const body = JSON.parse(event.body);\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({\n * model: 'gpt-4',\n * messages: [{ role: 'user', content: body.message }],\n * });\n * return {\n * statusCode: 200,\n * body: JSON.stringify(result.choices[0].message),\n * };\n * });\n *\n * @example\n * // With typed events\n * import type { APIGatewayProxyEvent, APIGatewayProxyResult } from 'aws-lambda';\n *\n * export const handler = withObserve<APIGatewayProxyEvent, APIGatewayProxyResult>(\n * async (event, context) => {\n * return { statusCode: 200, body: 'OK' };\n * }\n * );\n */\nexport function withObserve<TEvent = unknown, TResult = unknown>(\n handler: LambdaHandler<TEvent, TResult>\n): LambdaHandler<TEvent, TResult> {\n return async (event: TEvent, context: LambdaContext): Promise<TResult> => {\n try {\n return await handler(event, context);\n } finally {\n // Always flush - Lambda freezes immediately after return\n await flush();\n }\n };\n}\n","/**\n * Express Integration\n *\n * Middleware that automatically flushes traces when response finishes.\n *\n * @example\n * import express from 'express';\n * import { createMiddleware } from '@lelemondev/sdk/express';\n *\n * const app = express();\n * app.use(createMiddleware());\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring express as dependency)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Minimal Express request type (avoids requiring express as dependency)\n */\nexport interface ExpressRequest {\n [key: string]: unknown;\n}\n\n/**\n * Minimal Express response type (avoids requiring express as dependency)\n */\nexport interface ExpressResponse {\n on(event: 'finish' | 'close' | 'error', listener: () => void): this;\n [key: string]: unknown;\n}\n\n/**\n * Express next function type\n */\nexport type ExpressNextFunction = (error?: unknown) => void;\n\n/**\n * Express middleware function type\n *\n * @param req - Express request object\n * @param res - Express response object\n * @param next - Next function to pass control\n */\nexport type ExpressMiddleware = (\n req: ExpressRequest,\n res: ExpressResponse,\n next: ExpressNextFunction\n) => void;\n\n// ─────────────────────────────────────────────────────────────\n// Middleware\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create Express middleware for automatic trace flushing\n *\n * Flushes traces when the response finishes (after res.send/res.json).\n * This is fire-and-forget and doesn't block the response.\n *\n * @returns Express middleware function\n *\n * @example\n * // Global middleware\n * app.use(createMiddleware());\n *\n * @example\n * // Per-route middleware\n * app.post('/chat', createMiddleware(), async (req, res) => {\n * res.json({ ok: true });\n * });\n */\nexport function createMiddleware(): ExpressMiddleware {\n return (_req, res, next) => {\n // Flush when response is finished (after headers + body sent)\n res.on('finish', () => {\n flush().catch(() => {\n // Silently ignore flush errors - fire and forget\n });\n });\n\n next();\n };\n}\n","/**\n * Hono Integration\n *\n * Middleware for Hono framework (Cloudflare Workers, Deno, Bun, Node.js).\n * Uses executionCtx.waitUntil() when available for non-blocking flush.\n *\n * @example\n * import { Hono } from 'hono';\n * import { createMiddleware } from '@lelemondev/sdk/hono';\n *\n * const app = new Hono();\n * app.use(createMiddleware());\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring hono as dependency)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Execution context for edge runtimes (Cloudflare Workers, Deno Deploy)\n */\nexport interface ExecutionContext {\n waitUntil(promise: Promise<unknown>): void;\n passThroughOnException(): void;\n}\n\n/**\n * Minimal Hono context type (avoids requiring hono as dependency)\n */\nexport interface HonoContext {\n req: {\n raw: Request;\n [key: string]: unknown;\n };\n res: Response | undefined;\n executionCtx?: ExecutionContext;\n [key: string]: unknown;\n}\n\n/**\n * Hono next function type\n */\nexport type HonoNextFunction = () => Promise<void>;\n\n/**\n * Hono middleware function type\n *\n * @param c - Hono context object\n * @param next - Next function to continue middleware chain\n */\nexport type HonoMiddleware = (c: HonoContext, next: HonoNextFunction) => Promise<void>;\n\n// ─────────────────────────────────────────────────────────────\n// Middleware\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create Hono middleware for automatic trace flushing\n *\n * On Cloudflare Workers/Deno Deploy: uses executionCtx.waitUntil() for non-blocking flush\n * On Node.js/Bun: flushes after response (fire-and-forget)\n *\n * @returns Hono middleware function\n *\n * @example\n * import { Hono } from 'hono';\n * import { createMiddleware } from '@lelemondev/sdk/hono';\n *\n * const app = new Hono();\n *\n * // Global middleware\n * app.use(createMiddleware());\n *\n * app.post('/chat', async (c) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return c.json(result);\n * });\n *\n * export default app;\n */\nexport function createMiddleware(): HonoMiddleware {\n return async (c, next) => {\n await next();\n\n // Use waitUntil if available (Cloudflare Workers, Deno Deploy)\n if (c.executionCtx?.waitUntil) {\n c.executionCtx.waitUntil(flush());\n } else {\n // Fire-and-forget for Node.js/Bun\n flush().catch(() => {});\n }\n };\n}\n"]}
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/integrations/next.ts","../src/core/config.ts","../src/integrations/lambda.ts","../src/integrations/express.ts","../src/integrations/hono.ts"],"names":["next_exports","__export","createWrapper","withObserve","flush","handler","options","request","context","defaultOptions","lambda_exports","event","express_exports","createMiddleware","_req","res","next","hono_exports","c"],"mappings":";AAAA,IAAA,CAAA,CAAA,MAAA,CAAA,cAAA,CAAA,IAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,GAAA,CAAA,IAAA,IAAA,CAAA,IAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,GAAA,CAAA,CAAA,CAAA,CAAA,CAAA,CAAA,UAAA,CAAA,IAAA,CAAA,EAAA,CAAA,CAAA,IAAAA,EAAA,GAAAC,CAAAA,CAAAD,EAAA,CAAA,aAAA,CAAA,IAAAE,CAAAA,CAAA,gBAAAC,CAAAA,CAAAA,CAAAA,CCwGA,eAAsBC,GAAuB,CAI7C,CDhBO,SAASD,CAAAA,CACdE,CAAAA,CACAC,EAC4B,CAC5B,aAAcC,CAAAA,CAAkBC,CAAAA,GAA0C,CACxE,GAAI,CACF,OAAO,MAAMH,CAAAA,CAAQE,EAASC,CAAO,CACvC,QAAE,CAEIF,CAAAA,EAAS,MAEXA,CAAAA,CAAQ,KAAA,CAAM,IAAMF,CAAAA,EAAO,CAAA,CAClBE,CAAAA,EAAS,UAElBA,CAAAA,CAAQ,SAAA,CAAUF,GAAO,CAAA,CAGzB,MAAMA,CAAAA,GAEV,CACF,CACF,CAeO,SAASF,CAAAA,CAAcO,CAAAA,CAAoC,CAChE,OAAO,SACLJ,EACAC,CAAAA,CAC4B,CAC5B,OAAOH,CAAAA,CAAYE,CAAAA,CAAS,CAAE,GAAGI,CAAAA,CAAgB,GAAGH,CAAQ,CAAC,CAC/D,CACF,KEvIAI,CAAAA,CAAA,GAAAT,EAAAS,CAAAA,CAAA,CAAA,WAAA,CAAA,IAAAP,IAqFO,SAASA,CAAAA,CACdE,CAAAA,CACgC,CAChC,OAAO,MAAOM,CAAAA,CAAeH,IAA6C,CACxE,GAAI,CACF,OAAO,MAAMH,EAAQM,CAAAA,CAAOH,CAAO,CACrC,CAAA,OAAE,CAEA,MAAMJ,CAAAA,GACR,CACF,CACF,CChGA,IAAAQ,CAAAA,CAAA,GAAAX,CAAAA,CAAAW,CAAAA,CAAA,sBAAAC,CAAAA,CAAAA,CAAAA,CA0EO,SAASA,GAAsC,CACpD,OAAO,CAACC,CAAAA,CAAMC,CAAAA,CAAKC,IAAS,CAE1BD,CAAAA,CAAI,GAAG,QAAA,CAAU,IAAM,CACrBX,CAAAA,EAAM,CAAE,KAAA,CAAM,IAAM,CAEpB,CAAC,EACH,CAAC,CAAA,CAEDY,CAAAA,GACF,CACF,KCrFAC,CAAAA,CAAA,GAAAhB,EAAAgB,CAAAA,CAAA,CAAA,gBAAA,CAAA,IAAAJ,IAmFO,SAASA,CAAAA,EAAmC,CACjD,OAAO,MAAOK,EAAGF,CAAAA,GAAS,CACxB,MAAMA,CAAAA,EAAK,CAGPE,EAAE,YAAA,EAAc,SAAA,CAClBA,EAAE,YAAA,CAAa,SAAA,CAAUd,GAAO,CAAA,CAGhCA,GAAM,CAAE,KAAA,CAAM,IAAM,CAAC,CAAC,EAE1B,CACF","file":"integrations.mjs","sourcesContent":["/**\n * Next.js App Router Integration\n *\n * Wraps route handlers to automatically flush traces.\n * Supports Next.js 15+ `after()` and Vercel's `waitUntil()`.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/next';\n *\n * export const POST = withObserve(async (req) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return Response.json(result);\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Next.js App Router handler type\n *\n * @typeParam TContext - Optional context type for dynamic route parameters\n */\nexport type NextRouteHandler<TContext = unknown> = (\n request: Request,\n context?: TContext\n) => Response | Promise<Response>;\n\n/**\n * Options for the Next.js wrapper\n */\nexport interface NextObserveOptions {\n /**\n * Next.js 15+ after() function from 'next/server'\n * Preferred method - runs after response without blocking\n *\n * @example\n * import { after } from 'next/server';\n * export const POST = withObserve(handler, { after });\n */\n after?: (callback: () => void | Promise<void>) => void;\n\n /**\n * Vercel's waitUntil() from '@vercel/functions'\n * Alternative for Vercel deployments\n *\n * @example\n * import { waitUntil } from '@vercel/functions';\n * export const POST = withObserve(handler, { waitUntil });\n */\n waitUntil?: (promise: Promise<unknown>) => void;\n}\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap a Next.js App Router handler with automatic trace flushing\n *\n * @param handler - Your route handler function\n * @param options - Optional: pass `after` (Next.js 15+) or `waitUntil` (Vercel)\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // Basic usage (blocking flush)\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n *\n * @example\n * // Next.js 15+ with after() - non-blocking (recommended)\n * import { after } from 'next/server';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { after }\n * );\n *\n * @example\n * // Vercel with waitUntil() - non-blocking\n * import { waitUntil } from '@vercel/functions';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { waitUntil }\n * );\n */\nexport function withObserve<TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n): NextRouteHandler<TContext> {\n return async (request: Request, context?: TContext): Promise<Response> => {\n try {\n return await handler(request, context);\n } finally {\n // Priority: after() > waitUntil() > blocking flush\n if (options?.after) {\n // Next.js 15+ native - best option\n options.after(() => flush());\n } else if (options?.waitUntil) {\n // Vercel platform\n options.waitUntil(flush());\n } else {\n // Fallback: blocking flush\n await flush();\n }\n }\n };\n}\n\n/**\n * Create a pre-configured wrapper with default options\n *\n * @example\n * import { after } from 'next/server';\n * import { createWrapper } from '@lelemondev/sdk/next';\n *\n * const withObserve = createWrapper({ after });\n *\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n */\nexport function createWrapper(defaultOptions: NextObserveOptions) {\n return function <TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n ): NextRouteHandler<TContext> {\n return withObserve(handler, { ...defaultOptions, ...options });\n };\n}\n","/**\n * Global Configuration\n *\n * Manages SDK configuration and transport instance.\n */\n\nimport type { LelemonConfig, SDKTelemetry } from './types';\nimport { Transport } from './transport';\nimport { setDebug, info, warn, debug } from './logger';\nimport { buildTelemetry } from './telemetry';\n\n// ─────────────────────────────────────────────────────────────\n// Global State\n// ─────────────────────────────────────────────────────────────\n\nlet globalConfig: LelemonConfig = {};\nlet globalTransport: Transport | null = null;\nlet globalTelemetry: SDKTelemetry | null = null;\nlet initialized = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\nconst DEFAULT_ENDPOINT = 'https://www.lelemon.dev';\n\n/**\n * Initialize the SDK\n * Call once at app startup\n */\nexport function init(config: LelemonConfig = {}): void {\n globalConfig = config;\n\n // Configure debug mode\n if (config.debug) {\n setDebug(true);\n }\n\n // Build telemetry with service config\n globalTelemetry = buildTelemetry(config.service);\n\n info('Initializing SDK', {\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? false,\n telemetry: globalTelemetry,\n });\n\n globalTransport = createTransport(config);\n initialized = true;\n\n // Log status after transport is created\n if (globalTransport.isEnabled()) {\n info('SDK initialized - tracing enabled');\n } else {\n debug('SDK initialized - tracing disabled (no API key or explicitly disabled)');\n }\n}\n\n/**\n * Get current config\n */\nexport function getConfig(): LelemonConfig {\n return globalConfig;\n}\n\n/**\n * Get SDK telemetry\n */\nexport function getTelemetry(): SDKTelemetry | null {\n return globalTelemetry;\n}\n\n/**\n * Check if SDK is initialized\n */\nexport function isInitialized(): boolean {\n return initialized;\n}\n\n/**\n * Check if SDK is enabled\n */\nexport function isEnabled(): boolean {\n return getTransport().isEnabled();\n}\n\n// ─────────────────────────────────────────────────────────────\n// Transport\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get or create transport instance\n */\nexport function getTransport(): Transport {\n if (!globalTransport) {\n globalTransport = createTransport(globalConfig);\n }\n return globalTransport;\n}\n\n/**\n * Flush all pending traces\n */\nexport async function flush(): Promise<void> {\n if (globalTransport) {\n await globalTransport.flush();\n }\n}\n\n/**\n * Create transport instance\n */\nfunction createTransport(config: LelemonConfig): Transport {\n const apiKey = config.apiKey ?? getEnvVar('LELEMON_API_KEY');\n\n if (!apiKey && !config.disabled) {\n warn('No API key provided. Set apiKey in init() or LELEMON_API_KEY env var. Tracing disabled.');\n }\n\n return new Transport({\n apiKey: apiKey ?? '',\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? !apiKey,\n batchSize: config.batchSize,\n flushIntervalMs: config.flushIntervalMs,\n requestTimeoutMs: config.requestTimeoutMs,\n });\n}\n\n/**\n * Get environment variable (works in Node and edge)\n */\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * AWS Lambda Integration\n *\n * Wraps Lambda handlers to automatically flush traces before the function exits.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/lambda';\n *\n * export const handler = withObserve(async (event) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return { statusCode: 200, body: JSON.stringify(result) };\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring @types/aws-lambda)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * AWS Lambda Context object\n */\nexport interface LambdaContext {\n functionName: string;\n functionVersion: string;\n invokedFunctionArn: string;\n memoryLimitInMB: string;\n awsRequestId: string;\n logGroupName: string;\n logStreamName: string;\n getRemainingTimeInMillis(): number;\n [key: string]: unknown;\n}\n\n/**\n * Generic AWS Lambda handler type\n *\n * @typeParam TEvent - The event type (e.g., APIGatewayProxyEvent)\n * @typeParam TResult - The result type (e.g., APIGatewayProxyResult)\n */\nexport type LambdaHandler<TEvent = unknown, TResult = unknown> = (\n event: TEvent,\n context: LambdaContext\n) => Promise<TResult>;\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap an AWS Lambda handler with automatic trace flushing\n *\n * Always flushes before returning - Lambda freezes the container\n * immediately after the handler returns, so this is required.\n *\n * @param handler - Your Lambda handler function\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // API Gateway event\n * export const handler = withObserve(async (event) => {\n * const body = JSON.parse(event.body);\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({\n * model: 'gpt-4',\n * messages: [{ role: 'user', content: body.message }],\n * });\n * return {\n * statusCode: 200,\n * body: JSON.stringify(result.choices[0].message),\n * };\n * });\n *\n * @example\n * // With typed events\n * import type { APIGatewayProxyEvent, APIGatewayProxyResult } from 'aws-lambda';\n *\n * export const handler = withObserve<APIGatewayProxyEvent, APIGatewayProxyResult>(\n * async (event, context) => {\n * return { statusCode: 200, body: 'OK' };\n * }\n * );\n */\nexport function withObserve<TEvent = unknown, TResult = unknown>(\n handler: LambdaHandler<TEvent, TResult>\n): LambdaHandler<TEvent, TResult> {\n return async (event: TEvent, context: LambdaContext): Promise<TResult> => {\n try {\n return await handler(event, context);\n } finally {\n // Always flush - Lambda freezes immediately after return\n await flush();\n }\n };\n}\n","/**\n * Express Integration\n *\n * Middleware that automatically flushes traces when response finishes.\n *\n * @example\n * import express from 'express';\n * import { createMiddleware } from '@lelemondev/sdk/express';\n *\n * const app = express();\n * app.use(createMiddleware());\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring express as dependency)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Minimal Express request type (avoids requiring express as dependency)\n */\nexport interface ExpressRequest {\n [key: string]: unknown;\n}\n\n/**\n * Minimal Express response type (avoids requiring express as dependency)\n */\nexport interface ExpressResponse {\n on(event: 'finish' | 'close' | 'error', listener: () => void): this;\n [key: string]: unknown;\n}\n\n/**\n * Express next function type\n */\nexport type ExpressNextFunction = (error?: unknown) => void;\n\n/**\n * Express middleware function type\n *\n * @param req - Express request object\n * @param res - Express response object\n * @param next - Next function to pass control\n */\nexport type ExpressMiddleware = (\n req: ExpressRequest,\n res: ExpressResponse,\n next: ExpressNextFunction\n) => void;\n\n// ─────────────────────────────────────────────────────────────\n// Middleware\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create Express middleware for automatic trace flushing\n *\n * Flushes traces when the response finishes (after res.send/res.json).\n * This is fire-and-forget and doesn't block the response.\n *\n * @returns Express middleware function\n *\n * @example\n * // Global middleware\n * app.use(createMiddleware());\n *\n * @example\n * // Per-route middleware\n * app.post('/chat', createMiddleware(), async (req, res) => {\n * res.json({ ok: true });\n * });\n */\nexport function createMiddleware(): ExpressMiddleware {\n return (_req, res, next) => {\n // Flush when response is finished (after headers + body sent)\n res.on('finish', () => {\n flush().catch(() => {\n // Silently ignore flush errors - fire and forget\n });\n });\n\n next();\n };\n}\n","/**\n * Hono Integration\n *\n * Middleware for Hono framework (Cloudflare Workers, Deno, Bun, Node.js).\n * Uses executionCtx.waitUntil() when available for non-blocking flush.\n *\n * @example\n * import { Hono } from 'hono';\n * import { createMiddleware } from '@lelemondev/sdk/hono';\n *\n * const app = new Hono();\n * app.use(createMiddleware());\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring hono as dependency)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Execution context for edge runtimes (Cloudflare Workers, Deno Deploy)\n */\nexport interface ExecutionContext {\n waitUntil(promise: Promise<unknown>): void;\n passThroughOnException(): void;\n}\n\n/**\n * Minimal Hono context type (avoids requiring hono as dependency)\n */\nexport interface HonoContext {\n req: {\n raw: Request;\n [key: string]: unknown;\n };\n res: Response | undefined;\n executionCtx?: ExecutionContext;\n [key: string]: unknown;\n}\n\n/**\n * Hono next function type\n */\nexport type HonoNextFunction = () => Promise<void>;\n\n/**\n * Hono middleware function type\n *\n * @param c - Hono context object\n * @param next - Next function to continue middleware chain\n */\nexport type HonoMiddleware = (c: HonoContext, next: HonoNextFunction) => Promise<void>;\n\n// ─────────────────────────────────────────────────────────────\n// Middleware\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create Hono middleware for automatic trace flushing\n *\n * On Cloudflare Workers/Deno Deploy: uses executionCtx.waitUntil() for non-blocking flush\n * On Node.js/Bun: flushes after response (fire-and-forget)\n *\n * @returns Hono middleware function\n *\n * @example\n * import { Hono } from 'hono';\n * import { createMiddleware } from '@lelemondev/sdk/hono';\n *\n * const app = new Hono();\n *\n * // Global middleware\n * app.use(createMiddleware());\n *\n * app.post('/chat', async (c) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return c.json(result);\n * });\n *\n * export default app;\n */\nexport function createMiddleware(): HonoMiddleware {\n return async (c, next) => {\n await next();\n\n // Use waitUntil if available (Cloudflare Workers, Deno Deploy)\n if (c.executionCtx?.waitUntil) {\n c.executionCtx.waitUntil(flush());\n } else {\n // Fire-and-forget for Node.js/Bun\n flush().catch(() => {});\n }\n };\n}\n"]}
|
package/dist/lambda.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/core/config.ts","../src/integrations/lambda.ts"],"names":["flush","withObserve","handler","event","context"],"mappings":";AAwGA,eAAsBA,CAAAA,EAAuB,CAI7C,CCvBO,SAASC,CAAAA,CACdC,CAAAA,CACgC,CAChC,OAAO,MAAOC,CAAAA,CAAeC,CAAAA,GAA6C,CACxE,GAAI,CACF,OAAO,MAAMF,CAAAA,CAAQC,CAAAA,CAAOC,CAAO,CACrC,QAAE,CAEA,MAAMJ,CAAAA,GACR,CACF,CACF","file":"lambda.js","sourcesContent":["/**\n * Global Configuration\n *\n * Manages SDK configuration and transport instance.\n */\n\nimport type { LelemonConfig, SDKTelemetry } from './types';\nimport { Transport } from './transport';\nimport { setDebug, info, warn, debug } from './logger';\nimport { buildTelemetry } from './telemetry';\n\n// ─────────────────────────────────────────────────────────────\n// Global State\n// ─────────────────────────────────────────────────────────────\n\nlet globalConfig: LelemonConfig = {};\nlet globalTransport: Transport | null = null;\nlet globalTelemetry: SDKTelemetry | null = null;\nlet initialized = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\nconst DEFAULT_ENDPOINT = 'https://www.lelemon.dev';\n\n/**\n * Initialize the SDK\n * Call once at app startup\n */\nexport function init(config: LelemonConfig = {}): void {\n globalConfig = config;\n\n // Configure debug mode\n if (config.debug) {\n setDebug(true);\n }\n\n // Build telemetry with service config\n globalTelemetry = buildTelemetry(config.service);\n\n info('Initializing SDK', {\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? false,\n telemetry: globalTelemetry,\n });\n\n globalTransport = createTransport(config);\n initialized = true;\n\n // Log status after transport is created\n if (globalTransport.isEnabled()) {\n info('SDK initialized - tracing enabled');\n } else {\n debug('SDK initialized - tracing disabled (no API key or explicitly disabled)');\n }\n}\n\n/**\n * Get current config\n */\nexport function getConfig(): LelemonConfig {\n return globalConfig;\n}\n\n/**\n * Get SDK telemetry\n */\nexport function getTelemetry(): SDKTelemetry | null {\n return globalTelemetry;\n}\n\n/**\n * Check if SDK is initialized\n */\nexport function isInitialized(): boolean {\n return initialized;\n}\n\n/**\n * Check if SDK is enabled\n */\nexport function isEnabled(): boolean {\n return getTransport().isEnabled();\n}\n\n// ─────────────────────────────────────────────────────────────\n// Transport\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get or create transport instance\n */\nexport function getTransport(): Transport {\n if (!globalTransport) {\n globalTransport = createTransport(globalConfig);\n }\n return globalTransport;\n}\n\n/**\n * Flush all pending traces\n */\nexport async function flush(): Promise<void> {\n if (globalTransport) {\n await globalTransport.flush();\n }\n}\n\n/**\n * Create transport instance\n */\nfunction createTransport(config: LelemonConfig): Transport {\n const apiKey = config.apiKey ?? getEnvVar('LELEMON_API_KEY');\n\n if (!apiKey && !config.disabled) {\n warn('No API key provided. Set apiKey in init() or LELEMON_API_KEY env var. Tracing disabled.');\n }\n\n return new Transport({\n apiKey: apiKey ?? '',\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? !apiKey,\n batchSize: config.batchSize,\n flushIntervalMs: config.flushIntervalMs,\n requestTimeoutMs: config.requestTimeoutMs,\n });\n}\n\n/**\n * Get environment variable (works in Node and edge)\n */\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * AWS Lambda Integration\n *\n * Wraps Lambda handlers to automatically flush traces before the function exits.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/lambda';\n *\n * export const handler = withObserve(async (event) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return { statusCode: 200, body: JSON.stringify(result) };\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring @types/aws-lambda)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * AWS Lambda Context object\n */\nexport interface LambdaContext {\n functionName: string;\n functionVersion: string;\n invokedFunctionArn: string;\n memoryLimitInMB: string;\n awsRequestId: string;\n logGroupName: string;\n logStreamName: string;\n getRemainingTimeInMillis(): number;\n [key: string]: unknown;\n}\n\n/**\n * Generic AWS Lambda handler type\n *\n * @typeParam TEvent - The event type (e.g., APIGatewayProxyEvent)\n * @typeParam TResult - The result type (e.g., APIGatewayProxyResult)\n */\nexport type LambdaHandler<TEvent = unknown, TResult = unknown> = (\n event: TEvent,\n context: LambdaContext\n) => Promise<TResult>;\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap an AWS Lambda handler with automatic trace flushing\n *\n * Always flushes before returning - Lambda freezes the container\n * immediately after the handler returns, so this is required.\n *\n * @param handler - Your Lambda handler function\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // API Gateway event\n * export const handler = withObserve(async (event) => {\n * const body = JSON.parse(event.body);\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({\n * model: 'gpt-4',\n * messages: [{ role: 'user', content: body.message }],\n * });\n * return {\n * statusCode: 200,\n * body: JSON.stringify(result.choices[0].message),\n * };\n * });\n *\n * @example\n * // With typed events\n * import type { APIGatewayProxyEvent, APIGatewayProxyResult } from 'aws-lambda';\n *\n * export const handler = withObserve<APIGatewayProxyEvent, APIGatewayProxyResult>(\n * async (event, context) => {\n * return { statusCode: 200, body: 'OK' };\n * }\n * );\n */\nexport function withObserve<TEvent = unknown, TResult = unknown>(\n handler: LambdaHandler<TEvent, TResult>\n): LambdaHandler<TEvent, TResult> {\n return async (event: TEvent, context: LambdaContext): Promise<TResult> => {\n try {\n return await handler(event, context);\n } finally {\n // Always flush - Lambda freezes immediately after return\n await flush();\n }\n };\n}\n"]}
|
package/dist/lambda.mjs.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/core/config.ts","../src/integrations/lambda.ts"],"names":["flush","withObserve","handler","event","context"],"mappings":";AAwGA,eAAsBA,CAAAA,EAAuB,CAI7C,CCvBO,SAASC,CAAAA,CACdC,CAAAA,CACgC,CAChC,OAAO,MAAOC,CAAAA,CAAeC,CAAAA,GAA6C,CACxE,GAAI,CACF,OAAO,MAAMF,CAAAA,CAAQC,CAAAA,CAAOC,CAAO,CACrC,QAAE,CAEA,MAAMJ,CAAAA,GACR,CACF,CACF","file":"lambda.mjs","sourcesContent":["/**\n * Global Configuration\n *\n * Manages SDK configuration and transport instance.\n */\n\nimport type { LelemonConfig, SDKTelemetry } from './types';\nimport { Transport } from './transport';\nimport { setDebug, info, warn, debug } from './logger';\nimport { buildTelemetry } from './telemetry';\n\n// ─────────────────────────────────────────────────────────────\n// Global State\n// ─────────────────────────────────────────────────────────────\n\nlet globalConfig: LelemonConfig = {};\nlet globalTransport: Transport | null = null;\nlet globalTelemetry: SDKTelemetry | null = null;\nlet initialized = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\nconst DEFAULT_ENDPOINT = 'https://www.lelemon.dev';\n\n/**\n * Initialize the SDK\n * Call once at app startup\n */\nexport function init(config: LelemonConfig = {}): void {\n globalConfig = config;\n\n // Configure debug mode\n if (config.debug) {\n setDebug(true);\n }\n\n // Build telemetry with service config\n globalTelemetry = buildTelemetry(config.service);\n\n info('Initializing SDK', {\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? false,\n telemetry: globalTelemetry,\n });\n\n globalTransport = createTransport(config);\n initialized = true;\n\n // Log status after transport is created\n if (globalTransport.isEnabled()) {\n info('SDK initialized - tracing enabled');\n } else {\n debug('SDK initialized - tracing disabled (no API key or explicitly disabled)');\n }\n}\n\n/**\n * Get current config\n */\nexport function getConfig(): LelemonConfig {\n return globalConfig;\n}\n\n/**\n * Get SDK telemetry\n */\nexport function getTelemetry(): SDKTelemetry | null {\n return globalTelemetry;\n}\n\n/**\n * Check if SDK is initialized\n */\nexport function isInitialized(): boolean {\n return initialized;\n}\n\n/**\n * Check if SDK is enabled\n */\nexport function isEnabled(): boolean {\n return getTransport().isEnabled();\n}\n\n// ─────────────────────────────────────────────────────────────\n// Transport\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get or create transport instance\n */\nexport function getTransport(): Transport {\n if (!globalTransport) {\n globalTransport = createTransport(globalConfig);\n }\n return globalTransport;\n}\n\n/**\n * Flush all pending traces\n */\nexport async function flush(): Promise<void> {\n if (globalTransport) {\n await globalTransport.flush();\n }\n}\n\n/**\n * Create transport instance\n */\nfunction createTransport(config: LelemonConfig): Transport {\n const apiKey = config.apiKey ?? getEnvVar('LELEMON_API_KEY');\n\n if (!apiKey && !config.disabled) {\n warn('No API key provided. Set apiKey in init() or LELEMON_API_KEY env var. Tracing disabled.');\n }\n\n return new Transport({\n apiKey: apiKey ?? '',\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? !apiKey,\n batchSize: config.batchSize,\n flushIntervalMs: config.flushIntervalMs,\n requestTimeoutMs: config.requestTimeoutMs,\n });\n}\n\n/**\n * Get environment variable (works in Node and edge)\n */\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * AWS Lambda Integration\n *\n * Wraps Lambda handlers to automatically flush traces before the function exits.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/lambda';\n *\n * export const handler = withObserve(async (event) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return { statusCode: 200, body: JSON.stringify(result) };\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types (minimal to avoid requiring @types/aws-lambda)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * AWS Lambda Context object\n */\nexport interface LambdaContext {\n functionName: string;\n functionVersion: string;\n invokedFunctionArn: string;\n memoryLimitInMB: string;\n awsRequestId: string;\n logGroupName: string;\n logStreamName: string;\n getRemainingTimeInMillis(): number;\n [key: string]: unknown;\n}\n\n/**\n * Generic AWS Lambda handler type\n *\n * @typeParam TEvent - The event type (e.g., APIGatewayProxyEvent)\n * @typeParam TResult - The result type (e.g., APIGatewayProxyResult)\n */\nexport type LambdaHandler<TEvent = unknown, TResult = unknown> = (\n event: TEvent,\n context: LambdaContext\n) => Promise<TResult>;\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap an AWS Lambda handler with automatic trace flushing\n *\n * Always flushes before returning - Lambda freezes the container\n * immediately after the handler returns, so this is required.\n *\n * @param handler - Your Lambda handler function\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // API Gateway event\n * export const handler = withObserve(async (event) => {\n * const body = JSON.parse(event.body);\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({\n * model: 'gpt-4',\n * messages: [{ role: 'user', content: body.message }],\n * });\n * return {\n * statusCode: 200,\n * body: JSON.stringify(result.choices[0].message),\n * };\n * });\n *\n * @example\n * // With typed events\n * import type { APIGatewayProxyEvent, APIGatewayProxyResult } from 'aws-lambda';\n *\n * export const handler = withObserve<APIGatewayProxyEvent, APIGatewayProxyResult>(\n * async (event, context) => {\n * return { statusCode: 200, body: 'OK' };\n * }\n * );\n */\nexport function withObserve<TEvent = unknown, TResult = unknown>(\n handler: LambdaHandler<TEvent, TResult>\n): LambdaHandler<TEvent, TResult> {\n return async (event: TEvent, context: LambdaContext): Promise<TResult> => {\n try {\n return await handler(event, context);\n } finally {\n // Always flush - Lambda freezes immediately after return\n await flush();\n }\n };\n}\n"]}
|
package/dist/next.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/core/config.ts","../src/integrations/next.ts"],"names":["flush","withObserve","handler","options","request","context","createWrapper","defaultOptions"],"mappings":";AAwGA,eAAsBA,CAAAA,EAAuB,CAI7C,CChBO,SAASC,CAAAA,CACdC,EACAC,CAAAA,CAC4B,CAC5B,OAAO,MAAOC,CAAAA,CAAkBC,CAAAA,GAA0C,CACxE,GAAI,CACF,OAAO,MAAMH,CAAAA,CAAQE,CAAAA,CAASC,CAAO,CACvC,CAAA,OAAE,CAEIF,CAAAA,EAAS,MAEXA,CAAAA,CAAQ,KAAA,CAAM,IAAMH,CAAAA,EAAO,CAAA,CAClBG,CAAAA,EAAS,SAAA,CAElBA,EAAQ,SAAA,CAAUH,CAAAA,EAAO,CAAA,CAGzB,MAAMA,CAAAA,GAEV,CACF,CACF,CAeO,SAASM,CAAAA,CAAcC,CAAAA,CAAoC,CAChE,OAAO,SACLL,CAAAA,CACAC,CAAAA,CAC4B,CAC5B,OAAOF,CAAAA,CAAYC,CAAAA,CAAS,CAAE,GAAGK,CAAAA,CAAgB,GAAGJ,CAAQ,CAAC,CAC/D,CACF","file":"next.js","sourcesContent":["/**\n * Global Configuration\n *\n * Manages SDK configuration and transport instance.\n */\n\nimport type { LelemonConfig, SDKTelemetry } from './types';\nimport { Transport } from './transport';\nimport { setDebug, info, warn, debug } from './logger';\nimport { buildTelemetry } from './telemetry';\n\n// ─────────────────────────────────────────────────────────────\n// Global State\n// ─────────────────────────────────────────────────────────────\n\nlet globalConfig: LelemonConfig = {};\nlet globalTransport: Transport | null = null;\nlet globalTelemetry: SDKTelemetry | null = null;\nlet initialized = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\nconst DEFAULT_ENDPOINT = 'https://www.lelemon.dev';\n\n/**\n * Initialize the SDK\n * Call once at app startup\n */\nexport function init(config: LelemonConfig = {}): void {\n globalConfig = config;\n\n // Configure debug mode\n if (config.debug) {\n setDebug(true);\n }\n\n // Build telemetry with service config\n globalTelemetry = buildTelemetry(config.service);\n\n info('Initializing SDK', {\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? false,\n telemetry: globalTelemetry,\n });\n\n globalTransport = createTransport(config);\n initialized = true;\n\n // Log status after transport is created\n if (globalTransport.isEnabled()) {\n info('SDK initialized - tracing enabled');\n } else {\n debug('SDK initialized - tracing disabled (no API key or explicitly disabled)');\n }\n}\n\n/**\n * Get current config\n */\nexport function getConfig(): LelemonConfig {\n return globalConfig;\n}\n\n/**\n * Get SDK telemetry\n */\nexport function getTelemetry(): SDKTelemetry | null {\n return globalTelemetry;\n}\n\n/**\n * Check if SDK is initialized\n */\nexport function isInitialized(): boolean {\n return initialized;\n}\n\n/**\n * Check if SDK is enabled\n */\nexport function isEnabled(): boolean {\n return getTransport().isEnabled();\n}\n\n// ─────────────────────────────────────────────────────────────\n// Transport\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get or create transport instance\n */\nexport function getTransport(): Transport {\n if (!globalTransport) {\n globalTransport = createTransport(globalConfig);\n }\n return globalTransport;\n}\n\n/**\n * Flush all pending traces\n */\nexport async function flush(): Promise<void> {\n if (globalTransport) {\n await globalTransport.flush();\n }\n}\n\n/**\n * Create transport instance\n */\nfunction createTransport(config: LelemonConfig): Transport {\n const apiKey = config.apiKey ?? getEnvVar('LELEMON_API_KEY');\n\n if (!apiKey && !config.disabled) {\n warn('No API key provided. Set apiKey in init() or LELEMON_API_KEY env var. Tracing disabled.');\n }\n\n return new Transport({\n apiKey: apiKey ?? '',\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? !apiKey,\n batchSize: config.batchSize,\n flushIntervalMs: config.flushIntervalMs,\n requestTimeoutMs: config.requestTimeoutMs,\n });\n}\n\n/**\n * Get environment variable (works in Node and edge)\n */\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * Next.js App Router Integration\n *\n * Wraps route handlers to automatically flush traces.\n * Supports Next.js 15+ `after()` and Vercel's `waitUntil()`.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/next';\n *\n * export const POST = withObserve(async (req) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return Response.json(result);\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Next.js App Router handler type\n *\n * @typeParam TContext - Optional context type for dynamic route parameters\n */\nexport type NextRouteHandler<TContext = unknown> = (\n request: Request,\n context?: TContext\n) => Response | Promise<Response>;\n\n/**\n * Options for the Next.js wrapper\n */\nexport interface NextObserveOptions {\n /**\n * Next.js 15+ after() function from 'next/server'\n * Preferred method - runs after response without blocking\n *\n * @example\n * import { after } from 'next/server';\n * export const POST = withObserve(handler, { after });\n */\n after?: (callback: () => void | Promise<void>) => void;\n\n /**\n * Vercel's waitUntil() from '@vercel/functions'\n * Alternative for Vercel deployments\n *\n * @example\n * import { waitUntil } from '@vercel/functions';\n * export const POST = withObserve(handler, { waitUntil });\n */\n waitUntil?: (promise: Promise<unknown>) => void;\n}\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap a Next.js App Router handler with automatic trace flushing\n *\n * @param handler - Your route handler function\n * @param options - Optional: pass `after` (Next.js 15+) or `waitUntil` (Vercel)\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // Basic usage (blocking flush)\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n *\n * @example\n * // Next.js 15+ with after() - non-blocking (recommended)\n * import { after } from 'next/server';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { after }\n * );\n *\n * @example\n * // Vercel with waitUntil() - non-blocking\n * import { waitUntil } from '@vercel/functions';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { waitUntil }\n * );\n */\nexport function withObserve<TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n): NextRouteHandler<TContext> {\n return async (request: Request, context?: TContext): Promise<Response> => {\n try {\n return await handler(request, context);\n } finally {\n // Priority: after() > waitUntil() > blocking flush\n if (options?.after) {\n // Next.js 15+ native - best option\n options.after(() => flush());\n } else if (options?.waitUntil) {\n // Vercel platform\n options.waitUntil(flush());\n } else {\n // Fallback: blocking flush\n await flush();\n }\n }\n };\n}\n\n/**\n * Create a pre-configured wrapper with default options\n *\n * @example\n * import { after } from 'next/server';\n * import { createWrapper } from '@lelemondev/sdk/next';\n *\n * const withObserve = createWrapper({ after });\n *\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n */\nexport function createWrapper(defaultOptions: NextObserveOptions) {\n return function <TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n ): NextRouteHandler<TContext> {\n return withObserve(handler, { ...defaultOptions, ...options });\n };\n}\n"]}
|
package/dist/next.mjs.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/core/config.ts","../src/integrations/next.ts"],"names":["flush","withObserve","handler","options","request","context","createWrapper","defaultOptions"],"mappings":";AAwGA,eAAsBA,CAAAA,EAAuB,CAI7C,CChBO,SAASC,CAAAA,CACdC,EACAC,CAAAA,CAC4B,CAC5B,OAAO,MAAOC,CAAAA,CAAkBC,CAAAA,GAA0C,CACxE,GAAI,CACF,OAAO,MAAMH,CAAAA,CAAQE,CAAAA,CAASC,CAAO,CACvC,CAAA,OAAE,CAEIF,CAAAA,EAAS,MAEXA,CAAAA,CAAQ,KAAA,CAAM,IAAMH,CAAAA,EAAO,CAAA,CAClBG,CAAAA,EAAS,SAAA,CAElBA,EAAQ,SAAA,CAAUH,CAAAA,EAAO,CAAA,CAGzB,MAAMA,CAAAA,GAEV,CACF,CACF,CAeO,SAASM,CAAAA,CAAcC,CAAAA,CAAoC,CAChE,OAAO,SACLL,CAAAA,CACAC,CAAAA,CAC4B,CAC5B,OAAOF,CAAAA,CAAYC,CAAAA,CAAS,CAAE,GAAGK,CAAAA,CAAgB,GAAGJ,CAAQ,CAAC,CAC/D,CACF","file":"next.mjs","sourcesContent":["/**\n * Global Configuration\n *\n * Manages SDK configuration and transport instance.\n */\n\nimport type { LelemonConfig, SDKTelemetry } from './types';\nimport { Transport } from './transport';\nimport { setDebug, info, warn, debug } from './logger';\nimport { buildTelemetry } from './telemetry';\n\n// ─────────────────────────────────────────────────────────────\n// Global State\n// ─────────────────────────────────────────────────────────────\n\nlet globalConfig: LelemonConfig = {};\nlet globalTransport: Transport | null = null;\nlet globalTelemetry: SDKTelemetry | null = null;\nlet initialized = false;\n\n// ─────────────────────────────────────────────────────────────\n// Configuration\n// ─────────────────────────────────────────────────────────────\n\nconst DEFAULT_ENDPOINT = 'https://www.lelemon.dev';\n\n/**\n * Initialize the SDK\n * Call once at app startup\n */\nexport function init(config: LelemonConfig = {}): void {\n globalConfig = config;\n\n // Configure debug mode\n if (config.debug) {\n setDebug(true);\n }\n\n // Build telemetry with service config\n globalTelemetry = buildTelemetry(config.service);\n\n info('Initializing SDK', {\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? false,\n telemetry: globalTelemetry,\n });\n\n globalTransport = createTransport(config);\n initialized = true;\n\n // Log status after transport is created\n if (globalTransport.isEnabled()) {\n info('SDK initialized - tracing enabled');\n } else {\n debug('SDK initialized - tracing disabled (no API key or explicitly disabled)');\n }\n}\n\n/**\n * Get current config\n */\nexport function getConfig(): LelemonConfig {\n return globalConfig;\n}\n\n/**\n * Get SDK telemetry\n */\nexport function getTelemetry(): SDKTelemetry | null {\n return globalTelemetry;\n}\n\n/**\n * Check if SDK is initialized\n */\nexport function isInitialized(): boolean {\n return initialized;\n}\n\n/**\n * Check if SDK is enabled\n */\nexport function isEnabled(): boolean {\n return getTransport().isEnabled();\n}\n\n// ─────────────────────────────────────────────────────────────\n// Transport\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Get or create transport instance\n */\nexport function getTransport(): Transport {\n if (!globalTransport) {\n globalTransport = createTransport(globalConfig);\n }\n return globalTransport;\n}\n\n/**\n * Flush all pending traces\n */\nexport async function flush(): Promise<void> {\n if (globalTransport) {\n await globalTransport.flush();\n }\n}\n\n/**\n * Create transport instance\n */\nfunction createTransport(config: LelemonConfig): Transport {\n const apiKey = config.apiKey ?? getEnvVar('LELEMON_API_KEY');\n\n if (!apiKey && !config.disabled) {\n warn('No API key provided. Set apiKey in init() or LELEMON_API_KEY env var. Tracing disabled.');\n }\n\n return new Transport({\n apiKey: apiKey ?? '',\n endpoint: config.endpoint ?? DEFAULT_ENDPOINT,\n debug: config.debug ?? false,\n disabled: config.disabled ?? !apiKey,\n batchSize: config.batchSize,\n flushIntervalMs: config.flushIntervalMs,\n requestTimeoutMs: config.requestTimeoutMs,\n });\n}\n\n/**\n * Get environment variable (works in Node and edge)\n */\nfunction getEnvVar(name: string): string | undefined {\n if (typeof process !== 'undefined' && process.env) {\n return process.env[name];\n }\n return undefined;\n}\n","/**\n * Next.js App Router Integration\n *\n * Wraps route handlers to automatically flush traces.\n * Supports Next.js 15+ `after()` and Vercel's `waitUntil()`.\n *\n * @example\n * import { withObserve } from '@lelemondev/sdk/next';\n *\n * export const POST = withObserve(async (req) => {\n * const openai = observe(new OpenAI());\n * const result = await openai.chat.completions.create({...});\n * return Response.json(result);\n * });\n */\n\nimport { flush } from '../core/config';\n\n// ─────────────────────────────────────────────────────────────\n// Types\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Next.js App Router handler type\n *\n * @typeParam TContext - Optional context type for dynamic route parameters\n */\nexport type NextRouteHandler<TContext = unknown> = (\n request: Request,\n context?: TContext\n) => Response | Promise<Response>;\n\n/**\n * Options for the Next.js wrapper\n */\nexport interface NextObserveOptions {\n /**\n * Next.js 15+ after() function from 'next/server'\n * Preferred method - runs after response without blocking\n *\n * @example\n * import { after } from 'next/server';\n * export const POST = withObserve(handler, { after });\n */\n after?: (callback: () => void | Promise<void>) => void;\n\n /**\n * Vercel's waitUntil() from '@vercel/functions'\n * Alternative for Vercel deployments\n *\n * @example\n * import { waitUntil } from '@vercel/functions';\n * export const POST = withObserve(handler, { waitUntil });\n */\n waitUntil?: (promise: Promise<unknown>) => void;\n}\n\n// ─────────────────────────────────────────────────────────────\n// Wrapper\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Wrap a Next.js App Router handler with automatic trace flushing\n *\n * @param handler - Your route handler function\n * @param options - Optional: pass `after` (Next.js 15+) or `waitUntil` (Vercel)\n * @returns Wrapped handler that auto-flushes traces\n *\n * @example\n * // Basic usage (blocking flush)\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n *\n * @example\n * // Next.js 15+ with after() - non-blocking (recommended)\n * import { after } from 'next/server';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { after }\n * );\n *\n * @example\n * // Vercel with waitUntil() - non-blocking\n * import { waitUntil } from '@vercel/functions';\n *\n * export const POST = withObserve(\n * async (req) => Response.json({ ok: true }),\n * { waitUntil }\n * );\n */\nexport function withObserve<TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n): NextRouteHandler<TContext> {\n return async (request: Request, context?: TContext): Promise<Response> => {\n try {\n return await handler(request, context);\n } finally {\n // Priority: after() > waitUntil() > blocking flush\n if (options?.after) {\n // Next.js 15+ native - best option\n options.after(() => flush());\n } else if (options?.waitUntil) {\n // Vercel platform\n options.waitUntil(flush());\n } else {\n // Fallback: blocking flush\n await flush();\n }\n }\n };\n}\n\n/**\n * Create a pre-configured wrapper with default options\n *\n * @example\n * import { after } from 'next/server';\n * import { createWrapper } from '@lelemondev/sdk/next';\n *\n * const withObserve = createWrapper({ after });\n *\n * export const POST = withObserve(async (req) => {\n * return Response.json({ ok: true });\n * });\n */\nexport function createWrapper(defaultOptions: NextObserveOptions) {\n return function <TContext = unknown>(\n handler: NextRouteHandler<TContext>,\n options?: NextObserveOptions\n ): NextRouteHandler<TContext> {\n return withObserve(handler, { ...defaultOptions, ...options });\n };\n}\n"]}
|