@ai-sdk/lmnt 2.0.8 → 2.0.10
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +12 -0
- package/dist/index.js +1 -1
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +1 -1
- package/dist/index.mjs.map +1 -1
- package/docs/140-lmnt.mdx +140 -0
- package/package.json +8 -3
- package/src/index.ts +3 -0
- package/src/lmnt-api-types.ts +39 -0
- package/src/lmnt-config.ts +9 -0
- package/src/lmnt-error.test.ts +34 -0
- package/src/lmnt-error.ts +16 -0
- package/src/lmnt-provider.ts +83 -0
- package/src/lmnt-speech-model.test.ts +197 -0
- package/src/lmnt-speech-model.ts +206 -0
- package/src/lmnt-speech-options.ts +1 -0
- package/src/version.ts +6 -0
package/CHANGELOG.md
CHANGED
package/dist/index.js
CHANGED
package/dist/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/index.ts","../src/lmnt-provider.ts","../src/lmnt-speech-model.ts","../src/lmnt-error.ts","../src/version.ts"],"sourcesContent":["export { createLMNT, lmnt } from './lmnt-provider';\nexport type { LMNTProvider, LMNTProviderSettings } from './lmnt-provider';\nexport { VERSION } from './version';\n","import { SpeechModelV3, ProviderV3 } from '@ai-sdk/provider';\nimport {\n FetchFunction,\n loadApiKey,\n withUserAgentSuffix,\n} from '@ai-sdk/provider-utils';\nimport { LMNTSpeechModel } from './lmnt-speech-model';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { VERSION } from './version';\n\nexport interface LMNTProvider extends Pick<ProviderV3, 'speechModel'> {\n (\n modelId: 'aurora',\n settings?: {},\n ): {\n speech: LMNTSpeechModel;\n };\n\n /**\nCreates a model for speech synthesis.\n */\n speech(modelId: LMNTSpeechModelId): SpeechModelV3;\n}\n\nexport interface LMNTProviderSettings {\n /**\nAPI key for authenticating requests.\n */\n apiKey?: string;\n\n /**\nCustom headers to include in the requests.\n */\n headers?: Record<string, string>;\n\n /**\nCustom fetch implementation. You can use it as a middleware to intercept requests,\nor to provide a custom fetch implementation for e.g. testing.\n */\n fetch?: FetchFunction;\n}\n\n/**\nCreate an LMNT provider instance.\n */\nexport function createLMNT(options: LMNTProviderSettings = {}): LMNTProvider {\n const getHeaders = () =>\n withUserAgentSuffix(\n {\n 'x-api-key': loadApiKey({\n apiKey: options.apiKey,\n environmentVariableName: 'LMNT_API_KEY',\n description: 'LMNT',\n }),\n ...options.headers,\n },\n `ai-sdk/lmnt/${VERSION}`,\n );\n\n const createSpeechModel = (modelId: LMNTSpeechModelId) =>\n new LMNTSpeechModel(modelId, {\n provider: `lmnt.speech`,\n url: ({ path }) => `https://api.lmnt.com${path}`,\n headers: getHeaders,\n fetch: options.fetch,\n });\n\n const provider = function (modelId: LMNTSpeechModelId) {\n return {\n speech: createSpeechModel(modelId),\n };\n };\n\n provider.speech = createSpeechModel;\n provider.speechModel = createSpeechModel;\n\n return provider as LMNTProvider;\n}\n\n/**\nDefault LMNT provider instance.\n */\nexport const lmnt = createLMNT();\n","import { SpeechModelV3, SharedV3Warning } from '@ai-sdk/provider';\nimport {\n combineHeaders,\n createBinaryResponseHandler,\n parseProviderOptions,\n postJsonToApi,\n} from '@ai-sdk/provider-utils';\nimport { z } from 'zod/v4';\nimport { LMNTConfig } from './lmnt-config';\nimport { lmntFailedResponseHandler } from './lmnt-error';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { LMNTSpeechAPITypes } from './lmnt-api-types';\n\n// https://docs.lmnt.com/api-reference/speech/synthesize-speech-bytes\nconst lmntSpeechCallOptionsSchema = z.object({\n /**\n * The model to use for speech synthesis e.g. 'aurora' or 'blizzard'.\n * @default 'aurora'\n */\n model: z\n .union([z.enum(['aurora', 'blizzard']), z.string()])\n .nullish()\n .default('aurora'),\n\n /**\n * The audio format of the output.\n * @default 'mp3'\n */\n format: z\n .enum(['aac', 'mp3', 'mulaw', 'raw', 'wav'])\n .nullish()\n .default('mp3'),\n\n /**\n * The sample rate of the output audio in Hz.\n * @default 24000\n */\n sampleRate: z\n .union([z.literal(8000), z.literal(16000), z.literal(24000)])\n .nullish()\n .default(24000),\n\n /**\n * The speed of the speech. Range: 0.25 to 2.\n * @default 1\n */\n speed: z.number().min(0.25).max(2).nullish().default(1),\n\n /**\n * A seed value for deterministic generation.\n */\n seed: z.number().int().nullish(),\n\n /**\n * Whether to use a conversational style.\n * @default false\n */\n conversational: z.boolean().nullish().default(false),\n\n /**\n * Maximum length of the output in seconds (up to 300).\n */\n length: z.number().max(300).nullish(),\n\n /**\n * Top-p sampling parameter. Range: 0 to 1.\n * @default 1\n */\n topP: z.number().min(0).max(1).nullish().default(1),\n\n /**\n * Temperature for sampling. Higher values increase randomness.\n * @default 1\n */\n temperature: z.number().min(0).nullish().default(1),\n});\n\nexport type LMNTSpeechCallOptions = z.infer<typeof lmntSpeechCallOptionsSchema>;\n\ninterface LMNTSpeechModelConfig extends LMNTConfig {\n _internal?: {\n currentDate?: () => Date;\n };\n}\n\nexport class LMNTSpeechModel implements SpeechModelV3 {\n readonly specificationVersion = 'v3';\n\n get provider(): string {\n return this.config.provider;\n }\n\n constructor(\n readonly modelId: LMNTSpeechModelId,\n private readonly config: LMNTSpeechModelConfig,\n ) {}\n\n private async getArgs({\n text,\n voice = 'ava',\n outputFormat = 'mp3',\n speed,\n language,\n providerOptions,\n }: Parameters<SpeechModelV3['doGenerate']>[0]) {\n const warnings: SharedV3Warning[] = [];\n\n // Parse provider options\n const lmntOptions = await parseProviderOptions({\n provider: 'lmnt',\n providerOptions,\n schema: lmntSpeechCallOptionsSchema,\n });\n\n // Create request body\n const requestBody: Record<string, unknown> = {\n model: this.modelId,\n text,\n voice,\n response_format: 'mp3',\n speed,\n };\n\n if (outputFormat) {\n if (['mp3', 'aac', 'mulaw', 'raw', 'wav'].includes(outputFormat)) {\n requestBody.response_format = outputFormat;\n } else {\n warnings.push({\n type: 'unsupported',\n feature: 'outputFormat',\n details: `Unsupported output format: ${outputFormat}. Using mp3 instead.`,\n });\n }\n }\n\n // Add provider-specific options\n if (lmntOptions) {\n const speechModelOptions: Omit<LMNTSpeechAPITypes, 'voice' | 'text'> = {\n conversational: lmntOptions.conversational ?? undefined,\n length: lmntOptions.length ?? undefined,\n seed: lmntOptions.seed ?? undefined,\n speed: lmntOptions.speed ?? undefined,\n temperature: lmntOptions.temperature ?? undefined,\n top_p: lmntOptions.topP ?? undefined,\n sample_rate: lmntOptions.sampleRate ?? undefined,\n };\n\n for (const key in speechModelOptions) {\n const value =\n speechModelOptions[\n key as keyof Omit<LMNTSpeechAPITypes, 'voice' | 'text'>\n ];\n if (value !== undefined) {\n requestBody[key] = value;\n }\n }\n }\n\n if (language) {\n requestBody.language = language;\n }\n\n return {\n requestBody,\n warnings,\n };\n }\n\n async doGenerate(\n options: Parameters<SpeechModelV3['doGenerate']>[0],\n ): Promise<Awaited<ReturnType<SpeechModelV3['doGenerate']>>> {\n const currentDate = this.config._internal?.currentDate?.() ?? new Date();\n const { requestBody, warnings } = await this.getArgs(options);\n\n const {\n value: audio,\n responseHeaders,\n rawValue: rawResponse,\n } = await postJsonToApi({\n url: this.config.url({\n path: '/v1/ai/speech/bytes',\n modelId: this.modelId,\n }),\n headers: combineHeaders(this.config.headers(), options.headers),\n body: requestBody,\n failedResponseHandler: lmntFailedResponseHandler,\n successfulResponseHandler: createBinaryResponseHandler(),\n abortSignal: options.abortSignal,\n fetch: this.config.fetch,\n });\n\n return {\n audio,\n warnings,\n request: {\n body: JSON.stringify(requestBody),\n },\n response: {\n timestamp: currentDate,\n modelId: this.modelId,\n headers: responseHeaders,\n body: rawResponse,\n },\n };\n }\n}\n","import { z } from 'zod/v4';\nimport { createJsonErrorResponseHandler } from '@ai-sdk/provider-utils';\n\nexport const lmntErrorDataSchema = z.object({\n error: z.object({\n message: z.string(),\n code: z.number(),\n }),\n});\n\nexport type LMNTErrorData = z.infer<typeof lmntErrorDataSchema>;\n\nexport const lmntFailedResponseHandler = createJsonErrorResponseHandler({\n errorSchema: lmntErrorDataSchema,\n errorToMessage: data => data.error.message,\n});\n","// Version string of this package injected at build time.\ndeclare const __PACKAGE_VERSION__: string | undefined;\nexport const VERSION: string =\n typeof __PACKAGE_VERSION__ !== 'undefined'\n ? __PACKAGE_VERSION__\n : '0.0.0-test';\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACCA,IAAAA,yBAIO;;;ACJP,IAAAC,yBAKO;AACP,IAAAC,aAAkB;;;ACPlB,gBAAkB;AAClB,4BAA+C;AAExC,IAAM,sBAAsB,YAAE,OAAO;AAAA,EAC1C,OAAO,YAAE,OAAO;AAAA,IACd,SAAS,YAAE,OAAO;AAAA,IAClB,MAAM,YAAE,OAAO;AAAA,EACjB,CAAC;AACH,CAAC;AAIM,IAAM,gCAA4B,sDAA+B;AAAA,EACtE,aAAa;AAAA,EACb,gBAAgB,UAAQ,KAAK,MAAM;AACrC,CAAC;;;ADDD,IAAM,8BAA8B,aAAE,OAAO;AAAA;AAAA;AAAA;AAAA;AAAA,EAK3C,OAAO,aACJ,MAAM,CAAC,aAAE,KAAK,CAAC,UAAU,UAAU,CAAC,GAAG,aAAE,OAAO,CAAC,CAAC,EAClD,QAAQ,EACR,QAAQ,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMnB,QAAQ,aACL,KAAK,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,CAAC,EAC1C,QAAQ,EACR,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,YAAY,aACT,MAAM,CAAC,aAAE,QAAQ,GAAI,GAAG,aAAE,QAAQ,IAAK,GAAG,aAAE,QAAQ,IAAK,CAAC,CAAC,EAC3D,QAAQ,EACR,QAAQ,IAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,OAAO,aAAE,OAAO,EAAE,IAAI,IAAI,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA,EAKtD,MAAM,aAAE,OAAO,EAAE,IAAI,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAM/B,gBAAgB,aAAE,QAAQ,EAAE,QAAQ,EAAE,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA,EAKnD,QAAQ,aAAE,OAAO,EAAE,IAAI,GAAG,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMpC,MAAM,aAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA;AAAA,EAMlD,aAAa,aAAE,OAAO,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AACpD,CAAC;AAUM,IAAM,kBAAN,MAA+C;AAAA,EAOpD,YACW,SACQ,QACjB;AAFS;AACQ;AARnB,SAAS,uBAAuB;AAAA,EAS7B;AAAA,EAPH,IAAI,WAAmB;AACrB,WAAO,KAAK,OAAO;AAAA,EACrB;AAAA,EAOA,MAAc,QAAQ;AAAA,IACpB;AAAA,IACA,QAAQ;AAAA,IACR,eAAe;AAAA,IACf;AAAA,IACA;AAAA,IACA;AAAA,EACF,GAA+C;AAxGjD;AAyGI,UAAM,WAA8B,CAAC;AAGrC,UAAM,cAAc,UAAM,6CAAqB;AAAA,MAC7C,UAAU;AAAA,MACV;AAAA,MACA,QAAQ;AAAA,IACV,CAAC;AAGD,UAAM,cAAuC;AAAA,MAC3C,OAAO,KAAK;AAAA,MACZ;AAAA,MACA;AAAA,MACA,iBAAiB;AAAA,MACjB;AAAA,IACF;AAEA,QAAI,cAAc;AAChB,UAAI,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,EAAE,SAAS,YAAY,GAAG;AAChE,oBAAY,kBAAkB;AAAA,MAChC,OAAO;AACL,iBAAS,KAAK;AAAA,UACZ,MAAM;AAAA,UACN,SAAS;AAAA,UACT,SAAS,8BAA8B,YAAY;AAAA,QACrD,CAAC;AAAA,MACH;AAAA,IACF;AAGA,QAAI,aAAa;AACf,YAAM,qBAAiE;AAAA,QACrE,iBAAgB,iBAAY,mBAAZ,YAA8B;AAAA,QAC9C,SAAQ,iBAAY,WAAZ,YAAsB;AAAA,QAC9B,OAAM,iBAAY,SAAZ,YAAoB;AAAA,QAC1B,QAAO,iBAAY,UAAZ,YAAqB;AAAA,QAC5B,cAAa,iBAAY,gBAAZ,YAA2B;AAAA,QACxC,QAAO,iBAAY,SAAZ,YAAoB;AAAA,QAC3B,cAAa,iBAAY,eAAZ,YAA0B;AAAA,MACzC;AAEA,iBAAW,OAAO,oBAAoB;AACpC,cAAM,QACJ,mBACE,GACF;AACF,YAAI,UAAU,QAAW;AACvB,sBAAY,GAAG,IAAI;AAAA,QACrB;AAAA,MACF;AAAA,IACF;AAEA,QAAI,UAAU;AACZ,kBAAY,WAAW;AAAA,IACzB;AAEA,WAAO;AAAA,MACL;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,WACJ,SAC2D;AA1K/D;AA2KI,UAAM,eAAc,sBAAK,OAAO,cAAZ,mBAAuB,gBAAvB,4CAA0C,oBAAI,KAAK;AACvE,UAAM,EAAE,aAAa,SAAS,IAAI,MAAM,KAAK,QAAQ,OAAO;AAE5D,UAAM;AAAA,MACJ,OAAO;AAAA,MACP;AAAA,MACA,UAAU;AAAA,IACZ,IAAI,UAAM,sCAAc;AAAA,MACtB,KAAK,KAAK,OAAO,IAAI;AAAA,QACnB,MAAM;AAAA,QACN,SAAS,KAAK;AAAA,MAChB,CAAC;AAAA,MACD,aAAS,uCAAe,KAAK,OAAO,QAAQ,GAAG,QAAQ,OAAO;AAAA,MAC9D,MAAM;AAAA,MACN,uBAAuB;AAAA,MACvB,+BAA2B,oDAA4B;AAAA,MACvD,aAAa,QAAQ;AAAA,MACrB,OAAO,KAAK,OAAO;AAAA,IACrB,CAAC;AAED,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,SAAS;AAAA,QACP,MAAM,KAAK,UAAU,WAAW;AAAA,MAClC;AAAA,MACA,UAAU;AAAA,QACR,WAAW;AAAA,QACX,SAAS,KAAK;AAAA,QACd,SAAS;AAAA,QACT,MAAM;AAAA,MACR;AAAA,IACF;AAAA,EACF;AACF;;;AE3MO,IAAM,UACX,OACI,UACA;;;AHwCC,SAAS,WAAW,UAAgC,CAAC,GAAiB;AAC3E,QAAM,aAAa,UACjB;AAAA,IACE;AAAA,MACE,iBAAa,mCAAW;AAAA,QACtB,QAAQ,QAAQ;AAAA,QAChB,yBAAyB;AAAA,QACzB,aAAa;AAAA,MACf,CAAC;AAAA,MACD,GAAG,QAAQ;AAAA,IACb;AAAA,IACA,eAAe,OAAO;AAAA,EACxB;AAEF,QAAM,oBAAoB,CAAC,YACzB,IAAI,gBAAgB,SAAS;AAAA,IAC3B,UAAU;AAAA,IACV,KAAK,CAAC,EAAE,KAAK,MAAM,uBAAuB,IAAI;AAAA,IAC9C,SAAS;AAAA,IACT,OAAO,QAAQ;AAAA,EACjB,CAAC;AAEH,QAAM,WAAW,SAAU,SAA4B;AACrD,WAAO;AAAA,MACL,QAAQ,kBAAkB,OAAO;AAAA,IACnC;AAAA,EACF;AAEA,WAAS,SAAS;AAClB,WAAS,cAAc;AAEvB,SAAO;AACT;AAKO,IAAM,OAAO,WAAW;","names":["import_provider_utils","import_provider_utils","import_v4"]}
|
|
1
|
+
{"version":3,"sources":["../src/index.ts","../src/lmnt-provider.ts","../src/lmnt-speech-model.ts","../src/lmnt-error.ts","../src/version.ts"],"sourcesContent":["export { createLMNT, lmnt } from './lmnt-provider';\nexport type { LMNTProvider, LMNTProviderSettings } from './lmnt-provider';\nexport { VERSION } from './version';\n","import { SpeechModelV3, ProviderV3 } from '@ai-sdk/provider';\nimport {\n FetchFunction,\n loadApiKey,\n withUserAgentSuffix,\n} from '@ai-sdk/provider-utils';\nimport { LMNTSpeechModel } from './lmnt-speech-model';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { VERSION } from './version';\n\nexport interface LMNTProvider extends Pick<ProviderV3, 'speechModel'> {\n (\n modelId: 'aurora',\n settings?: {},\n ): {\n speech: LMNTSpeechModel;\n };\n\n /**\nCreates a model for speech synthesis.\n */\n speech(modelId: LMNTSpeechModelId): SpeechModelV3;\n}\n\nexport interface LMNTProviderSettings {\n /**\nAPI key for authenticating requests.\n */\n apiKey?: string;\n\n /**\nCustom headers to include in the requests.\n */\n headers?: Record<string, string>;\n\n /**\nCustom fetch implementation. You can use it as a middleware to intercept requests,\nor to provide a custom fetch implementation for e.g. testing.\n */\n fetch?: FetchFunction;\n}\n\n/**\nCreate an LMNT provider instance.\n */\nexport function createLMNT(options: LMNTProviderSettings = {}): LMNTProvider {\n const getHeaders = () =>\n withUserAgentSuffix(\n {\n 'x-api-key': loadApiKey({\n apiKey: options.apiKey,\n environmentVariableName: 'LMNT_API_KEY',\n description: 'LMNT',\n }),\n ...options.headers,\n },\n `ai-sdk/lmnt/${VERSION}`,\n );\n\n const createSpeechModel = (modelId: LMNTSpeechModelId) =>\n new LMNTSpeechModel(modelId, {\n provider: `lmnt.speech`,\n url: ({ path }) => `https://api.lmnt.com${path}`,\n headers: getHeaders,\n fetch: options.fetch,\n });\n\n const provider = function (modelId: LMNTSpeechModelId) {\n return {\n speech: createSpeechModel(modelId),\n };\n };\n\n provider.speech = createSpeechModel;\n provider.speechModel = createSpeechModel;\n\n return provider as LMNTProvider;\n}\n\n/**\nDefault LMNT provider instance.\n */\nexport const lmnt = createLMNT();\n","import { SpeechModelV3, SharedV3Warning } from '@ai-sdk/provider';\nimport {\n combineHeaders,\n createBinaryResponseHandler,\n parseProviderOptions,\n postJsonToApi,\n} from '@ai-sdk/provider-utils';\nimport { z } from 'zod/v4';\nimport { LMNTConfig } from './lmnt-config';\nimport { lmntFailedResponseHandler } from './lmnt-error';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { LMNTSpeechAPITypes } from './lmnt-api-types';\n\n// https://docs.lmnt.com/api-reference/speech/synthesize-speech-bytes\nconst lmntSpeechCallOptionsSchema = z.object({\n /**\n * The model to use for speech synthesis e.g. 'aurora' or 'blizzard'.\n * @default 'aurora'\n */\n model: z\n .union([z.enum(['aurora', 'blizzard']), z.string()])\n .nullish()\n .default('aurora'),\n\n /**\n * The audio format of the output.\n * @default 'mp3'\n */\n format: z\n .enum(['aac', 'mp3', 'mulaw', 'raw', 'wav'])\n .nullish()\n .default('mp3'),\n\n /**\n * The sample rate of the output audio in Hz.\n * @default 24000\n */\n sampleRate: z\n .union([z.literal(8000), z.literal(16000), z.literal(24000)])\n .nullish()\n .default(24000),\n\n /**\n * The speed of the speech. Range: 0.25 to 2.\n * @default 1\n */\n speed: z.number().min(0.25).max(2).nullish().default(1),\n\n /**\n * A seed value for deterministic generation.\n */\n seed: z.number().int().nullish(),\n\n /**\n * Whether to use a conversational style.\n * @default false\n */\n conversational: z.boolean().nullish().default(false),\n\n /**\n * Maximum length of the output in seconds (up to 300).\n */\n length: z.number().max(300).nullish(),\n\n /**\n * Top-p sampling parameter. Range: 0 to 1.\n * @default 1\n */\n topP: z.number().min(0).max(1).nullish().default(1),\n\n /**\n * Temperature for sampling. Higher values increase randomness.\n * @default 1\n */\n temperature: z.number().min(0).nullish().default(1),\n});\n\nexport type LMNTSpeechCallOptions = z.infer<typeof lmntSpeechCallOptionsSchema>;\n\ninterface LMNTSpeechModelConfig extends LMNTConfig {\n _internal?: {\n currentDate?: () => Date;\n };\n}\n\nexport class LMNTSpeechModel implements SpeechModelV3 {\n readonly specificationVersion = 'v3';\n\n get provider(): string {\n return this.config.provider;\n }\n\n constructor(\n readonly modelId: LMNTSpeechModelId,\n private readonly config: LMNTSpeechModelConfig,\n ) {}\n\n private async getArgs({\n text,\n voice = 'ava',\n outputFormat = 'mp3',\n speed,\n language,\n providerOptions,\n }: Parameters<SpeechModelV3['doGenerate']>[0]) {\n const warnings: SharedV3Warning[] = [];\n\n // Parse provider options\n const lmntOptions = await parseProviderOptions({\n provider: 'lmnt',\n providerOptions,\n schema: lmntSpeechCallOptionsSchema,\n });\n\n // Create request body\n const requestBody: Record<string, unknown> = {\n model: this.modelId,\n text,\n voice,\n response_format: 'mp3',\n speed,\n };\n\n if (outputFormat) {\n if (['mp3', 'aac', 'mulaw', 'raw', 'wav'].includes(outputFormat)) {\n requestBody.response_format = outputFormat;\n } else {\n warnings.push({\n type: 'unsupported',\n feature: 'outputFormat',\n details: `Unsupported output format: ${outputFormat}. Using mp3 instead.`,\n });\n }\n }\n\n // Add provider-specific options\n if (lmntOptions) {\n const speechModelOptions: Omit<LMNTSpeechAPITypes, 'voice' | 'text'> = {\n conversational: lmntOptions.conversational ?? undefined,\n length: lmntOptions.length ?? undefined,\n seed: lmntOptions.seed ?? undefined,\n speed: lmntOptions.speed ?? undefined,\n temperature: lmntOptions.temperature ?? undefined,\n top_p: lmntOptions.topP ?? undefined,\n sample_rate: lmntOptions.sampleRate ?? undefined,\n };\n\n for (const key in speechModelOptions) {\n const value =\n speechModelOptions[\n key as keyof Omit<LMNTSpeechAPITypes, 'voice' | 'text'>\n ];\n if (value !== undefined) {\n requestBody[key] = value;\n }\n }\n }\n\n if (language) {\n requestBody.language = language;\n }\n\n return {\n requestBody,\n warnings,\n };\n }\n\n async doGenerate(\n options: Parameters<SpeechModelV3['doGenerate']>[0],\n ): Promise<Awaited<ReturnType<SpeechModelV3['doGenerate']>>> {\n const currentDate = this.config._internal?.currentDate?.() ?? new Date();\n const { requestBody, warnings } = await this.getArgs(options);\n\n const {\n value: audio,\n responseHeaders,\n rawValue: rawResponse,\n } = await postJsonToApi({\n url: this.config.url({\n path: '/v1/ai/speech/bytes',\n modelId: this.modelId,\n }),\n headers: combineHeaders(this.config.headers(), options.headers),\n body: requestBody,\n failedResponseHandler: lmntFailedResponseHandler,\n successfulResponseHandler: createBinaryResponseHandler(),\n abortSignal: options.abortSignal,\n fetch: this.config.fetch,\n });\n\n return {\n audio,\n warnings,\n request: {\n body: JSON.stringify(requestBody),\n },\n response: {\n timestamp: currentDate,\n modelId: this.modelId,\n headers: responseHeaders,\n body: rawResponse,\n },\n };\n }\n}\n","import { z } from 'zod/v4';\nimport { createJsonErrorResponseHandler } from '@ai-sdk/provider-utils';\n\nexport const lmntErrorDataSchema = z.object({\n error: z.object({\n message: z.string(),\n code: z.number(),\n }),\n});\n\nexport type LMNTErrorData = z.infer<typeof lmntErrorDataSchema>;\n\nexport const lmntFailedResponseHandler = createJsonErrorResponseHandler({\n errorSchema: lmntErrorDataSchema,\n errorToMessage: data => data.error.message,\n});\n","// Version string of this package injected at build time.\ndeclare const __PACKAGE_VERSION__: string | undefined;\nexport const VERSION: string =\n typeof __PACKAGE_VERSION__ !== 'undefined'\n ? __PACKAGE_VERSION__\n : '0.0.0-test';\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACCA,IAAAA,yBAIO;;;ACJP,IAAAC,yBAKO;AACP,IAAAC,aAAkB;;;ACPlB,gBAAkB;AAClB,4BAA+C;AAExC,IAAM,sBAAsB,YAAE,OAAO;AAAA,EAC1C,OAAO,YAAE,OAAO;AAAA,IACd,SAAS,YAAE,OAAO;AAAA,IAClB,MAAM,YAAE,OAAO;AAAA,EACjB,CAAC;AACH,CAAC;AAIM,IAAM,gCAA4B,sDAA+B;AAAA,EACtE,aAAa;AAAA,EACb,gBAAgB,UAAQ,KAAK,MAAM;AACrC,CAAC;;;ADDD,IAAM,8BAA8B,aAAE,OAAO;AAAA;AAAA;AAAA;AAAA;AAAA,EAK3C,OAAO,aACJ,MAAM,CAAC,aAAE,KAAK,CAAC,UAAU,UAAU,CAAC,GAAG,aAAE,OAAO,CAAC,CAAC,EAClD,QAAQ,EACR,QAAQ,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMnB,QAAQ,aACL,KAAK,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,CAAC,EAC1C,QAAQ,EACR,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,YAAY,aACT,MAAM,CAAC,aAAE,QAAQ,GAAI,GAAG,aAAE,QAAQ,IAAK,GAAG,aAAE,QAAQ,IAAK,CAAC,CAAC,EAC3D,QAAQ,EACR,QAAQ,IAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,OAAO,aAAE,OAAO,EAAE,IAAI,IAAI,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA,EAKtD,MAAM,aAAE,OAAO,EAAE,IAAI,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAM/B,gBAAgB,aAAE,QAAQ,EAAE,QAAQ,EAAE,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA,EAKnD,QAAQ,aAAE,OAAO,EAAE,IAAI,GAAG,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMpC,MAAM,aAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA;AAAA,EAMlD,aAAa,aAAE,OAAO,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AACpD,CAAC;AAUM,IAAM,kBAAN,MAA+C;AAAA,EAOpD,YACW,SACQ,QACjB;AAFS;AACQ;AARnB,SAAS,uBAAuB;AAAA,EAS7B;AAAA,EAPH,IAAI,WAAmB;AACrB,WAAO,KAAK,OAAO;AAAA,EACrB;AAAA,EAOA,MAAc,QAAQ;AAAA,IACpB;AAAA,IACA,QAAQ;AAAA,IACR,eAAe;AAAA,IACf;AAAA,IACA;AAAA,IACA;AAAA,EACF,GAA+C;AAxGjD;AAyGI,UAAM,WAA8B,CAAC;AAGrC,UAAM,cAAc,UAAM,6CAAqB;AAAA,MAC7C,UAAU;AAAA,MACV;AAAA,MACA,QAAQ;AAAA,IACV,CAAC;AAGD,UAAM,cAAuC;AAAA,MAC3C,OAAO,KAAK;AAAA,MACZ;AAAA,MACA;AAAA,MACA,iBAAiB;AAAA,MACjB;AAAA,IACF;AAEA,QAAI,cAAc;AAChB,UAAI,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,EAAE,SAAS,YAAY,GAAG;AAChE,oBAAY,kBAAkB;AAAA,MAChC,OAAO;AACL,iBAAS,KAAK;AAAA,UACZ,MAAM;AAAA,UACN,SAAS;AAAA,UACT,SAAS,8BAA8B,YAAY;AAAA,QACrD,CAAC;AAAA,MACH;AAAA,IACF;AAGA,QAAI,aAAa;AACf,YAAM,qBAAiE;AAAA,QACrE,iBAAgB,iBAAY,mBAAZ,YAA8B;AAAA,QAC9C,SAAQ,iBAAY,WAAZ,YAAsB;AAAA,QAC9B,OAAM,iBAAY,SAAZ,YAAoB;AAAA,QAC1B,QAAO,iBAAY,UAAZ,YAAqB;AAAA,QAC5B,cAAa,iBAAY,gBAAZ,YAA2B;AAAA,QACxC,QAAO,iBAAY,SAAZ,YAAoB;AAAA,QAC3B,cAAa,iBAAY,eAAZ,YAA0B;AAAA,MACzC;AAEA,iBAAW,OAAO,oBAAoB;AACpC,cAAM,QACJ,mBACE,GACF;AACF,YAAI,UAAU,QAAW;AACvB,sBAAY,GAAG,IAAI;AAAA,QACrB;AAAA,MACF;AAAA,IACF;AAEA,QAAI,UAAU;AACZ,kBAAY,WAAW;AAAA,IACzB;AAEA,WAAO;AAAA,MACL;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,WACJ,SAC2D;AA1K/D;AA2KI,UAAM,eAAc,sBAAK,OAAO,cAAZ,mBAAuB,gBAAvB,4CAA0C,oBAAI,KAAK;AACvE,UAAM,EAAE,aAAa,SAAS,IAAI,MAAM,KAAK,QAAQ,OAAO;AAE5D,UAAM;AAAA,MACJ,OAAO;AAAA,MACP;AAAA,MACA,UAAU;AAAA,IACZ,IAAI,UAAM,sCAAc;AAAA,MACtB,KAAK,KAAK,OAAO,IAAI;AAAA,QACnB,MAAM;AAAA,QACN,SAAS,KAAK;AAAA,MAChB,CAAC;AAAA,MACD,aAAS,uCAAe,KAAK,OAAO,QAAQ,GAAG,QAAQ,OAAO;AAAA,MAC9D,MAAM;AAAA,MACN,uBAAuB;AAAA,MACvB,+BAA2B,oDAA4B;AAAA,MACvD,aAAa,QAAQ;AAAA,MACrB,OAAO,KAAK,OAAO;AAAA,IACrB,CAAC;AAED,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,SAAS;AAAA,QACP,MAAM,KAAK,UAAU,WAAW;AAAA,MAClC;AAAA,MACA,UAAU;AAAA,QACR,WAAW;AAAA,QACX,SAAS,KAAK;AAAA,QACd,SAAS;AAAA,QACT,MAAM;AAAA,MACR;AAAA,IACF;AAAA,EACF;AACF;;;AE3MO,IAAM,UACX,OACI,WACA;;;AHwCC,SAAS,WAAW,UAAgC,CAAC,GAAiB;AAC3E,QAAM,aAAa,UACjB;AAAA,IACE;AAAA,MACE,iBAAa,mCAAW;AAAA,QACtB,QAAQ,QAAQ;AAAA,QAChB,yBAAyB;AAAA,QACzB,aAAa;AAAA,MACf,CAAC;AAAA,MACD,GAAG,QAAQ;AAAA,IACb;AAAA,IACA,eAAe,OAAO;AAAA,EACxB;AAEF,QAAM,oBAAoB,CAAC,YACzB,IAAI,gBAAgB,SAAS;AAAA,IAC3B,UAAU;AAAA,IACV,KAAK,CAAC,EAAE,KAAK,MAAM,uBAAuB,IAAI;AAAA,IAC9C,SAAS;AAAA,IACT,OAAO,QAAQ;AAAA,EACjB,CAAC;AAEH,QAAM,WAAW,SAAU,SAA4B;AACrD,WAAO;AAAA,MACL,QAAQ,kBAAkB,OAAO;AAAA,IACnC;AAAA,EACF;AAEA,WAAS,SAAS;AAClB,WAAS,cAAc;AAEvB,SAAO;AACT;AAKO,IAAM,OAAO,WAAW;","names":["import_provider_utils","import_provider_utils","import_v4"]}
|
package/dist/index.mjs
CHANGED
package/dist/index.mjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/lmnt-provider.ts","../src/lmnt-speech-model.ts","../src/lmnt-error.ts","../src/version.ts"],"sourcesContent":["import { SpeechModelV3, ProviderV3 } from '@ai-sdk/provider';\nimport {\n FetchFunction,\n loadApiKey,\n withUserAgentSuffix,\n} from '@ai-sdk/provider-utils';\nimport { LMNTSpeechModel } from './lmnt-speech-model';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { VERSION } from './version';\n\nexport interface LMNTProvider extends Pick<ProviderV3, 'speechModel'> {\n (\n modelId: 'aurora',\n settings?: {},\n ): {\n speech: LMNTSpeechModel;\n };\n\n /**\nCreates a model for speech synthesis.\n */\n speech(modelId: LMNTSpeechModelId): SpeechModelV3;\n}\n\nexport interface LMNTProviderSettings {\n /**\nAPI key for authenticating requests.\n */\n apiKey?: string;\n\n /**\nCustom headers to include in the requests.\n */\n headers?: Record<string, string>;\n\n /**\nCustom fetch implementation. You can use it as a middleware to intercept requests,\nor to provide a custom fetch implementation for e.g. testing.\n */\n fetch?: FetchFunction;\n}\n\n/**\nCreate an LMNT provider instance.\n */\nexport function createLMNT(options: LMNTProviderSettings = {}): LMNTProvider {\n const getHeaders = () =>\n withUserAgentSuffix(\n {\n 'x-api-key': loadApiKey({\n apiKey: options.apiKey,\n environmentVariableName: 'LMNT_API_KEY',\n description: 'LMNT',\n }),\n ...options.headers,\n },\n `ai-sdk/lmnt/${VERSION}`,\n );\n\n const createSpeechModel = (modelId: LMNTSpeechModelId) =>\n new LMNTSpeechModel(modelId, {\n provider: `lmnt.speech`,\n url: ({ path }) => `https://api.lmnt.com${path}`,\n headers: getHeaders,\n fetch: options.fetch,\n });\n\n const provider = function (modelId: LMNTSpeechModelId) {\n return {\n speech: createSpeechModel(modelId),\n };\n };\n\n provider.speech = createSpeechModel;\n provider.speechModel = createSpeechModel;\n\n return provider as LMNTProvider;\n}\n\n/**\nDefault LMNT provider instance.\n */\nexport const lmnt = createLMNT();\n","import { SpeechModelV3, SharedV3Warning } from '@ai-sdk/provider';\nimport {\n combineHeaders,\n createBinaryResponseHandler,\n parseProviderOptions,\n postJsonToApi,\n} from '@ai-sdk/provider-utils';\nimport { z } from 'zod/v4';\nimport { LMNTConfig } from './lmnt-config';\nimport { lmntFailedResponseHandler } from './lmnt-error';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { LMNTSpeechAPITypes } from './lmnt-api-types';\n\n// https://docs.lmnt.com/api-reference/speech/synthesize-speech-bytes\nconst lmntSpeechCallOptionsSchema = z.object({\n /**\n * The model to use for speech synthesis e.g. 'aurora' or 'blizzard'.\n * @default 'aurora'\n */\n model: z\n .union([z.enum(['aurora', 'blizzard']), z.string()])\n .nullish()\n .default('aurora'),\n\n /**\n * The audio format of the output.\n * @default 'mp3'\n */\n format: z\n .enum(['aac', 'mp3', 'mulaw', 'raw', 'wav'])\n .nullish()\n .default('mp3'),\n\n /**\n * The sample rate of the output audio in Hz.\n * @default 24000\n */\n sampleRate: z\n .union([z.literal(8000), z.literal(16000), z.literal(24000)])\n .nullish()\n .default(24000),\n\n /**\n * The speed of the speech. Range: 0.25 to 2.\n * @default 1\n */\n speed: z.number().min(0.25).max(2).nullish().default(1),\n\n /**\n * A seed value for deterministic generation.\n */\n seed: z.number().int().nullish(),\n\n /**\n * Whether to use a conversational style.\n * @default false\n */\n conversational: z.boolean().nullish().default(false),\n\n /**\n * Maximum length of the output in seconds (up to 300).\n */\n length: z.number().max(300).nullish(),\n\n /**\n * Top-p sampling parameter. Range: 0 to 1.\n * @default 1\n */\n topP: z.number().min(0).max(1).nullish().default(1),\n\n /**\n * Temperature for sampling. Higher values increase randomness.\n * @default 1\n */\n temperature: z.number().min(0).nullish().default(1),\n});\n\nexport type LMNTSpeechCallOptions = z.infer<typeof lmntSpeechCallOptionsSchema>;\n\ninterface LMNTSpeechModelConfig extends LMNTConfig {\n _internal?: {\n currentDate?: () => Date;\n };\n}\n\nexport class LMNTSpeechModel implements SpeechModelV3 {\n readonly specificationVersion = 'v3';\n\n get provider(): string {\n return this.config.provider;\n }\n\n constructor(\n readonly modelId: LMNTSpeechModelId,\n private readonly config: LMNTSpeechModelConfig,\n ) {}\n\n private async getArgs({\n text,\n voice = 'ava',\n outputFormat = 'mp3',\n speed,\n language,\n providerOptions,\n }: Parameters<SpeechModelV3['doGenerate']>[0]) {\n const warnings: SharedV3Warning[] = [];\n\n // Parse provider options\n const lmntOptions = await parseProviderOptions({\n provider: 'lmnt',\n providerOptions,\n schema: lmntSpeechCallOptionsSchema,\n });\n\n // Create request body\n const requestBody: Record<string, unknown> = {\n model: this.modelId,\n text,\n voice,\n response_format: 'mp3',\n speed,\n };\n\n if (outputFormat) {\n if (['mp3', 'aac', 'mulaw', 'raw', 'wav'].includes(outputFormat)) {\n requestBody.response_format = outputFormat;\n } else {\n warnings.push({\n type: 'unsupported',\n feature: 'outputFormat',\n details: `Unsupported output format: ${outputFormat}. Using mp3 instead.`,\n });\n }\n }\n\n // Add provider-specific options\n if (lmntOptions) {\n const speechModelOptions: Omit<LMNTSpeechAPITypes, 'voice' | 'text'> = {\n conversational: lmntOptions.conversational ?? undefined,\n length: lmntOptions.length ?? undefined,\n seed: lmntOptions.seed ?? undefined,\n speed: lmntOptions.speed ?? undefined,\n temperature: lmntOptions.temperature ?? undefined,\n top_p: lmntOptions.topP ?? undefined,\n sample_rate: lmntOptions.sampleRate ?? undefined,\n };\n\n for (const key in speechModelOptions) {\n const value =\n speechModelOptions[\n key as keyof Omit<LMNTSpeechAPITypes, 'voice' | 'text'>\n ];\n if (value !== undefined) {\n requestBody[key] = value;\n }\n }\n }\n\n if (language) {\n requestBody.language = language;\n }\n\n return {\n requestBody,\n warnings,\n };\n }\n\n async doGenerate(\n options: Parameters<SpeechModelV3['doGenerate']>[0],\n ): Promise<Awaited<ReturnType<SpeechModelV3['doGenerate']>>> {\n const currentDate = this.config._internal?.currentDate?.() ?? new Date();\n const { requestBody, warnings } = await this.getArgs(options);\n\n const {\n value: audio,\n responseHeaders,\n rawValue: rawResponse,\n } = await postJsonToApi({\n url: this.config.url({\n path: '/v1/ai/speech/bytes',\n modelId: this.modelId,\n }),\n headers: combineHeaders(this.config.headers(), options.headers),\n body: requestBody,\n failedResponseHandler: lmntFailedResponseHandler,\n successfulResponseHandler: createBinaryResponseHandler(),\n abortSignal: options.abortSignal,\n fetch: this.config.fetch,\n });\n\n return {\n audio,\n warnings,\n request: {\n body: JSON.stringify(requestBody),\n },\n response: {\n timestamp: currentDate,\n modelId: this.modelId,\n headers: responseHeaders,\n body: rawResponse,\n },\n };\n }\n}\n","import { z } from 'zod/v4';\nimport { createJsonErrorResponseHandler } from '@ai-sdk/provider-utils';\n\nexport const lmntErrorDataSchema = z.object({\n error: z.object({\n message: z.string(),\n code: z.number(),\n }),\n});\n\nexport type LMNTErrorData = z.infer<typeof lmntErrorDataSchema>;\n\nexport const lmntFailedResponseHandler = createJsonErrorResponseHandler({\n errorSchema: lmntErrorDataSchema,\n errorToMessage: data => data.error.message,\n});\n","// Version string of this package injected at build time.\ndeclare const __PACKAGE_VERSION__: string | undefined;\nexport const VERSION: string =\n typeof __PACKAGE_VERSION__ !== 'undefined'\n ? __PACKAGE_VERSION__\n : '0.0.0-test';\n"],"mappings":";AACA;AAAA,EAEE;AAAA,EACA;AAAA,OACK;;;ACJP;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,OACK;AACP,SAAS,KAAAA,UAAS;;;ACPlB,SAAS,SAAS;AAClB,SAAS,sCAAsC;AAExC,IAAM,sBAAsB,EAAE,OAAO;AAAA,EAC1C,OAAO,EAAE,OAAO;AAAA,IACd,SAAS,EAAE,OAAO;AAAA,IAClB,MAAM,EAAE,OAAO;AAAA,EACjB,CAAC;AACH,CAAC;AAIM,IAAM,4BAA4B,+BAA+B;AAAA,EACtE,aAAa;AAAA,EACb,gBAAgB,UAAQ,KAAK,MAAM;AACrC,CAAC;;;ADDD,IAAM,8BAA8BC,GAAE,OAAO;AAAA;AAAA;AAAA;AAAA;AAAA,EAK3C,OAAOA,GACJ,MAAM,CAACA,GAAE,KAAK,CAAC,UAAU,UAAU,CAAC,GAAGA,GAAE,OAAO,CAAC,CAAC,EAClD,QAAQ,EACR,QAAQ,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMnB,QAAQA,GACL,KAAK,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,CAAC,EAC1C,QAAQ,EACR,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,YAAYA,GACT,MAAM,CAACA,GAAE,QAAQ,GAAI,GAAGA,GAAE,QAAQ,IAAK,GAAGA,GAAE,QAAQ,IAAK,CAAC,CAAC,EAC3D,QAAQ,EACR,QAAQ,IAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,OAAOA,GAAE,OAAO,EAAE,IAAI,IAAI,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA,EAKtD,MAAMA,GAAE,OAAO,EAAE,IAAI,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAM/B,gBAAgBA,GAAE,QAAQ,EAAE,QAAQ,EAAE,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA,EAKnD,QAAQA,GAAE,OAAO,EAAE,IAAI,GAAG,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMpC,MAAMA,GAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA;AAAA,EAMlD,aAAaA,GAAE,OAAO,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AACpD,CAAC;AAUM,IAAM,kBAAN,MAA+C;AAAA,EAOpD,YACW,SACQ,QACjB;AAFS;AACQ;AARnB,SAAS,uBAAuB;AAAA,EAS7B;AAAA,EAPH,IAAI,WAAmB;AACrB,WAAO,KAAK,OAAO;AAAA,EACrB;AAAA,EAOA,MAAc,QAAQ;AAAA,IACpB;AAAA,IACA,QAAQ;AAAA,IACR,eAAe;AAAA,IACf;AAAA,IACA;AAAA,IACA;AAAA,EACF,GAA+C;AAxGjD;AAyGI,UAAM,WAA8B,CAAC;AAGrC,UAAM,cAAc,MAAM,qBAAqB;AAAA,MAC7C,UAAU;AAAA,MACV;AAAA,MACA,QAAQ;AAAA,IACV,CAAC;AAGD,UAAM,cAAuC;AAAA,MAC3C,OAAO,KAAK;AAAA,MACZ;AAAA,MACA;AAAA,MACA,iBAAiB;AAAA,MACjB;AAAA,IACF;AAEA,QAAI,cAAc;AAChB,UAAI,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,EAAE,SAAS,YAAY,GAAG;AAChE,oBAAY,kBAAkB;AAAA,MAChC,OAAO;AACL,iBAAS,KAAK;AAAA,UACZ,MAAM;AAAA,UACN,SAAS;AAAA,UACT,SAAS,8BAA8B,YAAY;AAAA,QACrD,CAAC;AAAA,MACH;AAAA,IACF;AAGA,QAAI,aAAa;AACf,YAAM,qBAAiE;AAAA,QACrE,iBAAgB,iBAAY,mBAAZ,YAA8B;AAAA,QAC9C,SAAQ,iBAAY,WAAZ,YAAsB;AAAA,QAC9B,OAAM,iBAAY,SAAZ,YAAoB;AAAA,QAC1B,QAAO,iBAAY,UAAZ,YAAqB;AAAA,QAC5B,cAAa,iBAAY,gBAAZ,YAA2B;AAAA,QACxC,QAAO,iBAAY,SAAZ,YAAoB;AAAA,QAC3B,cAAa,iBAAY,eAAZ,YAA0B;AAAA,MACzC;AAEA,iBAAW,OAAO,oBAAoB;AACpC,cAAM,QACJ,mBACE,GACF;AACF,YAAI,UAAU,QAAW;AACvB,sBAAY,GAAG,IAAI;AAAA,QACrB;AAAA,MACF;AAAA,IACF;AAEA,QAAI,UAAU;AACZ,kBAAY,WAAW;AAAA,IACzB;AAEA,WAAO;AAAA,MACL;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,WACJ,SAC2D;AA1K/D;AA2KI,UAAM,eAAc,sBAAK,OAAO,cAAZ,mBAAuB,gBAAvB,4CAA0C,oBAAI,KAAK;AACvE,UAAM,EAAE,aAAa,SAAS,IAAI,MAAM,KAAK,QAAQ,OAAO;AAE5D,UAAM;AAAA,MACJ,OAAO;AAAA,MACP;AAAA,MACA,UAAU;AAAA,IACZ,IAAI,MAAM,cAAc;AAAA,MACtB,KAAK,KAAK,OAAO,IAAI;AAAA,QACnB,MAAM;AAAA,QACN,SAAS,KAAK;AAAA,MAChB,CAAC;AAAA,MACD,SAAS,eAAe,KAAK,OAAO,QAAQ,GAAG,QAAQ,OAAO;AAAA,MAC9D,MAAM;AAAA,MACN,uBAAuB;AAAA,MACvB,2BAA2B,4BAA4B;AAAA,MACvD,aAAa,QAAQ;AAAA,MACrB,OAAO,KAAK,OAAO;AAAA,IACrB,CAAC;AAED,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,SAAS;AAAA,QACP,MAAM,KAAK,UAAU,WAAW;AAAA,MAClC;AAAA,MACA,UAAU;AAAA,QACR,WAAW;AAAA,QACX,SAAS,KAAK;AAAA,QACd,SAAS;AAAA,QACT,MAAM;AAAA,MACR;AAAA,IACF;AAAA,EACF;AACF;;;AE3MO,IAAM,UACX,OACI,UACA;;;AHwCC,SAAS,WAAW,UAAgC,CAAC,GAAiB;AAC3E,QAAM,aAAa,MACjB;AAAA,IACE;AAAA,MACE,aAAa,WAAW;AAAA,QACtB,QAAQ,QAAQ;AAAA,QAChB,yBAAyB;AAAA,QACzB,aAAa;AAAA,MACf,CAAC;AAAA,MACD,GAAG,QAAQ;AAAA,IACb;AAAA,IACA,eAAe,OAAO;AAAA,EACxB;AAEF,QAAM,oBAAoB,CAAC,YACzB,IAAI,gBAAgB,SAAS;AAAA,IAC3B,UAAU;AAAA,IACV,KAAK,CAAC,EAAE,KAAK,MAAM,uBAAuB,IAAI;AAAA,IAC9C,SAAS;AAAA,IACT,OAAO,QAAQ;AAAA,EACjB,CAAC;AAEH,QAAM,WAAW,SAAU,SAA4B;AACrD,WAAO;AAAA,MACL,QAAQ,kBAAkB,OAAO;AAAA,IACnC;AAAA,EACF;AAEA,WAAS,SAAS;AAClB,WAAS,cAAc;AAEvB,SAAO;AACT;AAKO,IAAM,OAAO,WAAW;","names":["z","z"]}
|
|
1
|
+
{"version":3,"sources":["../src/lmnt-provider.ts","../src/lmnt-speech-model.ts","../src/lmnt-error.ts","../src/version.ts"],"sourcesContent":["import { SpeechModelV3, ProviderV3 } from '@ai-sdk/provider';\nimport {\n FetchFunction,\n loadApiKey,\n withUserAgentSuffix,\n} from '@ai-sdk/provider-utils';\nimport { LMNTSpeechModel } from './lmnt-speech-model';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { VERSION } from './version';\n\nexport interface LMNTProvider extends Pick<ProviderV3, 'speechModel'> {\n (\n modelId: 'aurora',\n settings?: {},\n ): {\n speech: LMNTSpeechModel;\n };\n\n /**\nCreates a model for speech synthesis.\n */\n speech(modelId: LMNTSpeechModelId): SpeechModelV3;\n}\n\nexport interface LMNTProviderSettings {\n /**\nAPI key for authenticating requests.\n */\n apiKey?: string;\n\n /**\nCustom headers to include in the requests.\n */\n headers?: Record<string, string>;\n\n /**\nCustom fetch implementation. You can use it as a middleware to intercept requests,\nor to provide a custom fetch implementation for e.g. testing.\n */\n fetch?: FetchFunction;\n}\n\n/**\nCreate an LMNT provider instance.\n */\nexport function createLMNT(options: LMNTProviderSettings = {}): LMNTProvider {\n const getHeaders = () =>\n withUserAgentSuffix(\n {\n 'x-api-key': loadApiKey({\n apiKey: options.apiKey,\n environmentVariableName: 'LMNT_API_KEY',\n description: 'LMNT',\n }),\n ...options.headers,\n },\n `ai-sdk/lmnt/${VERSION}`,\n );\n\n const createSpeechModel = (modelId: LMNTSpeechModelId) =>\n new LMNTSpeechModel(modelId, {\n provider: `lmnt.speech`,\n url: ({ path }) => `https://api.lmnt.com${path}`,\n headers: getHeaders,\n fetch: options.fetch,\n });\n\n const provider = function (modelId: LMNTSpeechModelId) {\n return {\n speech: createSpeechModel(modelId),\n };\n };\n\n provider.speech = createSpeechModel;\n provider.speechModel = createSpeechModel;\n\n return provider as LMNTProvider;\n}\n\n/**\nDefault LMNT provider instance.\n */\nexport const lmnt = createLMNT();\n","import { SpeechModelV3, SharedV3Warning } from '@ai-sdk/provider';\nimport {\n combineHeaders,\n createBinaryResponseHandler,\n parseProviderOptions,\n postJsonToApi,\n} from '@ai-sdk/provider-utils';\nimport { z } from 'zod/v4';\nimport { LMNTConfig } from './lmnt-config';\nimport { lmntFailedResponseHandler } from './lmnt-error';\nimport { LMNTSpeechModelId } from './lmnt-speech-options';\nimport { LMNTSpeechAPITypes } from './lmnt-api-types';\n\n// https://docs.lmnt.com/api-reference/speech/synthesize-speech-bytes\nconst lmntSpeechCallOptionsSchema = z.object({\n /**\n * The model to use for speech synthesis e.g. 'aurora' or 'blizzard'.\n * @default 'aurora'\n */\n model: z\n .union([z.enum(['aurora', 'blizzard']), z.string()])\n .nullish()\n .default('aurora'),\n\n /**\n * The audio format of the output.\n * @default 'mp3'\n */\n format: z\n .enum(['aac', 'mp3', 'mulaw', 'raw', 'wav'])\n .nullish()\n .default('mp3'),\n\n /**\n * The sample rate of the output audio in Hz.\n * @default 24000\n */\n sampleRate: z\n .union([z.literal(8000), z.literal(16000), z.literal(24000)])\n .nullish()\n .default(24000),\n\n /**\n * The speed of the speech. Range: 0.25 to 2.\n * @default 1\n */\n speed: z.number().min(0.25).max(2).nullish().default(1),\n\n /**\n * A seed value for deterministic generation.\n */\n seed: z.number().int().nullish(),\n\n /**\n * Whether to use a conversational style.\n * @default false\n */\n conversational: z.boolean().nullish().default(false),\n\n /**\n * Maximum length of the output in seconds (up to 300).\n */\n length: z.number().max(300).nullish(),\n\n /**\n * Top-p sampling parameter. Range: 0 to 1.\n * @default 1\n */\n topP: z.number().min(0).max(1).nullish().default(1),\n\n /**\n * Temperature for sampling. Higher values increase randomness.\n * @default 1\n */\n temperature: z.number().min(0).nullish().default(1),\n});\n\nexport type LMNTSpeechCallOptions = z.infer<typeof lmntSpeechCallOptionsSchema>;\n\ninterface LMNTSpeechModelConfig extends LMNTConfig {\n _internal?: {\n currentDate?: () => Date;\n };\n}\n\nexport class LMNTSpeechModel implements SpeechModelV3 {\n readonly specificationVersion = 'v3';\n\n get provider(): string {\n return this.config.provider;\n }\n\n constructor(\n readonly modelId: LMNTSpeechModelId,\n private readonly config: LMNTSpeechModelConfig,\n ) {}\n\n private async getArgs({\n text,\n voice = 'ava',\n outputFormat = 'mp3',\n speed,\n language,\n providerOptions,\n }: Parameters<SpeechModelV3['doGenerate']>[0]) {\n const warnings: SharedV3Warning[] = [];\n\n // Parse provider options\n const lmntOptions = await parseProviderOptions({\n provider: 'lmnt',\n providerOptions,\n schema: lmntSpeechCallOptionsSchema,\n });\n\n // Create request body\n const requestBody: Record<string, unknown> = {\n model: this.modelId,\n text,\n voice,\n response_format: 'mp3',\n speed,\n };\n\n if (outputFormat) {\n if (['mp3', 'aac', 'mulaw', 'raw', 'wav'].includes(outputFormat)) {\n requestBody.response_format = outputFormat;\n } else {\n warnings.push({\n type: 'unsupported',\n feature: 'outputFormat',\n details: `Unsupported output format: ${outputFormat}. Using mp3 instead.`,\n });\n }\n }\n\n // Add provider-specific options\n if (lmntOptions) {\n const speechModelOptions: Omit<LMNTSpeechAPITypes, 'voice' | 'text'> = {\n conversational: lmntOptions.conversational ?? undefined,\n length: lmntOptions.length ?? undefined,\n seed: lmntOptions.seed ?? undefined,\n speed: lmntOptions.speed ?? undefined,\n temperature: lmntOptions.temperature ?? undefined,\n top_p: lmntOptions.topP ?? undefined,\n sample_rate: lmntOptions.sampleRate ?? undefined,\n };\n\n for (const key in speechModelOptions) {\n const value =\n speechModelOptions[\n key as keyof Omit<LMNTSpeechAPITypes, 'voice' | 'text'>\n ];\n if (value !== undefined) {\n requestBody[key] = value;\n }\n }\n }\n\n if (language) {\n requestBody.language = language;\n }\n\n return {\n requestBody,\n warnings,\n };\n }\n\n async doGenerate(\n options: Parameters<SpeechModelV3['doGenerate']>[0],\n ): Promise<Awaited<ReturnType<SpeechModelV3['doGenerate']>>> {\n const currentDate = this.config._internal?.currentDate?.() ?? new Date();\n const { requestBody, warnings } = await this.getArgs(options);\n\n const {\n value: audio,\n responseHeaders,\n rawValue: rawResponse,\n } = await postJsonToApi({\n url: this.config.url({\n path: '/v1/ai/speech/bytes',\n modelId: this.modelId,\n }),\n headers: combineHeaders(this.config.headers(), options.headers),\n body: requestBody,\n failedResponseHandler: lmntFailedResponseHandler,\n successfulResponseHandler: createBinaryResponseHandler(),\n abortSignal: options.abortSignal,\n fetch: this.config.fetch,\n });\n\n return {\n audio,\n warnings,\n request: {\n body: JSON.stringify(requestBody),\n },\n response: {\n timestamp: currentDate,\n modelId: this.modelId,\n headers: responseHeaders,\n body: rawResponse,\n },\n };\n }\n}\n","import { z } from 'zod/v4';\nimport { createJsonErrorResponseHandler } from '@ai-sdk/provider-utils';\n\nexport const lmntErrorDataSchema = z.object({\n error: z.object({\n message: z.string(),\n code: z.number(),\n }),\n});\n\nexport type LMNTErrorData = z.infer<typeof lmntErrorDataSchema>;\n\nexport const lmntFailedResponseHandler = createJsonErrorResponseHandler({\n errorSchema: lmntErrorDataSchema,\n errorToMessage: data => data.error.message,\n});\n","// Version string of this package injected at build time.\ndeclare const __PACKAGE_VERSION__: string | undefined;\nexport const VERSION: string =\n typeof __PACKAGE_VERSION__ !== 'undefined'\n ? __PACKAGE_VERSION__\n : '0.0.0-test';\n"],"mappings":";AACA;AAAA,EAEE;AAAA,EACA;AAAA,OACK;;;ACJP;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,OACK;AACP,SAAS,KAAAA,UAAS;;;ACPlB,SAAS,SAAS;AAClB,SAAS,sCAAsC;AAExC,IAAM,sBAAsB,EAAE,OAAO;AAAA,EAC1C,OAAO,EAAE,OAAO;AAAA,IACd,SAAS,EAAE,OAAO;AAAA,IAClB,MAAM,EAAE,OAAO;AAAA,EACjB,CAAC;AACH,CAAC;AAIM,IAAM,4BAA4B,+BAA+B;AAAA,EACtE,aAAa;AAAA,EACb,gBAAgB,UAAQ,KAAK,MAAM;AACrC,CAAC;;;ADDD,IAAM,8BAA8BC,GAAE,OAAO;AAAA;AAAA;AAAA;AAAA;AAAA,EAK3C,OAAOA,GACJ,MAAM,CAACA,GAAE,KAAK,CAAC,UAAU,UAAU,CAAC,GAAGA,GAAE,OAAO,CAAC,CAAC,EAClD,QAAQ,EACR,QAAQ,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMnB,QAAQA,GACL,KAAK,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,CAAC,EAC1C,QAAQ,EACR,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,YAAYA,GACT,MAAM,CAACA,GAAE,QAAQ,GAAI,GAAGA,GAAE,QAAQ,IAAK,GAAGA,GAAE,QAAQ,IAAK,CAAC,CAAC,EAC3D,QAAQ,EACR,QAAQ,IAAK;AAAA;AAAA;AAAA;AAAA;AAAA,EAMhB,OAAOA,GAAE,OAAO,EAAE,IAAI,IAAI,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA,EAKtD,MAAMA,GAAE,OAAO,EAAE,IAAI,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAM/B,gBAAgBA,GAAE,QAAQ,EAAE,QAAQ,EAAE,QAAQ,KAAK;AAAA;AAAA;AAAA;AAAA,EAKnD,QAAQA,GAAE,OAAO,EAAE,IAAI,GAAG,EAAE,QAAQ;AAAA;AAAA;AAAA;AAAA;AAAA,EAMpC,MAAMA,GAAE,OAAO,EAAE,IAAI,CAAC,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AAAA;AAAA;AAAA;AAAA;AAAA,EAMlD,aAAaA,GAAE,OAAO,EAAE,IAAI,CAAC,EAAE,QAAQ,EAAE,QAAQ,CAAC;AACpD,CAAC;AAUM,IAAM,kBAAN,MAA+C;AAAA,EAOpD,YACW,SACQ,QACjB;AAFS;AACQ;AARnB,SAAS,uBAAuB;AAAA,EAS7B;AAAA,EAPH,IAAI,WAAmB;AACrB,WAAO,KAAK,OAAO;AAAA,EACrB;AAAA,EAOA,MAAc,QAAQ;AAAA,IACpB;AAAA,IACA,QAAQ;AAAA,IACR,eAAe;AAAA,IACf;AAAA,IACA;AAAA,IACA;AAAA,EACF,GAA+C;AAxGjD;AAyGI,UAAM,WAA8B,CAAC;AAGrC,UAAM,cAAc,MAAM,qBAAqB;AAAA,MAC7C,UAAU;AAAA,MACV;AAAA,MACA,QAAQ;AAAA,IACV,CAAC;AAGD,UAAM,cAAuC;AAAA,MAC3C,OAAO,KAAK;AAAA,MACZ;AAAA,MACA;AAAA,MACA,iBAAiB;AAAA,MACjB;AAAA,IACF;AAEA,QAAI,cAAc;AAChB,UAAI,CAAC,OAAO,OAAO,SAAS,OAAO,KAAK,EAAE,SAAS,YAAY,GAAG;AAChE,oBAAY,kBAAkB;AAAA,MAChC,OAAO;AACL,iBAAS,KAAK;AAAA,UACZ,MAAM;AAAA,UACN,SAAS;AAAA,UACT,SAAS,8BAA8B,YAAY;AAAA,QACrD,CAAC;AAAA,MACH;AAAA,IACF;AAGA,QAAI,aAAa;AACf,YAAM,qBAAiE;AAAA,QACrE,iBAAgB,iBAAY,mBAAZ,YAA8B;AAAA,QAC9C,SAAQ,iBAAY,WAAZ,YAAsB;AAAA,QAC9B,OAAM,iBAAY,SAAZ,YAAoB;AAAA,QAC1B,QAAO,iBAAY,UAAZ,YAAqB;AAAA,QAC5B,cAAa,iBAAY,gBAAZ,YAA2B;AAAA,QACxC,QAAO,iBAAY,SAAZ,YAAoB;AAAA,QAC3B,cAAa,iBAAY,eAAZ,YAA0B;AAAA,MACzC;AAEA,iBAAW,OAAO,oBAAoB;AACpC,cAAM,QACJ,mBACE,GACF;AACF,YAAI,UAAU,QAAW;AACvB,sBAAY,GAAG,IAAI;AAAA,QACrB;AAAA,MACF;AAAA,IACF;AAEA,QAAI,UAAU;AACZ,kBAAY,WAAW;AAAA,IACzB;AAEA,WAAO;AAAA,MACL;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,WACJ,SAC2D;AA1K/D;AA2KI,UAAM,eAAc,sBAAK,OAAO,cAAZ,mBAAuB,gBAAvB,4CAA0C,oBAAI,KAAK;AACvE,UAAM,EAAE,aAAa,SAAS,IAAI,MAAM,KAAK,QAAQ,OAAO;AAE5D,UAAM;AAAA,MACJ,OAAO;AAAA,MACP;AAAA,MACA,UAAU;AAAA,IACZ,IAAI,MAAM,cAAc;AAAA,MACtB,KAAK,KAAK,OAAO,IAAI;AAAA,QACnB,MAAM;AAAA,QACN,SAAS,KAAK;AAAA,MAChB,CAAC;AAAA,MACD,SAAS,eAAe,KAAK,OAAO,QAAQ,GAAG,QAAQ,OAAO;AAAA,MAC9D,MAAM;AAAA,MACN,uBAAuB;AAAA,MACvB,2BAA2B,4BAA4B;AAAA,MACvD,aAAa,QAAQ;AAAA,MACrB,OAAO,KAAK,OAAO;AAAA,IACrB,CAAC;AAED,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA,SAAS;AAAA,QACP,MAAM,KAAK,UAAU,WAAW;AAAA,MAClC;AAAA,MACA,UAAU;AAAA,QACR,WAAW;AAAA,QACX,SAAS,KAAK;AAAA,QACd,SAAS;AAAA,QACT,MAAM;AAAA,MACR;AAAA,IACF;AAAA,EACF;AACF;;;AE3MO,IAAM,UACX,OACI,WACA;;;AHwCC,SAAS,WAAW,UAAgC,CAAC,GAAiB;AAC3E,QAAM,aAAa,MACjB;AAAA,IACE;AAAA,MACE,aAAa,WAAW;AAAA,QACtB,QAAQ,QAAQ;AAAA,QAChB,yBAAyB;AAAA,QACzB,aAAa;AAAA,MACf,CAAC;AAAA,MACD,GAAG,QAAQ;AAAA,IACb;AAAA,IACA,eAAe,OAAO;AAAA,EACxB;AAEF,QAAM,oBAAoB,CAAC,YACzB,IAAI,gBAAgB,SAAS;AAAA,IAC3B,UAAU;AAAA,IACV,KAAK,CAAC,EAAE,KAAK,MAAM,uBAAuB,IAAI;AAAA,IAC9C,SAAS;AAAA,IACT,OAAO,QAAQ;AAAA,EACjB,CAAC;AAEH,QAAM,WAAW,SAAU,SAA4B;AACrD,WAAO;AAAA,MACL,QAAQ,kBAAkB,OAAO;AAAA,IACnC;AAAA,EACF;AAEA,WAAS,SAAS;AAClB,WAAS,cAAc;AAEvB,SAAO;AACT;AAKO,IAAM,OAAO,WAAW;","names":["z","z"]}
|
|
@@ -0,0 +1,140 @@
|
|
|
1
|
+
---
|
|
2
|
+
title: LMNT
|
|
3
|
+
description: Learn how to use the LMNT provider for the AI SDK.
|
|
4
|
+
---
|
|
5
|
+
|
|
6
|
+
# LMNT Provider
|
|
7
|
+
|
|
8
|
+
The [LMNT](https://lmnt.com/) provider contains language model support for the LMNT transcription API.
|
|
9
|
+
|
|
10
|
+
## Setup
|
|
11
|
+
|
|
12
|
+
The LMNT provider is available in the `@ai-sdk/lmnt` module. You can install it with
|
|
13
|
+
|
|
14
|
+
<Tabs items={['pnpm', 'npm', 'yarn', 'bun']}>
|
|
15
|
+
<Tab>
|
|
16
|
+
<Snippet text="pnpm add @ai-sdk/lmnt" dark />
|
|
17
|
+
</Tab>
|
|
18
|
+
<Tab>
|
|
19
|
+
<Snippet text="npm install @ai-sdk/lmnt" dark />
|
|
20
|
+
</Tab>
|
|
21
|
+
<Tab>
|
|
22
|
+
<Snippet text="yarn add @ai-sdk/lmnt" dark />
|
|
23
|
+
</Tab>
|
|
24
|
+
|
|
25
|
+
<Tab>
|
|
26
|
+
<Snippet text="bun add @ai-sdk/lmnt" dark />
|
|
27
|
+
</Tab>
|
|
28
|
+
</Tabs>
|
|
29
|
+
|
|
30
|
+
## Provider Instance
|
|
31
|
+
|
|
32
|
+
You can import the default provider instance `lmnt` from `@ai-sdk/lmnt`:
|
|
33
|
+
|
|
34
|
+
```ts
|
|
35
|
+
import { lmnt } from '@ai-sdk/lmnt';
|
|
36
|
+
```
|
|
37
|
+
|
|
38
|
+
If you need a customized setup, you can import `createLMNT` from `@ai-sdk/lmnt` and create a provider instance with your settings:
|
|
39
|
+
|
|
40
|
+
```ts
|
|
41
|
+
import { createLMNT } from '@ai-sdk/lmnt';
|
|
42
|
+
|
|
43
|
+
const lmnt = createLMNT({
|
|
44
|
+
// custom settings, e.g.
|
|
45
|
+
fetch: customFetch,
|
|
46
|
+
});
|
|
47
|
+
```
|
|
48
|
+
|
|
49
|
+
You can use the following optional settings to customize the LMNT provider instance:
|
|
50
|
+
|
|
51
|
+
- **apiKey** _string_
|
|
52
|
+
|
|
53
|
+
API key that is being sent using the `Authorization` header.
|
|
54
|
+
It defaults to the `LMNT_API_KEY` environment variable.
|
|
55
|
+
|
|
56
|
+
- **headers** _Record<string,string>_
|
|
57
|
+
|
|
58
|
+
Custom headers to include in the requests.
|
|
59
|
+
|
|
60
|
+
- **fetch** _(input: RequestInfo, init?: RequestInit) => Promise<Response>_
|
|
61
|
+
|
|
62
|
+
Custom [fetch](https://developer.mozilla.org/en-US/docs/Web/API/fetch) implementation.
|
|
63
|
+
Defaults to the global `fetch` function.
|
|
64
|
+
You can use it as a middleware to intercept requests,
|
|
65
|
+
or to provide a custom fetch implementation for e.g. testing.
|
|
66
|
+
|
|
67
|
+
## Speech Models
|
|
68
|
+
|
|
69
|
+
You can create models that call the [LMNT speech API](https://docs.lmnt.com/api-reference/speech/synthesize-speech-bytes)
|
|
70
|
+
using the `.speech()` factory method.
|
|
71
|
+
|
|
72
|
+
The first argument is the model id e.g. `aurora`.
|
|
73
|
+
|
|
74
|
+
```ts
|
|
75
|
+
const model = lmnt.speech('aurora');
|
|
76
|
+
```
|
|
77
|
+
|
|
78
|
+
You can also pass additional provider-specific options using the `providerOptions` argument. For example, supplying a voice to use for the generated audio.
|
|
79
|
+
|
|
80
|
+
```ts highlight="6"
|
|
81
|
+
import { experimental_generateSpeech as generateSpeech } from 'ai';
|
|
82
|
+
import { lmnt } from '@ai-sdk/lmnt';
|
|
83
|
+
|
|
84
|
+
const result = await generateSpeech({
|
|
85
|
+
model: lmnt.speech('aurora'),
|
|
86
|
+
text: 'Hello, world!',
|
|
87
|
+
language: 'en', // Standardized language parameter
|
|
88
|
+
});
|
|
89
|
+
```
|
|
90
|
+
|
|
91
|
+
### Provider Options
|
|
92
|
+
|
|
93
|
+
The LMNT provider accepts the following options:
|
|
94
|
+
|
|
95
|
+
- **model** _'aurora' | 'blizzard'_
|
|
96
|
+
|
|
97
|
+
The LMNT model to use. Defaults to `'aurora'`.
|
|
98
|
+
|
|
99
|
+
- **language** _'auto' | 'en' | 'es' | 'pt' | 'fr' | 'de' | 'zh' | 'ko' | 'hi' | 'ja' | 'ru' | 'it' | 'tr'_
|
|
100
|
+
|
|
101
|
+
The language to use for speech synthesis. Defaults to `'auto'`.
|
|
102
|
+
|
|
103
|
+
- **format** _'aac' | 'mp3' | 'mulaw' | 'raw' | 'wav'_
|
|
104
|
+
|
|
105
|
+
The audio format to return. Defaults to `'mp3'`.
|
|
106
|
+
|
|
107
|
+
- **sampleRate** _number_
|
|
108
|
+
|
|
109
|
+
The sample rate of the audio in Hz. Defaults to `24000`.
|
|
110
|
+
|
|
111
|
+
- **speed** _number_
|
|
112
|
+
|
|
113
|
+
The speed of the speech. Must be between 0.25 and 2. Defaults to `1`.
|
|
114
|
+
|
|
115
|
+
- **seed** _number_
|
|
116
|
+
|
|
117
|
+
An optional seed for deterministic generation.
|
|
118
|
+
|
|
119
|
+
- **conversational** _boolean_
|
|
120
|
+
|
|
121
|
+
Whether to use a conversational style. Defaults to `false`.
|
|
122
|
+
|
|
123
|
+
- **length** _number_
|
|
124
|
+
|
|
125
|
+
Maximum length of the audio in seconds. Maximum value is 300.
|
|
126
|
+
|
|
127
|
+
- **topP** _number_
|
|
128
|
+
|
|
129
|
+
Top-p sampling parameter. Must be between 0 and 1. Defaults to `1`.
|
|
130
|
+
|
|
131
|
+
- **temperature** _number_
|
|
132
|
+
|
|
133
|
+
Temperature parameter for sampling. Must be at least 0. Defaults to `1`.
|
|
134
|
+
|
|
135
|
+
### Model Capabilities
|
|
136
|
+
|
|
137
|
+
| Model | Instructions |
|
|
138
|
+
| ---------- | ------------------- |
|
|
139
|
+
| `aurora` | <Check size={18} /> |
|
|
140
|
+
| `blizzard` | <Check size={18} /> |
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@ai-sdk/lmnt",
|
|
3
|
-
"version": "2.0.
|
|
3
|
+
"version": "2.0.10",
|
|
4
4
|
"license": "Apache-2.0",
|
|
5
5
|
"sideEffects": false,
|
|
6
6
|
"main": "./dist/index.js",
|
|
@@ -8,9 +8,14 @@
|
|
|
8
8
|
"types": "./dist/index.d.ts",
|
|
9
9
|
"files": [
|
|
10
10
|
"dist/**/*",
|
|
11
|
+
"docs/**/*",
|
|
12
|
+
"src",
|
|
11
13
|
"CHANGELOG.md",
|
|
12
14
|
"README.md"
|
|
13
15
|
],
|
|
16
|
+
"directories": {
|
|
17
|
+
"doc": "./docs"
|
|
18
|
+
},
|
|
14
19
|
"exports": {
|
|
15
20
|
"./package.json": "./package.json",
|
|
16
21
|
".": {
|
|
@@ -28,7 +33,7 @@
|
|
|
28
33
|
"tsup": "^8",
|
|
29
34
|
"typescript": "5.6.3",
|
|
30
35
|
"zod": "3.25.76",
|
|
31
|
-
"@ai-sdk/test-server": "1.0.
|
|
36
|
+
"@ai-sdk/test-server": "1.0.2",
|
|
32
37
|
"@vercel/ai-tsconfig": "0.0.0"
|
|
33
38
|
},
|
|
34
39
|
"peerDependencies": {
|
|
@@ -54,7 +59,7 @@
|
|
|
54
59
|
"scripts": {
|
|
55
60
|
"build": "tsup --tsconfig tsconfig.build.json",
|
|
56
61
|
"build:watch": "tsup --tsconfig tsconfig.build.json --watch",
|
|
57
|
-
"clean": "del-cli dist",
|
|
62
|
+
"clean": "del-cli dist docs",
|
|
58
63
|
"lint": "eslint \"./**/*.ts*\"",
|
|
59
64
|
"type-check": "tsc --noEmit",
|
|
60
65
|
"prettier-check": "prettier --check \"./**/*.ts*\"",
|
package/src/index.ts
ADDED
|
@@ -0,0 +1,39 @@
|
|
|
1
|
+
export type LMNTSpeechAPITypes = {
|
|
2
|
+
/** The voice id of the voice to use; voice ids can be retrieved by calls to List voices or Voice info. */
|
|
3
|
+
voice: string;
|
|
4
|
+
/** The text to synthesize; max 5000 characters per request (including spaces) */
|
|
5
|
+
text: string;
|
|
6
|
+
/** The model to use for synthesis. One of aurora (default) or blizzard. */
|
|
7
|
+
model?: 'aurora' | 'blizzard';
|
|
8
|
+
/** The desired language. Two letter ISO 639-1 code. Does not work with professional clones. Not all languages work with all models. Defaults to auto language detection. */
|
|
9
|
+
language?:
|
|
10
|
+
| 'auto'
|
|
11
|
+
| 'en'
|
|
12
|
+
| 'es'
|
|
13
|
+
| 'pt'
|
|
14
|
+
| 'fr'
|
|
15
|
+
| 'de'
|
|
16
|
+
| 'zh'
|
|
17
|
+
| 'ko'
|
|
18
|
+
| 'hi'
|
|
19
|
+
| 'ja'
|
|
20
|
+
| 'ru'
|
|
21
|
+
| 'it'
|
|
22
|
+
| 'tr';
|
|
23
|
+
/** The file format of the audio output */
|
|
24
|
+
format?: 'aac' | 'mp3' | 'mulaw' | 'raw' | 'wav';
|
|
25
|
+
/** The desired output sample rate in Hz */
|
|
26
|
+
sample_rate?: 8000 | 16000 | 24000;
|
|
27
|
+
/** The talking speed of the generated speech, a floating point value between 0.25 (slow) and 2.0 (fast). */
|
|
28
|
+
speed?: number;
|
|
29
|
+
/** Seed used to specify a different take; defaults to random */
|
|
30
|
+
seed?: number;
|
|
31
|
+
/** Set this to true to generate conversational-style speech rather than reading-style speech. Does not work with the blizzard model. */
|
|
32
|
+
conversational?: boolean;
|
|
33
|
+
/** Produce speech of this length in seconds; maximum 300.0 (5 minutes). Does not work with the blizzard model. */
|
|
34
|
+
length?: number;
|
|
35
|
+
/** Controls the stability of the generated speech. A lower value (like 0.3) produces more consistent, reliable speech. A higher value (like 0.9) gives more flexibility in how words are spoken, but might occasionally produce unusual intonations or speech patterns. */
|
|
36
|
+
top_p?: number;
|
|
37
|
+
/** Influences how expressive and emotionally varied the speech becomes. Lower values (like 0.3) create more neutral, consistent speaking styles. Higher values (like 1.0) allow for more dynamic emotional range and speaking styles. */
|
|
38
|
+
temperature?: number;
|
|
39
|
+
};
|
|
@@ -0,0 +1,9 @@
|
|
|
1
|
+
import { FetchFunction } from '@ai-sdk/provider-utils';
|
|
2
|
+
|
|
3
|
+
export type LMNTConfig = {
|
|
4
|
+
provider: string;
|
|
5
|
+
url: (options: { modelId: string; path: string }) => string;
|
|
6
|
+
headers: () => Record<string, string | undefined>;
|
|
7
|
+
fetch?: FetchFunction;
|
|
8
|
+
generateId?: () => string;
|
|
9
|
+
};
|
|
@@ -0,0 +1,34 @@
|
|
|
1
|
+
import { safeParseJSON } from '@ai-sdk/provider-utils';
|
|
2
|
+
import { lmntErrorDataSchema } from './lmnt-error';
|
|
3
|
+
import { describe, it, expect } from 'vitest';
|
|
4
|
+
|
|
5
|
+
describe('lmntErrorDataSchema', () => {
|
|
6
|
+
it('should parse LMNT resource exhausted error', async () => {
|
|
7
|
+
const error = `
|
|
8
|
+
{"error":{"message":"{\\n \\"error\\": {\\n \\"code\\": 429,\\n \\"message\\": \\"Resource has been exhausted (e.g. check quota).\\",\\n \\"status\\": \\"RESOURCE_EXHAUSTED\\"\\n }\\n}\\n","code":429}}
|
|
9
|
+
`;
|
|
10
|
+
|
|
11
|
+
const result = await safeParseJSON({
|
|
12
|
+
text: error,
|
|
13
|
+
schema: lmntErrorDataSchema,
|
|
14
|
+
});
|
|
15
|
+
|
|
16
|
+
expect(result).toStrictEqual({
|
|
17
|
+
success: true,
|
|
18
|
+
value: {
|
|
19
|
+
error: {
|
|
20
|
+
message:
|
|
21
|
+
'{\n "error": {\n "code": 429,\n "message": "Resource has been exhausted (e.g. check quota).",\n "status": "RESOURCE_EXHAUSTED"\n }\n}\n',
|
|
22
|
+
code: 429,
|
|
23
|
+
},
|
|
24
|
+
},
|
|
25
|
+
rawValue: {
|
|
26
|
+
error: {
|
|
27
|
+
message:
|
|
28
|
+
'{\n "error": {\n "code": 429,\n "message": "Resource has been exhausted (e.g. check quota).",\n "status": "RESOURCE_EXHAUSTED"\n }\n}\n',
|
|
29
|
+
code: 429,
|
|
30
|
+
},
|
|
31
|
+
},
|
|
32
|
+
});
|
|
33
|
+
});
|
|
34
|
+
});
|
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
import { z } from 'zod/v4';
|
|
2
|
+
import { createJsonErrorResponseHandler } from '@ai-sdk/provider-utils';
|
|
3
|
+
|
|
4
|
+
export const lmntErrorDataSchema = z.object({
|
|
5
|
+
error: z.object({
|
|
6
|
+
message: z.string(),
|
|
7
|
+
code: z.number(),
|
|
8
|
+
}),
|
|
9
|
+
});
|
|
10
|
+
|
|
11
|
+
export type LMNTErrorData = z.infer<typeof lmntErrorDataSchema>;
|
|
12
|
+
|
|
13
|
+
export const lmntFailedResponseHandler = createJsonErrorResponseHandler({
|
|
14
|
+
errorSchema: lmntErrorDataSchema,
|
|
15
|
+
errorToMessage: data => data.error.message,
|
|
16
|
+
});
|
|
@@ -0,0 +1,83 @@
|
|
|
1
|
+
import { SpeechModelV3, ProviderV3 } from '@ai-sdk/provider';
|
|
2
|
+
import {
|
|
3
|
+
FetchFunction,
|
|
4
|
+
loadApiKey,
|
|
5
|
+
withUserAgentSuffix,
|
|
6
|
+
} from '@ai-sdk/provider-utils';
|
|
7
|
+
import { LMNTSpeechModel } from './lmnt-speech-model';
|
|
8
|
+
import { LMNTSpeechModelId } from './lmnt-speech-options';
|
|
9
|
+
import { VERSION } from './version';
|
|
10
|
+
|
|
11
|
+
export interface LMNTProvider extends Pick<ProviderV3, 'speechModel'> {
|
|
12
|
+
(
|
|
13
|
+
modelId: 'aurora',
|
|
14
|
+
settings?: {},
|
|
15
|
+
): {
|
|
16
|
+
speech: LMNTSpeechModel;
|
|
17
|
+
};
|
|
18
|
+
|
|
19
|
+
/**
|
|
20
|
+
Creates a model for speech synthesis.
|
|
21
|
+
*/
|
|
22
|
+
speech(modelId: LMNTSpeechModelId): SpeechModelV3;
|
|
23
|
+
}
|
|
24
|
+
|
|
25
|
+
export interface LMNTProviderSettings {
|
|
26
|
+
/**
|
|
27
|
+
API key for authenticating requests.
|
|
28
|
+
*/
|
|
29
|
+
apiKey?: string;
|
|
30
|
+
|
|
31
|
+
/**
|
|
32
|
+
Custom headers to include in the requests.
|
|
33
|
+
*/
|
|
34
|
+
headers?: Record<string, string>;
|
|
35
|
+
|
|
36
|
+
/**
|
|
37
|
+
Custom fetch implementation. You can use it as a middleware to intercept requests,
|
|
38
|
+
or to provide a custom fetch implementation for e.g. testing.
|
|
39
|
+
*/
|
|
40
|
+
fetch?: FetchFunction;
|
|
41
|
+
}
|
|
42
|
+
|
|
43
|
+
/**
|
|
44
|
+
Create an LMNT provider instance.
|
|
45
|
+
*/
|
|
46
|
+
export function createLMNT(options: LMNTProviderSettings = {}): LMNTProvider {
|
|
47
|
+
const getHeaders = () =>
|
|
48
|
+
withUserAgentSuffix(
|
|
49
|
+
{
|
|
50
|
+
'x-api-key': loadApiKey({
|
|
51
|
+
apiKey: options.apiKey,
|
|
52
|
+
environmentVariableName: 'LMNT_API_KEY',
|
|
53
|
+
description: 'LMNT',
|
|
54
|
+
}),
|
|
55
|
+
...options.headers,
|
|
56
|
+
},
|
|
57
|
+
`ai-sdk/lmnt/${VERSION}`,
|
|
58
|
+
);
|
|
59
|
+
|
|
60
|
+
const createSpeechModel = (modelId: LMNTSpeechModelId) =>
|
|
61
|
+
new LMNTSpeechModel(modelId, {
|
|
62
|
+
provider: `lmnt.speech`,
|
|
63
|
+
url: ({ path }) => `https://api.lmnt.com${path}`,
|
|
64
|
+
headers: getHeaders,
|
|
65
|
+
fetch: options.fetch,
|
|
66
|
+
});
|
|
67
|
+
|
|
68
|
+
const provider = function (modelId: LMNTSpeechModelId) {
|
|
69
|
+
return {
|
|
70
|
+
speech: createSpeechModel(modelId),
|
|
71
|
+
};
|
|
72
|
+
};
|
|
73
|
+
|
|
74
|
+
provider.speech = createSpeechModel;
|
|
75
|
+
provider.speechModel = createSpeechModel;
|
|
76
|
+
|
|
77
|
+
return provider as LMNTProvider;
|
|
78
|
+
}
|
|
79
|
+
|
|
80
|
+
/**
|
|
81
|
+
Default LMNT provider instance.
|
|
82
|
+
*/
|
|
83
|
+
export const lmnt = createLMNT();
|
|
@@ -0,0 +1,197 @@
|
|
|
1
|
+
import { createTestServer } from '@ai-sdk/test-server/with-vitest';
|
|
2
|
+
import { LMNTSpeechModel } from './lmnt-speech-model';
|
|
3
|
+
import { createLMNT } from './lmnt-provider';
|
|
4
|
+
import { describe, it, expect, vi } from 'vitest';
|
|
5
|
+
|
|
6
|
+
vi.mock('./version', () => ({
|
|
7
|
+
VERSION: '0.0.0-test',
|
|
8
|
+
}));
|
|
9
|
+
|
|
10
|
+
const provider = createLMNT({ apiKey: 'test-api-key' });
|
|
11
|
+
const model = provider.speech('aurora');
|
|
12
|
+
|
|
13
|
+
const server = createTestServer({
|
|
14
|
+
'https://api.lmnt.com/v1/ai/speech/bytes': {},
|
|
15
|
+
});
|
|
16
|
+
|
|
17
|
+
describe('doGenerate', () => {
|
|
18
|
+
function prepareAudioResponse({
|
|
19
|
+
headers,
|
|
20
|
+
format = 'mp3',
|
|
21
|
+
}: {
|
|
22
|
+
headers?: Record<string, string>;
|
|
23
|
+
format?: 'aac' | 'mp3' | 'mulaw' | 'raw' | 'wav';
|
|
24
|
+
} = {}) {
|
|
25
|
+
const audioBuffer = new Uint8Array(100); // Mock audio data
|
|
26
|
+
server.urls['https://api.lmnt.com/v1/ai/speech/bytes'].response = {
|
|
27
|
+
type: 'binary',
|
|
28
|
+
headers: {
|
|
29
|
+
'content-type': `audio/${format}`,
|
|
30
|
+
...headers,
|
|
31
|
+
},
|
|
32
|
+
body: Buffer.from(audioBuffer),
|
|
33
|
+
};
|
|
34
|
+
return audioBuffer;
|
|
35
|
+
}
|
|
36
|
+
|
|
37
|
+
it('should pass the model and text', async () => {
|
|
38
|
+
prepareAudioResponse();
|
|
39
|
+
|
|
40
|
+
await model.doGenerate({
|
|
41
|
+
text: 'Hello from the AI SDK!',
|
|
42
|
+
});
|
|
43
|
+
|
|
44
|
+
expect(await server.calls[0].requestBodyJson).toMatchObject({
|
|
45
|
+
model: 'aurora',
|
|
46
|
+
text: 'Hello from the AI SDK!',
|
|
47
|
+
});
|
|
48
|
+
});
|
|
49
|
+
|
|
50
|
+
it('should pass headers', async () => {
|
|
51
|
+
prepareAudioResponse();
|
|
52
|
+
|
|
53
|
+
const provider = createLMNT({
|
|
54
|
+
apiKey: 'test-api-key',
|
|
55
|
+
headers: {
|
|
56
|
+
'Custom-Provider-Header': 'provider-header-value',
|
|
57
|
+
},
|
|
58
|
+
});
|
|
59
|
+
|
|
60
|
+
await provider.speech('aurora').doGenerate({
|
|
61
|
+
text: 'Hello from the AI SDK!',
|
|
62
|
+
headers: {
|
|
63
|
+
'Custom-Request-Header': 'request-header-value',
|
|
64
|
+
},
|
|
65
|
+
});
|
|
66
|
+
|
|
67
|
+
expect(server.calls[0].requestHeaders).toMatchObject({
|
|
68
|
+
'x-api-key': 'test-api-key',
|
|
69
|
+
'content-type': 'application/json',
|
|
70
|
+
'custom-provider-header': 'provider-header-value',
|
|
71
|
+
'custom-request-header': 'request-header-value',
|
|
72
|
+
});
|
|
73
|
+
expect(server.calls[0].requestUserAgent).toContain(
|
|
74
|
+
`ai-sdk/lmnt/0.0.0-test`,
|
|
75
|
+
);
|
|
76
|
+
});
|
|
77
|
+
|
|
78
|
+
it('should pass options', async () => {
|
|
79
|
+
prepareAudioResponse();
|
|
80
|
+
|
|
81
|
+
await model.doGenerate({
|
|
82
|
+
text: 'Hello from the AI SDK!',
|
|
83
|
+
voice: 'nova',
|
|
84
|
+
outputFormat: 'mp3',
|
|
85
|
+
speed: 1.5,
|
|
86
|
+
});
|
|
87
|
+
|
|
88
|
+
expect(await server.calls[0].requestBodyJson).toMatchObject({
|
|
89
|
+
model: 'aurora',
|
|
90
|
+
text: 'Hello from the AI SDK!',
|
|
91
|
+
voice: 'nova',
|
|
92
|
+
speed: 1.5,
|
|
93
|
+
response_format: 'mp3',
|
|
94
|
+
});
|
|
95
|
+
});
|
|
96
|
+
|
|
97
|
+
it('should return audio data with correct content type', async () => {
|
|
98
|
+
const audio = new Uint8Array(100); // Mock audio data
|
|
99
|
+
prepareAudioResponse({
|
|
100
|
+
format: 'mp3',
|
|
101
|
+
headers: {
|
|
102
|
+
'x-request-id': 'test-request-id',
|
|
103
|
+
'x-ratelimit-remaining': '123',
|
|
104
|
+
},
|
|
105
|
+
});
|
|
106
|
+
|
|
107
|
+
const result = await model.doGenerate({
|
|
108
|
+
text: 'Hello from the AI SDK!',
|
|
109
|
+
outputFormat: 'mp3',
|
|
110
|
+
});
|
|
111
|
+
|
|
112
|
+
expect(result.audio).toStrictEqual(audio);
|
|
113
|
+
});
|
|
114
|
+
|
|
115
|
+
it('should include response data with timestamp, modelId and headers', async () => {
|
|
116
|
+
prepareAudioResponse({
|
|
117
|
+
headers: {
|
|
118
|
+
'x-request-id': 'test-request-id',
|
|
119
|
+
'x-ratelimit-remaining': '123',
|
|
120
|
+
},
|
|
121
|
+
});
|
|
122
|
+
|
|
123
|
+
const testDate = new Date(0);
|
|
124
|
+
const customModel = new LMNTSpeechModel('aurora', {
|
|
125
|
+
provider: 'test-provider',
|
|
126
|
+
url: () => 'https://api.lmnt.com/v1/ai/speech/bytes',
|
|
127
|
+
headers: () => ({}),
|
|
128
|
+
_internal: {
|
|
129
|
+
currentDate: () => testDate,
|
|
130
|
+
},
|
|
131
|
+
});
|
|
132
|
+
|
|
133
|
+
const result = await customModel.doGenerate({
|
|
134
|
+
text: 'Hello from the AI SDK!',
|
|
135
|
+
});
|
|
136
|
+
|
|
137
|
+
expect(result.response).toMatchObject({
|
|
138
|
+
timestamp: testDate,
|
|
139
|
+
modelId: 'aurora',
|
|
140
|
+
headers: {
|
|
141
|
+
'content-type': 'audio/mp3',
|
|
142
|
+
'x-request-id': 'test-request-id',
|
|
143
|
+
'x-ratelimit-remaining': '123',
|
|
144
|
+
},
|
|
145
|
+
});
|
|
146
|
+
});
|
|
147
|
+
|
|
148
|
+
it('should use real date when no custom date provider is specified', async () => {
|
|
149
|
+
prepareAudioResponse();
|
|
150
|
+
|
|
151
|
+
const testDate = new Date(0);
|
|
152
|
+
const customModel = new LMNTSpeechModel('aurora', {
|
|
153
|
+
provider: 'test-provider',
|
|
154
|
+
url: () => 'https://api.lmnt.com/v1/ai/speech/bytes',
|
|
155
|
+
headers: () => ({}),
|
|
156
|
+
_internal: {
|
|
157
|
+
currentDate: () => testDate,
|
|
158
|
+
},
|
|
159
|
+
});
|
|
160
|
+
|
|
161
|
+
const result = await customModel.doGenerate({
|
|
162
|
+
text: 'Hello from the AI SDK!',
|
|
163
|
+
});
|
|
164
|
+
|
|
165
|
+
expect(result.response.timestamp.getTime()).toEqual(testDate.getTime());
|
|
166
|
+
expect(result.response.modelId).toBe('aurora');
|
|
167
|
+
});
|
|
168
|
+
|
|
169
|
+
it('should handle different audio formats', async () => {
|
|
170
|
+
const formats = ['aac', 'mp3', 'mulaw', 'raw', 'wav'] as const;
|
|
171
|
+
|
|
172
|
+
for (const format of formats) {
|
|
173
|
+
const audio = prepareAudioResponse({ format });
|
|
174
|
+
|
|
175
|
+
const result = await model.doGenerate({
|
|
176
|
+
text: 'Hello from the AI SDK!',
|
|
177
|
+
providerOptions: {
|
|
178
|
+
lmnt: {
|
|
179
|
+
format,
|
|
180
|
+
},
|
|
181
|
+
},
|
|
182
|
+
});
|
|
183
|
+
|
|
184
|
+
expect(result.audio).toStrictEqual(audio);
|
|
185
|
+
}
|
|
186
|
+
});
|
|
187
|
+
|
|
188
|
+
it('should include warnings if any are generated', async () => {
|
|
189
|
+
prepareAudioResponse();
|
|
190
|
+
|
|
191
|
+
const result = await model.doGenerate({
|
|
192
|
+
text: 'Hello from the AI SDK!',
|
|
193
|
+
});
|
|
194
|
+
|
|
195
|
+
expect(result.warnings).toEqual([]);
|
|
196
|
+
});
|
|
197
|
+
});
|
|
@@ -0,0 +1,206 @@
|
|
|
1
|
+
import { SpeechModelV3, SharedV3Warning } from '@ai-sdk/provider';
|
|
2
|
+
import {
|
|
3
|
+
combineHeaders,
|
|
4
|
+
createBinaryResponseHandler,
|
|
5
|
+
parseProviderOptions,
|
|
6
|
+
postJsonToApi,
|
|
7
|
+
} from '@ai-sdk/provider-utils';
|
|
8
|
+
import { z } from 'zod/v4';
|
|
9
|
+
import { LMNTConfig } from './lmnt-config';
|
|
10
|
+
import { lmntFailedResponseHandler } from './lmnt-error';
|
|
11
|
+
import { LMNTSpeechModelId } from './lmnt-speech-options';
|
|
12
|
+
import { LMNTSpeechAPITypes } from './lmnt-api-types';
|
|
13
|
+
|
|
14
|
+
// https://docs.lmnt.com/api-reference/speech/synthesize-speech-bytes
|
|
15
|
+
const lmntSpeechCallOptionsSchema = z.object({
|
|
16
|
+
/**
|
|
17
|
+
* The model to use for speech synthesis e.g. 'aurora' or 'blizzard'.
|
|
18
|
+
* @default 'aurora'
|
|
19
|
+
*/
|
|
20
|
+
model: z
|
|
21
|
+
.union([z.enum(['aurora', 'blizzard']), z.string()])
|
|
22
|
+
.nullish()
|
|
23
|
+
.default('aurora'),
|
|
24
|
+
|
|
25
|
+
/**
|
|
26
|
+
* The audio format of the output.
|
|
27
|
+
* @default 'mp3'
|
|
28
|
+
*/
|
|
29
|
+
format: z
|
|
30
|
+
.enum(['aac', 'mp3', 'mulaw', 'raw', 'wav'])
|
|
31
|
+
.nullish()
|
|
32
|
+
.default('mp3'),
|
|
33
|
+
|
|
34
|
+
/**
|
|
35
|
+
* The sample rate of the output audio in Hz.
|
|
36
|
+
* @default 24000
|
|
37
|
+
*/
|
|
38
|
+
sampleRate: z
|
|
39
|
+
.union([z.literal(8000), z.literal(16000), z.literal(24000)])
|
|
40
|
+
.nullish()
|
|
41
|
+
.default(24000),
|
|
42
|
+
|
|
43
|
+
/**
|
|
44
|
+
* The speed of the speech. Range: 0.25 to 2.
|
|
45
|
+
* @default 1
|
|
46
|
+
*/
|
|
47
|
+
speed: z.number().min(0.25).max(2).nullish().default(1),
|
|
48
|
+
|
|
49
|
+
/**
|
|
50
|
+
* A seed value for deterministic generation.
|
|
51
|
+
*/
|
|
52
|
+
seed: z.number().int().nullish(),
|
|
53
|
+
|
|
54
|
+
/**
|
|
55
|
+
* Whether to use a conversational style.
|
|
56
|
+
* @default false
|
|
57
|
+
*/
|
|
58
|
+
conversational: z.boolean().nullish().default(false),
|
|
59
|
+
|
|
60
|
+
/**
|
|
61
|
+
* Maximum length of the output in seconds (up to 300).
|
|
62
|
+
*/
|
|
63
|
+
length: z.number().max(300).nullish(),
|
|
64
|
+
|
|
65
|
+
/**
|
|
66
|
+
* Top-p sampling parameter. Range: 0 to 1.
|
|
67
|
+
* @default 1
|
|
68
|
+
*/
|
|
69
|
+
topP: z.number().min(0).max(1).nullish().default(1),
|
|
70
|
+
|
|
71
|
+
/**
|
|
72
|
+
* Temperature for sampling. Higher values increase randomness.
|
|
73
|
+
* @default 1
|
|
74
|
+
*/
|
|
75
|
+
temperature: z.number().min(0).nullish().default(1),
|
|
76
|
+
});
|
|
77
|
+
|
|
78
|
+
export type LMNTSpeechCallOptions = z.infer<typeof lmntSpeechCallOptionsSchema>;
|
|
79
|
+
|
|
80
|
+
interface LMNTSpeechModelConfig extends LMNTConfig {
|
|
81
|
+
_internal?: {
|
|
82
|
+
currentDate?: () => Date;
|
|
83
|
+
};
|
|
84
|
+
}
|
|
85
|
+
|
|
86
|
+
export class LMNTSpeechModel implements SpeechModelV3 {
|
|
87
|
+
readonly specificationVersion = 'v3';
|
|
88
|
+
|
|
89
|
+
get provider(): string {
|
|
90
|
+
return this.config.provider;
|
|
91
|
+
}
|
|
92
|
+
|
|
93
|
+
constructor(
|
|
94
|
+
readonly modelId: LMNTSpeechModelId,
|
|
95
|
+
private readonly config: LMNTSpeechModelConfig,
|
|
96
|
+
) {}
|
|
97
|
+
|
|
98
|
+
private async getArgs({
|
|
99
|
+
text,
|
|
100
|
+
voice = 'ava',
|
|
101
|
+
outputFormat = 'mp3',
|
|
102
|
+
speed,
|
|
103
|
+
language,
|
|
104
|
+
providerOptions,
|
|
105
|
+
}: Parameters<SpeechModelV3['doGenerate']>[0]) {
|
|
106
|
+
const warnings: SharedV3Warning[] = [];
|
|
107
|
+
|
|
108
|
+
// Parse provider options
|
|
109
|
+
const lmntOptions = await parseProviderOptions({
|
|
110
|
+
provider: 'lmnt',
|
|
111
|
+
providerOptions,
|
|
112
|
+
schema: lmntSpeechCallOptionsSchema,
|
|
113
|
+
});
|
|
114
|
+
|
|
115
|
+
// Create request body
|
|
116
|
+
const requestBody: Record<string, unknown> = {
|
|
117
|
+
model: this.modelId,
|
|
118
|
+
text,
|
|
119
|
+
voice,
|
|
120
|
+
response_format: 'mp3',
|
|
121
|
+
speed,
|
|
122
|
+
};
|
|
123
|
+
|
|
124
|
+
if (outputFormat) {
|
|
125
|
+
if (['mp3', 'aac', 'mulaw', 'raw', 'wav'].includes(outputFormat)) {
|
|
126
|
+
requestBody.response_format = outputFormat;
|
|
127
|
+
} else {
|
|
128
|
+
warnings.push({
|
|
129
|
+
type: 'unsupported',
|
|
130
|
+
feature: 'outputFormat',
|
|
131
|
+
details: `Unsupported output format: ${outputFormat}. Using mp3 instead.`,
|
|
132
|
+
});
|
|
133
|
+
}
|
|
134
|
+
}
|
|
135
|
+
|
|
136
|
+
// Add provider-specific options
|
|
137
|
+
if (lmntOptions) {
|
|
138
|
+
const speechModelOptions: Omit<LMNTSpeechAPITypes, 'voice' | 'text'> = {
|
|
139
|
+
conversational: lmntOptions.conversational ?? undefined,
|
|
140
|
+
length: lmntOptions.length ?? undefined,
|
|
141
|
+
seed: lmntOptions.seed ?? undefined,
|
|
142
|
+
speed: lmntOptions.speed ?? undefined,
|
|
143
|
+
temperature: lmntOptions.temperature ?? undefined,
|
|
144
|
+
top_p: lmntOptions.topP ?? undefined,
|
|
145
|
+
sample_rate: lmntOptions.sampleRate ?? undefined,
|
|
146
|
+
};
|
|
147
|
+
|
|
148
|
+
for (const key in speechModelOptions) {
|
|
149
|
+
const value =
|
|
150
|
+
speechModelOptions[
|
|
151
|
+
key as keyof Omit<LMNTSpeechAPITypes, 'voice' | 'text'>
|
|
152
|
+
];
|
|
153
|
+
if (value !== undefined) {
|
|
154
|
+
requestBody[key] = value;
|
|
155
|
+
}
|
|
156
|
+
}
|
|
157
|
+
}
|
|
158
|
+
|
|
159
|
+
if (language) {
|
|
160
|
+
requestBody.language = language;
|
|
161
|
+
}
|
|
162
|
+
|
|
163
|
+
return {
|
|
164
|
+
requestBody,
|
|
165
|
+
warnings,
|
|
166
|
+
};
|
|
167
|
+
}
|
|
168
|
+
|
|
169
|
+
async doGenerate(
|
|
170
|
+
options: Parameters<SpeechModelV3['doGenerate']>[0],
|
|
171
|
+
): Promise<Awaited<ReturnType<SpeechModelV3['doGenerate']>>> {
|
|
172
|
+
const currentDate = this.config._internal?.currentDate?.() ?? new Date();
|
|
173
|
+
const { requestBody, warnings } = await this.getArgs(options);
|
|
174
|
+
|
|
175
|
+
const {
|
|
176
|
+
value: audio,
|
|
177
|
+
responseHeaders,
|
|
178
|
+
rawValue: rawResponse,
|
|
179
|
+
} = await postJsonToApi({
|
|
180
|
+
url: this.config.url({
|
|
181
|
+
path: '/v1/ai/speech/bytes',
|
|
182
|
+
modelId: this.modelId,
|
|
183
|
+
}),
|
|
184
|
+
headers: combineHeaders(this.config.headers(), options.headers),
|
|
185
|
+
body: requestBody,
|
|
186
|
+
failedResponseHandler: lmntFailedResponseHandler,
|
|
187
|
+
successfulResponseHandler: createBinaryResponseHandler(),
|
|
188
|
+
abortSignal: options.abortSignal,
|
|
189
|
+
fetch: this.config.fetch,
|
|
190
|
+
});
|
|
191
|
+
|
|
192
|
+
return {
|
|
193
|
+
audio,
|
|
194
|
+
warnings,
|
|
195
|
+
request: {
|
|
196
|
+
body: JSON.stringify(requestBody),
|
|
197
|
+
},
|
|
198
|
+
response: {
|
|
199
|
+
timestamp: currentDate,
|
|
200
|
+
modelId: this.modelId,
|
|
201
|
+
headers: responseHeaders,
|
|
202
|
+
body: rawResponse,
|
|
203
|
+
},
|
|
204
|
+
};
|
|
205
|
+
}
|
|
206
|
+
}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
export type LMNTSpeechModelId = 'aurora' | 'blizzard' | (string & {});
|