@ai-sdk/deepseek 0.0.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +7 -0
- package/LICENSE +13 -0
- package/README.md +35 -0
- package/dist/index.d.mts +46 -0
- package/dist/index.d.ts +46 -0
- package/dist/index.js +68 -0
- package/dist/index.js.map +1 -0
- package/dist/index.mjs +45 -0
- package/dist/index.mjs.map +1 -0
- package/package.json +64 -0
package/CHANGELOG.md
ADDED
package/LICENSE
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
1
|
+
Copyright 2023 Vercel, Inc.
|
|
2
|
+
|
|
3
|
+
Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
+
you may not use this file except in compliance with the License.
|
|
5
|
+
You may obtain a copy of the License at
|
|
6
|
+
|
|
7
|
+
http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
+
|
|
9
|
+
Unless required by applicable law or agreed to in writing, software
|
|
10
|
+
distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
+
See the License for the specific language governing permissions and
|
|
13
|
+
limitations under the License.
|
package/README.md
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
# AI SDK - DeepSeek Provider
|
|
2
|
+
|
|
3
|
+
The **[DeepSeek provider](https://sdk.vercel.ai/providers/ai-sdk-providers/deepseek)** for the [AI SDK](https://sdk.vercel.ai/docs) contains language model support for the [DeepSeek](https://www.deepseek.com) platform.
|
|
4
|
+
|
|
5
|
+
## Setup
|
|
6
|
+
|
|
7
|
+
The DeepSeek provider is available in the `@ai-sdk/deepseek` module. You can install it with
|
|
8
|
+
|
|
9
|
+
```bash
|
|
10
|
+
npm i @ai-sdk/deepseek
|
|
11
|
+
```
|
|
12
|
+
|
|
13
|
+
## Provider Instance
|
|
14
|
+
|
|
15
|
+
You can import the default provider instance `deepseek` from `@ai-sdk/deepseek`:
|
|
16
|
+
|
|
17
|
+
```ts
|
|
18
|
+
import { deepseek } from '@ai-sdk/deepseek';
|
|
19
|
+
```
|
|
20
|
+
|
|
21
|
+
## Example
|
|
22
|
+
|
|
23
|
+
```ts
|
|
24
|
+
import { deepseek } from '@ai-sdk/deepseek';
|
|
25
|
+
import { generateText } from 'ai';
|
|
26
|
+
|
|
27
|
+
const { text } = await generateText({
|
|
28
|
+
model: deepseek('deepseek-chat'),
|
|
29
|
+
prompt: 'Write a JavaScript function that sorts a list:',
|
|
30
|
+
});
|
|
31
|
+
```
|
|
32
|
+
|
|
33
|
+
## Documentation
|
|
34
|
+
|
|
35
|
+
Please check out the **[DeepSeek provider](https://sdk.vercel.ai/providers/ai-sdk-providers/deepseek)** for more information.
|
package/dist/index.d.mts
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
1
|
+
import { ProviderV1, LanguageModelV1 } from '@ai-sdk/provider';
|
|
2
|
+
import { FetchFunction } from '@ai-sdk/provider-utils';
|
|
3
|
+
import { OpenAICompatibleChatSettings } from '@ai-sdk/openai-compatible';
|
|
4
|
+
export { OpenAICompatibleErrorData as DeepSeekErrorData } from '@ai-sdk/openai-compatible';
|
|
5
|
+
|
|
6
|
+
type DeepSeekChatModelId = 'deepseek-chat' | (string & {});
|
|
7
|
+
interface DeepSeekChatSettings extends OpenAICompatibleChatSettings {
|
|
8
|
+
}
|
|
9
|
+
|
|
10
|
+
interface DeepSeekProviderSettings {
|
|
11
|
+
/**
|
|
12
|
+
DeepSeek API key.
|
|
13
|
+
*/
|
|
14
|
+
apiKey?: string;
|
|
15
|
+
/**
|
|
16
|
+
Base URL for the API calls.
|
|
17
|
+
*/
|
|
18
|
+
baseURL?: string;
|
|
19
|
+
/**
|
|
20
|
+
Custom headers to include in the requests.
|
|
21
|
+
*/
|
|
22
|
+
headers?: Record<string, string>;
|
|
23
|
+
/**
|
|
24
|
+
Custom fetch implementation. You can use it as a middleware to intercept requests,
|
|
25
|
+
or to provide a custom fetch implementation for e.g. testing.
|
|
26
|
+
*/
|
|
27
|
+
fetch?: FetchFunction;
|
|
28
|
+
}
|
|
29
|
+
interface DeepSeekProvider extends ProviderV1 {
|
|
30
|
+
/**
|
|
31
|
+
Creates a DeepSeek model for text generation.
|
|
32
|
+
*/
|
|
33
|
+
(modelId: DeepSeekChatModelId, settings?: DeepSeekChatSettings): LanguageModelV1;
|
|
34
|
+
/**
|
|
35
|
+
Creates a DeepSeek model for text generation.
|
|
36
|
+
*/
|
|
37
|
+
languageModel(modelId: DeepSeekChatModelId, settings?: DeepSeekChatSettings): LanguageModelV1;
|
|
38
|
+
/**
|
|
39
|
+
Creates a DeepSeek chat model for text generation.
|
|
40
|
+
*/
|
|
41
|
+
chat(modelId: DeepSeekChatModelId, settings?: DeepSeekChatSettings): LanguageModelV1;
|
|
42
|
+
}
|
|
43
|
+
declare function createDeepSeek(options?: DeepSeekProviderSettings): DeepSeekProvider;
|
|
44
|
+
declare const deepseek: DeepSeekProvider;
|
|
45
|
+
|
|
46
|
+
export { type DeepSeekProvider, type DeepSeekProviderSettings, createDeepSeek, deepseek };
|
package/dist/index.d.ts
ADDED
|
@@ -0,0 +1,46 @@
|
|
|
1
|
+
import { ProviderV1, LanguageModelV1 } from '@ai-sdk/provider';
|
|
2
|
+
import { FetchFunction } from '@ai-sdk/provider-utils';
|
|
3
|
+
import { OpenAICompatibleChatSettings } from '@ai-sdk/openai-compatible';
|
|
4
|
+
export { OpenAICompatibleErrorData as DeepSeekErrorData } from '@ai-sdk/openai-compatible';
|
|
5
|
+
|
|
6
|
+
type DeepSeekChatModelId = 'deepseek-chat' | (string & {});
|
|
7
|
+
interface DeepSeekChatSettings extends OpenAICompatibleChatSettings {
|
|
8
|
+
}
|
|
9
|
+
|
|
10
|
+
interface DeepSeekProviderSettings {
|
|
11
|
+
/**
|
|
12
|
+
DeepSeek API key.
|
|
13
|
+
*/
|
|
14
|
+
apiKey?: string;
|
|
15
|
+
/**
|
|
16
|
+
Base URL for the API calls.
|
|
17
|
+
*/
|
|
18
|
+
baseURL?: string;
|
|
19
|
+
/**
|
|
20
|
+
Custom headers to include in the requests.
|
|
21
|
+
*/
|
|
22
|
+
headers?: Record<string, string>;
|
|
23
|
+
/**
|
|
24
|
+
Custom fetch implementation. You can use it as a middleware to intercept requests,
|
|
25
|
+
or to provide a custom fetch implementation for e.g. testing.
|
|
26
|
+
*/
|
|
27
|
+
fetch?: FetchFunction;
|
|
28
|
+
}
|
|
29
|
+
interface DeepSeekProvider extends ProviderV1 {
|
|
30
|
+
/**
|
|
31
|
+
Creates a DeepSeek model for text generation.
|
|
32
|
+
*/
|
|
33
|
+
(modelId: DeepSeekChatModelId, settings?: DeepSeekChatSettings): LanguageModelV1;
|
|
34
|
+
/**
|
|
35
|
+
Creates a DeepSeek model for text generation.
|
|
36
|
+
*/
|
|
37
|
+
languageModel(modelId: DeepSeekChatModelId, settings?: DeepSeekChatSettings): LanguageModelV1;
|
|
38
|
+
/**
|
|
39
|
+
Creates a DeepSeek chat model for text generation.
|
|
40
|
+
*/
|
|
41
|
+
chat(modelId: DeepSeekChatModelId, settings?: DeepSeekChatSettings): LanguageModelV1;
|
|
42
|
+
}
|
|
43
|
+
declare function createDeepSeek(options?: DeepSeekProviderSettings): DeepSeekProvider;
|
|
44
|
+
declare const deepseek: DeepSeekProvider;
|
|
45
|
+
|
|
46
|
+
export { type DeepSeekProvider, type DeepSeekProviderSettings, createDeepSeek, deepseek };
|
package/dist/index.js
ADDED
|
@@ -0,0 +1,68 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __defProp = Object.defineProperty;
|
|
3
|
+
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
4
|
+
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
5
|
+
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
6
|
+
var __export = (target, all) => {
|
|
7
|
+
for (var name in all)
|
|
8
|
+
__defProp(target, name, { get: all[name], enumerable: true });
|
|
9
|
+
};
|
|
10
|
+
var __copyProps = (to, from, except, desc) => {
|
|
11
|
+
if (from && typeof from === "object" || typeof from === "function") {
|
|
12
|
+
for (let key of __getOwnPropNames(from))
|
|
13
|
+
if (!__hasOwnProp.call(to, key) && key !== except)
|
|
14
|
+
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
15
|
+
}
|
|
16
|
+
return to;
|
|
17
|
+
};
|
|
18
|
+
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
19
|
+
|
|
20
|
+
// src/index.ts
|
|
21
|
+
var src_exports = {};
|
|
22
|
+
__export(src_exports, {
|
|
23
|
+
createDeepSeek: () => createDeepSeek,
|
|
24
|
+
deepseek: () => deepseek
|
|
25
|
+
});
|
|
26
|
+
module.exports = __toCommonJS(src_exports);
|
|
27
|
+
|
|
28
|
+
// src/deepseek-provider.ts
|
|
29
|
+
var import_openai_compatible = require("@ai-sdk/openai-compatible");
|
|
30
|
+
var import_provider = require("@ai-sdk/provider");
|
|
31
|
+
var import_provider_utils = require("@ai-sdk/provider-utils");
|
|
32
|
+
function createDeepSeek(options = {}) {
|
|
33
|
+
var _a;
|
|
34
|
+
const baseURL = (0, import_provider_utils.withoutTrailingSlash)(
|
|
35
|
+
(_a = options.baseURL) != null ? _a : "https://api.deepseek.com/v1"
|
|
36
|
+
);
|
|
37
|
+
const getHeaders = () => ({
|
|
38
|
+
Authorization: `Bearer ${(0, import_provider_utils.loadApiKey)({
|
|
39
|
+
apiKey: options.apiKey,
|
|
40
|
+
environmentVariableName: "DEEPSEEK_API_KEY",
|
|
41
|
+
description: "DeepSeek API key"
|
|
42
|
+
})}`,
|
|
43
|
+
...options.headers
|
|
44
|
+
});
|
|
45
|
+
const createLanguageModel = (modelId, settings = {}) => {
|
|
46
|
+
return new import_openai_compatible.OpenAICompatibleChatLanguageModel(modelId, settings, {
|
|
47
|
+
provider: `deepseek.chat`,
|
|
48
|
+
url: ({ path }) => `${baseURL}${path}`,
|
|
49
|
+
headers: getHeaders,
|
|
50
|
+
fetch: options.fetch,
|
|
51
|
+
defaultObjectGenerationMode: "json"
|
|
52
|
+
});
|
|
53
|
+
};
|
|
54
|
+
const provider = (modelId, settings) => createLanguageModel(modelId, settings);
|
|
55
|
+
provider.languageModel = createLanguageModel;
|
|
56
|
+
provider.chat = createLanguageModel;
|
|
57
|
+
provider.textEmbeddingModel = (modelId) => {
|
|
58
|
+
throw new import_provider.NoSuchModelError({ modelId, modelType: "textEmbeddingModel" });
|
|
59
|
+
};
|
|
60
|
+
return provider;
|
|
61
|
+
}
|
|
62
|
+
var deepseek = createDeepSeek();
|
|
63
|
+
// Annotate the CommonJS export names for ESM import in node:
|
|
64
|
+
0 && (module.exports = {
|
|
65
|
+
createDeepSeek,
|
|
66
|
+
deepseek
|
|
67
|
+
});
|
|
68
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/index.ts","../src/deepseek-provider.ts"],"sourcesContent":["export { createDeepSeek, deepseek } from './deepseek-provider';\nexport type {\n DeepSeekProvider,\n DeepSeekProviderSettings,\n} from './deepseek-provider';\nexport type { OpenAICompatibleErrorData as DeepSeekErrorData } from '@ai-sdk/openai-compatible';\n","import { OpenAICompatibleChatLanguageModel } from '@ai-sdk/openai-compatible';\nimport {\n LanguageModelV1,\n NoSuchModelError,\n ProviderV1,\n} from '@ai-sdk/provider';\nimport {\n FetchFunction,\n loadApiKey,\n withoutTrailingSlash,\n} from '@ai-sdk/provider-utils';\nimport {\n DeepSeekChatModelId,\n DeepSeekChatSettings,\n} from './deepseek-chat-settings';\n\nexport interface DeepSeekProviderSettings {\n /**\nDeepSeek API key.\n*/\n apiKey?: string;\n /**\nBase URL for the API calls.\n*/\n baseURL?: string;\n /**\nCustom headers to include in the requests.\n*/\n headers?: Record<string, string>;\n /**\nCustom fetch implementation. You can use it as a middleware to intercept requests,\nor to provide a custom fetch implementation for e.g. testing.\n*/\n fetch?: FetchFunction;\n}\n\nexport interface DeepSeekProvider extends ProviderV1 {\n /**\nCreates a DeepSeek model for text generation.\n*/\n (\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ): LanguageModelV1;\n\n /**\nCreates a DeepSeek model for text generation.\n*/\n languageModel(\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ): LanguageModelV1;\n\n /**\nCreates a DeepSeek chat model for text generation.\n*/\n chat(\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ): LanguageModelV1;\n}\n\nexport function createDeepSeek(\n options: DeepSeekProviderSettings = {},\n): DeepSeekProvider {\n const baseURL = withoutTrailingSlash(\n options.baseURL ?? 'https://api.deepseek.com/v1',\n );\n const getHeaders = () => ({\n Authorization: `Bearer ${loadApiKey({\n apiKey: options.apiKey,\n environmentVariableName: 'DEEPSEEK_API_KEY',\n description: 'DeepSeek API key',\n })}`,\n ...options.headers,\n });\n\n const createLanguageModel = (\n modelId: DeepSeekChatModelId,\n settings: DeepSeekChatSettings = {},\n ) => {\n return new OpenAICompatibleChatLanguageModel(modelId, settings, {\n provider: `deepseek.chat`,\n url: ({ path }) => `${baseURL}${path}`,\n headers: getHeaders,\n fetch: options.fetch,\n defaultObjectGenerationMode: 'json',\n });\n };\n\n const provider = (\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ) => createLanguageModel(modelId, settings);\n\n provider.languageModel = createLanguageModel;\n provider.chat = createLanguageModel;\n provider.textEmbeddingModel = (modelId: string) => {\n throw new NoSuchModelError({ modelId, modelType: 'textEmbeddingModel' });\n };\n\n return provider as DeepSeekProvider;\n}\n\nexport const deepseek = createDeepSeek();\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,+BAAkD;AAClD,sBAIO;AACP,4BAIO;AAoDA,SAAS,eACd,UAAoC,CAAC,GACnB;AAhEpB;AAiEE,QAAM,cAAU;AAAA,KACd,aAAQ,YAAR,YAAmB;AAAA,EACrB;AACA,QAAM,aAAa,OAAO;AAAA,IACxB,eAAe,cAAU,kCAAW;AAAA,MAClC,QAAQ,QAAQ;AAAA,MAChB,yBAAyB;AAAA,MACzB,aAAa;AAAA,IACf,CAAC,CAAC;AAAA,IACF,GAAG,QAAQ;AAAA,EACb;AAEA,QAAM,sBAAsB,CAC1B,SACA,WAAiC,CAAC,MAC/B;AACH,WAAO,IAAI,2DAAkC,SAAS,UAAU;AAAA,MAC9D,UAAU;AAAA,MACV,KAAK,CAAC,EAAE,KAAK,MAAM,GAAG,OAAO,GAAG,IAAI;AAAA,MACpC,SAAS;AAAA,MACT,OAAO,QAAQ;AAAA,MACf,6BAA6B;AAAA,IAC/B,CAAC;AAAA,EACH;AAEA,QAAM,WAAW,CACf,SACA,aACG,oBAAoB,SAAS,QAAQ;AAE1C,WAAS,gBAAgB;AACzB,WAAS,OAAO;AAChB,WAAS,qBAAqB,CAAC,YAAoB;AACjD,UAAM,IAAI,iCAAiB,EAAE,SAAS,WAAW,qBAAqB,CAAC;AAAA,EACzE;AAEA,SAAO;AACT;AAEO,IAAM,WAAW,eAAe;","names":[]}
|
package/dist/index.mjs
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
1
|
+
// src/deepseek-provider.ts
|
|
2
|
+
import { OpenAICompatibleChatLanguageModel } from "@ai-sdk/openai-compatible";
|
|
3
|
+
import {
|
|
4
|
+
NoSuchModelError
|
|
5
|
+
} from "@ai-sdk/provider";
|
|
6
|
+
import {
|
|
7
|
+
loadApiKey,
|
|
8
|
+
withoutTrailingSlash
|
|
9
|
+
} from "@ai-sdk/provider-utils";
|
|
10
|
+
function createDeepSeek(options = {}) {
|
|
11
|
+
var _a;
|
|
12
|
+
const baseURL = withoutTrailingSlash(
|
|
13
|
+
(_a = options.baseURL) != null ? _a : "https://api.deepseek.com/v1"
|
|
14
|
+
);
|
|
15
|
+
const getHeaders = () => ({
|
|
16
|
+
Authorization: `Bearer ${loadApiKey({
|
|
17
|
+
apiKey: options.apiKey,
|
|
18
|
+
environmentVariableName: "DEEPSEEK_API_KEY",
|
|
19
|
+
description: "DeepSeek API key"
|
|
20
|
+
})}`,
|
|
21
|
+
...options.headers
|
|
22
|
+
});
|
|
23
|
+
const createLanguageModel = (modelId, settings = {}) => {
|
|
24
|
+
return new OpenAICompatibleChatLanguageModel(modelId, settings, {
|
|
25
|
+
provider: `deepseek.chat`,
|
|
26
|
+
url: ({ path }) => `${baseURL}${path}`,
|
|
27
|
+
headers: getHeaders,
|
|
28
|
+
fetch: options.fetch,
|
|
29
|
+
defaultObjectGenerationMode: "json"
|
|
30
|
+
});
|
|
31
|
+
};
|
|
32
|
+
const provider = (modelId, settings) => createLanguageModel(modelId, settings);
|
|
33
|
+
provider.languageModel = createLanguageModel;
|
|
34
|
+
provider.chat = createLanguageModel;
|
|
35
|
+
provider.textEmbeddingModel = (modelId) => {
|
|
36
|
+
throw new NoSuchModelError({ modelId, modelType: "textEmbeddingModel" });
|
|
37
|
+
};
|
|
38
|
+
return provider;
|
|
39
|
+
}
|
|
40
|
+
var deepseek = createDeepSeek();
|
|
41
|
+
export {
|
|
42
|
+
createDeepSeek,
|
|
43
|
+
deepseek
|
|
44
|
+
};
|
|
45
|
+
//# sourceMappingURL=index.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/deepseek-provider.ts"],"sourcesContent":["import { OpenAICompatibleChatLanguageModel } from '@ai-sdk/openai-compatible';\nimport {\n LanguageModelV1,\n NoSuchModelError,\n ProviderV1,\n} from '@ai-sdk/provider';\nimport {\n FetchFunction,\n loadApiKey,\n withoutTrailingSlash,\n} from '@ai-sdk/provider-utils';\nimport {\n DeepSeekChatModelId,\n DeepSeekChatSettings,\n} from './deepseek-chat-settings';\n\nexport interface DeepSeekProviderSettings {\n /**\nDeepSeek API key.\n*/\n apiKey?: string;\n /**\nBase URL for the API calls.\n*/\n baseURL?: string;\n /**\nCustom headers to include in the requests.\n*/\n headers?: Record<string, string>;\n /**\nCustom fetch implementation. You can use it as a middleware to intercept requests,\nor to provide a custom fetch implementation for e.g. testing.\n*/\n fetch?: FetchFunction;\n}\n\nexport interface DeepSeekProvider extends ProviderV1 {\n /**\nCreates a DeepSeek model for text generation.\n*/\n (\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ): LanguageModelV1;\n\n /**\nCreates a DeepSeek model for text generation.\n*/\n languageModel(\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ): LanguageModelV1;\n\n /**\nCreates a DeepSeek chat model for text generation.\n*/\n chat(\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ): LanguageModelV1;\n}\n\nexport function createDeepSeek(\n options: DeepSeekProviderSettings = {},\n): DeepSeekProvider {\n const baseURL = withoutTrailingSlash(\n options.baseURL ?? 'https://api.deepseek.com/v1',\n );\n const getHeaders = () => ({\n Authorization: `Bearer ${loadApiKey({\n apiKey: options.apiKey,\n environmentVariableName: 'DEEPSEEK_API_KEY',\n description: 'DeepSeek API key',\n })}`,\n ...options.headers,\n });\n\n const createLanguageModel = (\n modelId: DeepSeekChatModelId,\n settings: DeepSeekChatSettings = {},\n ) => {\n return new OpenAICompatibleChatLanguageModel(modelId, settings, {\n provider: `deepseek.chat`,\n url: ({ path }) => `${baseURL}${path}`,\n headers: getHeaders,\n fetch: options.fetch,\n defaultObjectGenerationMode: 'json',\n });\n };\n\n const provider = (\n modelId: DeepSeekChatModelId,\n settings?: DeepSeekChatSettings,\n ) => createLanguageModel(modelId, settings);\n\n provider.languageModel = createLanguageModel;\n provider.chat = createLanguageModel;\n provider.textEmbeddingModel = (modelId: string) => {\n throw new NoSuchModelError({ modelId, modelType: 'textEmbeddingModel' });\n };\n\n return provider as DeepSeekProvider;\n}\n\nexport const deepseek = createDeepSeek();\n"],"mappings":";AAAA,SAAS,yCAAyC;AAClD;AAAA,EAEE;AAAA,OAEK;AACP;AAAA,EAEE;AAAA,EACA;AAAA,OACK;AAoDA,SAAS,eACd,UAAoC,CAAC,GACnB;AAhEpB;AAiEE,QAAM,UAAU;AAAA,KACd,aAAQ,YAAR,YAAmB;AAAA,EACrB;AACA,QAAM,aAAa,OAAO;AAAA,IACxB,eAAe,UAAU,WAAW;AAAA,MAClC,QAAQ,QAAQ;AAAA,MAChB,yBAAyB;AAAA,MACzB,aAAa;AAAA,IACf,CAAC,CAAC;AAAA,IACF,GAAG,QAAQ;AAAA,EACb;AAEA,QAAM,sBAAsB,CAC1B,SACA,WAAiC,CAAC,MAC/B;AACH,WAAO,IAAI,kCAAkC,SAAS,UAAU;AAAA,MAC9D,UAAU;AAAA,MACV,KAAK,CAAC,EAAE,KAAK,MAAM,GAAG,OAAO,GAAG,IAAI;AAAA,MACpC,SAAS;AAAA,MACT,OAAO,QAAQ;AAAA,MACf,6BAA6B;AAAA,IAC/B,CAAC;AAAA,EACH;AAEA,QAAM,WAAW,CACf,SACA,aACG,oBAAoB,SAAS,QAAQ;AAE1C,WAAS,gBAAgB;AACzB,WAAS,OAAO;AAChB,WAAS,qBAAqB,CAAC,YAAoB;AACjD,UAAM,IAAI,iBAAiB,EAAE,SAAS,WAAW,qBAAqB,CAAC;AAAA,EACzE;AAEA,SAAO;AACT;AAEO,IAAM,WAAW,eAAe;","names":[]}
|
package/package.json
ADDED
|
@@ -0,0 +1,64 @@
|
|
|
1
|
+
{
|
|
2
|
+
"name": "@ai-sdk/deepseek",
|
|
3
|
+
"version": "0.0.1",
|
|
4
|
+
"license": "Apache-2.0",
|
|
5
|
+
"sideEffects": false,
|
|
6
|
+
"main": "./dist/index.js",
|
|
7
|
+
"module": "./dist/index.mjs",
|
|
8
|
+
"types": "./dist/index.d.ts",
|
|
9
|
+
"files": [
|
|
10
|
+
"dist/**/*",
|
|
11
|
+
"CHANGELOG.md"
|
|
12
|
+
],
|
|
13
|
+
"exports": {
|
|
14
|
+
"./package.json": "./package.json",
|
|
15
|
+
".": {
|
|
16
|
+
"types": "./dist/index.d.ts",
|
|
17
|
+
"import": "./dist/index.mjs",
|
|
18
|
+
"require": "./dist/index.js"
|
|
19
|
+
}
|
|
20
|
+
},
|
|
21
|
+
"dependencies": {
|
|
22
|
+
"@ai-sdk/openai-compatible": "0.0.13",
|
|
23
|
+
"@ai-sdk/provider": "1.0.3",
|
|
24
|
+
"@ai-sdk/provider-utils": "2.0.5"
|
|
25
|
+
},
|
|
26
|
+
"devDependencies": {
|
|
27
|
+
"@types/node": "^18",
|
|
28
|
+
"tsup": "^8",
|
|
29
|
+
"typescript": "5.6.3",
|
|
30
|
+
"zod": "3.23.8",
|
|
31
|
+
"@vercel/ai-tsconfig": "0.0.0"
|
|
32
|
+
},
|
|
33
|
+
"peerDependencies": {
|
|
34
|
+
"zod": "^3.0.0"
|
|
35
|
+
},
|
|
36
|
+
"engines": {
|
|
37
|
+
"node": ">=18"
|
|
38
|
+
},
|
|
39
|
+
"publishConfig": {
|
|
40
|
+
"access": "public"
|
|
41
|
+
},
|
|
42
|
+
"homepage": "https://sdk.vercel.ai/docs",
|
|
43
|
+
"repository": {
|
|
44
|
+
"type": "git",
|
|
45
|
+
"url": "git+https://github.com/vercel/ai.git"
|
|
46
|
+
},
|
|
47
|
+
"bugs": {
|
|
48
|
+
"url": "https://github.com/vercel/ai/issues"
|
|
49
|
+
},
|
|
50
|
+
"keywords": [
|
|
51
|
+
"ai"
|
|
52
|
+
],
|
|
53
|
+
"scripts": {
|
|
54
|
+
"build": "tsup",
|
|
55
|
+
"build:watch": "tsup --watch",
|
|
56
|
+
"clean": "rm -rf dist",
|
|
57
|
+
"lint": "eslint \"./**/*.ts*\"",
|
|
58
|
+
"type-check": "tsc --noEmit",
|
|
59
|
+
"prettier-check": "prettier --check \"./**/*.ts*\"",
|
|
60
|
+
"test": "pnpm test:node && pnpm test:edge",
|
|
61
|
+
"test:edge": "vitest --config vitest.edge.config.js --run",
|
|
62
|
+
"test:node": "vitest --config vitest.node.config.js --run"
|
|
63
|
+
}
|
|
64
|
+
}
|