@ai-sdk/togetherai 2.0.24 → 2.0.26
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +14 -0
- package/README.md +8 -0
- package/dist/index.js +1 -1
- package/dist/index.mjs +1 -1
- package/docs/24-togetherai.mdx +35 -14
- package/package.json +2 -2
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,19 @@
|
|
|
1
1
|
# @ai-sdk/togetherai
|
|
2
2
|
|
|
3
|
+
## 2.0.26
|
|
4
|
+
|
|
5
|
+
### Patch Changes
|
|
6
|
+
|
|
7
|
+
- 1524271: chore: add skill information to README files
|
|
8
|
+
- Updated dependencies [1524271]
|
|
9
|
+
- @ai-sdk/openai-compatible@2.0.23
|
|
10
|
+
|
|
11
|
+
## 2.0.25
|
|
12
|
+
|
|
13
|
+
### Patch Changes
|
|
14
|
+
|
|
15
|
+
- 3988c08: docs: fix incorrect and outdated provider docs
|
|
16
|
+
|
|
3
17
|
## 2.0.24
|
|
4
18
|
|
|
5
19
|
### Patch Changes
|
package/README.md
CHANGED
|
@@ -10,6 +10,14 @@ The Together.ai provider is available in the `@ai-sdk/togetherai` module. You ca
|
|
|
10
10
|
npm i @ai-sdk/togetherai
|
|
11
11
|
```
|
|
12
12
|
|
|
13
|
+
## Skill for Coding Agents
|
|
14
|
+
|
|
15
|
+
If you use coding agents such as Claude Code or Cursor, we highly recommend adding the AI SDK skill to your repository:
|
|
16
|
+
|
|
17
|
+
```shell
|
|
18
|
+
npx skills add vercel/ai
|
|
19
|
+
```
|
|
20
|
+
|
|
13
21
|
## Provider Instance
|
|
14
22
|
|
|
15
23
|
You can import the default provider instance `togetherai` from `@ai-sdk/togetherai`:
|
package/dist/index.js
CHANGED
|
@@ -272,7 +272,7 @@ var togetheraiImageProviderOptionsSchema = (0, import_provider_utils4.lazySchema
|
|
|
272
272
|
);
|
|
273
273
|
|
|
274
274
|
// src/version.ts
|
|
275
|
-
var VERSION = true ? "2.0.
|
|
275
|
+
var VERSION = true ? "2.0.26" : "0.0.0-test";
|
|
276
276
|
|
|
277
277
|
// src/togetherai-provider.ts
|
|
278
278
|
function createTogetherAI(options = {}) {
|
package/dist/index.mjs
CHANGED
|
@@ -267,7 +267,7 @@ var togetheraiImageProviderOptionsSchema = lazySchema3(
|
|
|
267
267
|
);
|
|
268
268
|
|
|
269
269
|
// src/version.ts
|
|
270
|
-
var VERSION = true ? "2.0.
|
|
270
|
+
var VERSION = true ? "2.0.26" : "0.0.0-test";
|
|
271
271
|
|
|
272
272
|
// src/togetherai-provider.ts
|
|
273
273
|
function createTogetherAI(options = {}) {
|
package/docs/24-togetherai.mdx
CHANGED
|
@@ -112,20 +112,20 @@ const { text } = await generateText({
|
|
|
112
112
|
Together.ai language models can also be used in the `streamText` function
|
|
113
113
|
(see [AI SDK Core](/docs/ai-sdk-core)).
|
|
114
114
|
|
|
115
|
-
The Together.ai provider also supports [completion models](https://docs.together.ai/docs/serverless-models#language-models) via (following the above example code) `togetherai.
|
|
115
|
+
The Together.ai provider also supports [completion models](https://docs.together.ai/docs/serverless-models#language-models) via (following the above example code) `togetherai.completionModel()` and [embedding models](https://docs.together.ai/docs/serverless-models#embedding-models) via `togetherai.embeddingModel()`.
|
|
116
116
|
|
|
117
117
|
## Model Capabilities
|
|
118
118
|
|
|
119
|
-
| Model
|
|
120
|
-
|
|
|
121
|
-
| `meta-llama/
|
|
122
|
-
| `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo`
|
|
123
|
-
| `mistralai/Mixtral-8x22B-Instruct-v0.1`
|
|
124
|
-
| `mistralai/Mistral-7B-Instruct-v0.3`
|
|
125
|
-
| `deepseek-ai/DeepSeek-V3`
|
|
126
|
-
| `google/gemma-2b-it`
|
|
127
|
-
| `Qwen/Qwen2.5-72B-Instruct-Turbo`
|
|
128
|
-
| `databricks/dbrx-instruct`
|
|
119
|
+
| Model | Image Input | Object Generation | Tool Usage | Tool Streaming |
|
|
120
|
+
| --------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- |
|
|
121
|
+
| `meta-llama/Llama-3.3-70B-Instruct-Turbo` | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> |
|
|
122
|
+
| `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | <Cross size={18} /> | <Cross size={18} /> | <Check size={18} /> | <Check size={18} /> |
|
|
123
|
+
| `mistralai/Mixtral-8x22B-Instruct-v0.1` | <Cross size={18} /> | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> |
|
|
124
|
+
| `mistralai/Mistral-7B-Instruct-v0.3` | <Cross size={18} /> | <Check size={18} /> | <Check size={18} /> | <Check size={18} /> |
|
|
125
|
+
| `deepseek-ai/DeepSeek-V3` | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> |
|
|
126
|
+
| `google/gemma-2b-it` | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> |
|
|
127
|
+
| `Qwen/Qwen2.5-72B-Instruct-Turbo` | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> |
|
|
128
|
+
| `databricks/dbrx-instruct` | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> | <Cross size={18} /> |
|
|
129
129
|
|
|
130
130
|
<Note>
|
|
131
131
|
The table above lists popular models. Please see the [Together.ai
|
|
@@ -168,7 +168,25 @@ const { images } = await generateImage({
|
|
|
168
168
|
});
|
|
169
169
|
```
|
|
170
170
|
|
|
171
|
-
|
|
171
|
+
The following provider options are available:
|
|
172
|
+
|
|
173
|
+
- **steps** _number_
|
|
174
|
+
|
|
175
|
+
Number of generation steps. Higher values can improve quality.
|
|
176
|
+
|
|
177
|
+
- **guidance** _number_
|
|
178
|
+
|
|
179
|
+
Guidance scale for image generation.
|
|
180
|
+
|
|
181
|
+
- **negative_prompt** _string_
|
|
182
|
+
|
|
183
|
+
Negative prompt to guide what to avoid.
|
|
184
|
+
|
|
185
|
+
- **disable_safety_checker** _boolean_
|
|
186
|
+
|
|
187
|
+
Disable the safety checker for image generation.
|
|
188
|
+
When true, the API will not reject images flagged as potentially NSFW.
|
|
189
|
+
Not available for Flux Schnell Free and Flux Pro models.
|
|
172
190
|
|
|
173
191
|
### Image Editing
|
|
174
192
|
|
|
@@ -251,6 +269,9 @@ Together.ai image models support various image dimensions that vary by model. Co
|
|
|
251
269
|
| `black-forest-labs/FLUX.1.1-pro` |
|
|
252
270
|
| `black-forest-labs/FLUX.1-pro` |
|
|
253
271
|
| `black-forest-labs/FLUX.1-schnell-Free` |
|
|
272
|
+
| `black-forest-labs/FLUX.1-kontext-pro` |
|
|
273
|
+
| `black-forest-labs/FLUX.1-kontext-max` |
|
|
274
|
+
| `black-forest-labs/FLUX.1-kontext-dev` |
|
|
254
275
|
|
|
255
276
|
<Note>
|
|
256
277
|
Please see the [Together.ai models
|
|
@@ -260,7 +281,7 @@ Together.ai image models support various image dimensions that vary by model. Co
|
|
|
260
281
|
|
|
261
282
|
## Embedding Models
|
|
262
283
|
|
|
263
|
-
You can create Together.ai embedding models using the `.
|
|
284
|
+
You can create Together.ai embedding models using the `.embeddingModel()` factory method.
|
|
264
285
|
For more on embedding models with the AI SDK see [embed()](/docs/reference/ai-sdk-core/embed).
|
|
265
286
|
|
|
266
287
|
```ts
|
|
@@ -268,7 +289,7 @@ import { togetherai } from '@ai-sdk/togetherai';
|
|
|
268
289
|
import { embed } from 'ai';
|
|
269
290
|
|
|
270
291
|
const { embedding } = await embed({
|
|
271
|
-
model: togetherai.
|
|
292
|
+
model: togetherai.embeddingModel('togethercomputer/m2-bert-80M-2k-retrieval'),
|
|
272
293
|
value: 'sunny day at the beach',
|
|
273
294
|
});
|
|
274
295
|
```
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@ai-sdk/togetherai",
|
|
3
|
-
"version": "2.0.
|
|
3
|
+
"version": "2.0.26",
|
|
4
4
|
"license": "Apache-2.0",
|
|
5
5
|
"sideEffects": false,
|
|
6
6
|
"main": "./dist/index.js",
|
|
@@ -29,7 +29,7 @@
|
|
|
29
29
|
}
|
|
30
30
|
},
|
|
31
31
|
"dependencies": {
|
|
32
|
-
"@ai-sdk/openai-compatible": "2.0.
|
|
32
|
+
"@ai-sdk/openai-compatible": "2.0.23",
|
|
33
33
|
"@ai-sdk/provider": "3.0.5",
|
|
34
34
|
"@ai-sdk/provider-utils": "4.0.10"
|
|
35
35
|
},
|