opik-vercel 1.10.13 → 1.10.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -5
- package/package.json +1 -1
package/README.md
CHANGED
|
@@ -67,7 +67,7 @@ sdk.start();
|
|
|
67
67
|
|
|
68
68
|
async function main() {
|
|
69
69
|
const result = await generateText({
|
|
70
|
-
model: openai("gpt-
|
|
70
|
+
model: openai("gpt-5-nano"),
|
|
71
71
|
maxTokens: 50,
|
|
72
72
|
prompt: "What is love?",
|
|
73
73
|
experimental_telemetry: OpikExporter.getSettings({
|
|
@@ -107,7 +107,7 @@ import { openai } from "@ai-sdk/openai";
|
|
|
107
107
|
import { generateText } from "ai";
|
|
108
108
|
|
|
109
109
|
const result = await generateText({
|
|
110
|
-
model: openai("gpt-
|
|
110
|
+
model: openai("gpt-5-nano"),
|
|
111
111
|
prompt: "What is love?",
|
|
112
112
|
experimental_telemetry: { isEnabled: true },
|
|
113
113
|
});
|
|
@@ -122,7 +122,7 @@ You can add custom tags and metadata to all traces generated by the OpikExporter
|
|
|
122
122
|
```typescript
|
|
123
123
|
const exporter = new OpikExporter({
|
|
124
124
|
// Optional: add custom tags to all traces
|
|
125
|
-
tags: ["production", "gpt-
|
|
125
|
+
tags: ["production", "gpt-5-nano"],
|
|
126
126
|
// Optional: add custom metadata to all traces
|
|
127
127
|
metadata: {
|
|
128
128
|
environment: "production",
|
|
@@ -142,7 +142,7 @@ Use `OpikExporter.getSettings()` to configure telemetry for individual AI SDK ca
|
|
|
142
142
|
|
|
143
143
|
```typescript
|
|
144
144
|
const result = await generateText({
|
|
145
|
-
model: openai("gpt-
|
|
145
|
+
model: openai("gpt-5-nano"),
|
|
146
146
|
prompt: "Tell a joke",
|
|
147
147
|
experimental_telemetry: OpikExporter.getSettings({
|
|
148
148
|
name: "custom-trace-name",
|
|
@@ -158,7 +158,7 @@ Or use the basic telemetry settings:
|
|
|
158
158
|
|
|
159
159
|
```typescript
|
|
160
160
|
const result = await generateText({
|
|
161
|
-
model: openai("gpt-
|
|
161
|
+
model: openai("gpt-5-nano"),
|
|
162
162
|
prompt: "Tell a joke",
|
|
163
163
|
experimental_telemetry: { isEnabled: true },
|
|
164
164
|
});
|