@gammatech/aijsx 0.5.0-dev.2024-03-19 → 0.6.1-dev.2024-04-17
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.mts +24 -6
- package/dist/index.d.ts +24 -6
- package/dist/index.js +470 -361
- package/dist/index.mjs +465 -361
- package/dist/{jsx-dev-runtime-j4JCaPsB.d.mts → jsx-dev-runtime-n4ojN2eR.d.mts} +4 -2
- package/dist/{jsx-dev-runtime-j4JCaPsB.d.ts → jsx-dev-runtime-n4ojN2eR.d.ts} +4 -2
- package/dist/jsx-dev-runtime.d.mts +1 -1
- package/dist/jsx-dev-runtime.d.ts +1 -1
- package/dist/jsx-runtime.d.mts +1 -1
- package/dist/jsx-runtime.d.ts +1 -1
- package/package.json +2 -2
package/dist/index.d.mts
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
import { R as RenderContext, L as LogImplementation, S as SpanProcessor, C as ContextValues,
|
|
2
|
-
export { I as AIElement, k as AIFragment, o as AssistantMessage, B as BoundLogger, r as ChatCompletionError, s as ChatCompletionRequestPayloads, m as ChatCompletionRole, z as CombinedLogger, y as ConsoleLogger, D as Literal, t as LogChatCompletionRequest, u as LogChatCompletionResponse, v as LogLevel, w as Logger, x as NoopLogImplementation, _ as OutputParser, M as PropsOfAIComponent, G as RenderResult, K as Renderable, p as RenderedConversationMessage, V as Span, O as SpanContext, W as SpanEvent, Q as SpanStatus, n as SystemMessage, Y as TracingContext, X as TracingContextKey, Z as TracingContextManager, U as UserMessage, H as attachedContextSymbol, q as computeUsage, j as createAIElement, l as createContext } from './jsx-dev-runtime-
|
|
1
|
+
import { R as RenderContext, L as LogImplementation, S as SpanProcessor, C as ContextValues, a as Context, A as AINode, T as Tracer, b as ReadableSpan, c as SpanExporter, d as AIComponent, e as SpanAttributes, E as EvaluatorFn, P as PromptParsed, f as Prompt, N as NotAsyncGenerator, F as FunctionChain, g as StreamChain, h as EvaluatorResult, i as ChatCompletionClientAndProvider, J as JSX } from './jsx-dev-runtime-n4ojN2eR.mjs';
|
|
2
|
+
export { I as AIElement, k as AIFragment, o as AssistantMessage, B as BoundLogger, r as ChatCompletionError, s as ChatCompletionRequestPayloads, m as ChatCompletionRole, z as CombinedLogger, y as ConsoleLogger, D as Literal, t as LogChatCompletionRequest, u as LogChatCompletionResponse, v as LogLevel, w as Logger, x as NoopLogImplementation, _ as OutputParser, M as PropsOfAIComponent, G as RenderResult, K as Renderable, p as RenderedConversationMessage, V as Span, O as SpanContext, W as SpanEvent, Q as SpanStatus, n as SystemMessage, Y as TracingContext, X as TracingContextKey, Z as TracingContextManager, U as UserMessage, H as attachedContextSymbol, q as computeUsage, j as createAIElement, l as createContext } from './jsx-dev-runtime-n4ojN2eR.mjs';
|
|
3
3
|
import { ZodObject, ZodRawShape, ZodTypeAny, ZodString, z } from 'zod';
|
|
4
4
|
import { OpenAI } from 'openai';
|
|
5
5
|
export { OpenAI as OpenAIClient } from 'openai';
|
|
@@ -17,6 +17,16 @@ type CreateRenderContextOptions = {
|
|
|
17
17
|
};
|
|
18
18
|
declare function createRenderContext({ logger, traceId, processor, contextValues, }?: CreateRenderContextOptions): RenderContext;
|
|
19
19
|
|
|
20
|
+
declare const RetryCountContext: Context<number>;
|
|
21
|
+
declare const DefaultMaxRetriesContext: Context<number>;
|
|
22
|
+
type RetryProps = {
|
|
23
|
+
shouldRetry: (error: Error) => boolean;
|
|
24
|
+
retries?: number;
|
|
25
|
+
maxRetries?: number;
|
|
26
|
+
children: AINode;
|
|
27
|
+
};
|
|
28
|
+
declare function Retry({ shouldRetry, retries, maxRetries, children }: RetryProps, ctx: RenderContext): AsyncGenerator<string, void, unknown>;
|
|
29
|
+
|
|
20
30
|
declare class AITraceAPI {
|
|
21
31
|
private static _instance?;
|
|
22
32
|
private enabled;
|
|
@@ -141,10 +151,12 @@ type OpenAIChatCompletionProps = {
|
|
|
141
151
|
model: ValidOpenAIChatModel;
|
|
142
152
|
maxTokens?: number;
|
|
143
153
|
temperature?: number;
|
|
154
|
+
stop?: string | string[];
|
|
144
155
|
responseFormat?: ChatCompletionCreateParams.ResponseFormat['type'];
|
|
156
|
+
maxRetries?: number;
|
|
145
157
|
children: AINode;
|
|
146
158
|
};
|
|
147
|
-
declare function OpenAIChatCompletion(props: OpenAIChatCompletionProps,
|
|
159
|
+
declare function OpenAIChatCompletion(props: OpenAIChatCompletionProps, ctx: RenderContext): AINode;
|
|
148
160
|
|
|
149
161
|
type ValidOpenAIVisionModel = 'gpt-4-vision-preview';
|
|
150
162
|
declare const ContentTypeImage: (_props: {
|
|
@@ -158,10 +170,13 @@ declare const ContentTypeImage: (_props: {
|
|
|
158
170
|
type OpenAIVisionChatCompletionProps = {
|
|
159
171
|
model?: ValidOpenAIVisionModel;
|
|
160
172
|
maxTokens?: number;
|
|
173
|
+
stop?: string | string[];
|
|
161
174
|
temperature?: number;
|
|
175
|
+
maxRetries?: number;
|
|
162
176
|
children: AINode;
|
|
163
177
|
};
|
|
164
|
-
declare function OpenAIVisionChatCompletion(props: OpenAIVisionChatCompletionProps,
|
|
178
|
+
declare function OpenAIVisionChatCompletion(props: OpenAIVisionChatCompletionProps, ctx: RenderContext): AINode;
|
|
179
|
+
declare function OpenAIVisionChatCompletionInner(props: OpenAIVisionChatCompletionProps, { logger, render, tracer, getContext }: RenderContext): AsyncGenerator<string, void, unknown>;
|
|
165
180
|
|
|
166
181
|
declare const tokenizer: {
|
|
167
182
|
encode: (text: string) => number[];
|
|
@@ -187,8 +202,11 @@ type AnthropicChatCompletionProps = {
|
|
|
187
202
|
model: ValidAnthropicChatModel;
|
|
188
203
|
maxTokens?: number;
|
|
189
204
|
temperature?: number;
|
|
205
|
+
stop?: string | string[];
|
|
206
|
+
maxRetries?: number;
|
|
190
207
|
children: AINode;
|
|
191
208
|
};
|
|
209
|
+
declare function AnthropicChatCompletion(props: AnthropicChatCompletionProps, ctx: RenderContext): JSX.Element;
|
|
192
210
|
/**
|
|
193
211
|
* An AI.JSX component that invokes an Anthropic Large Language Model.
|
|
194
212
|
* @param children The children to render.
|
|
@@ -196,7 +214,7 @@ type AnthropicChatCompletionProps = {
|
|
|
196
214
|
* @param completionModel The completion model to use.
|
|
197
215
|
* @param client The Anthropic client.
|
|
198
216
|
*/
|
|
199
|
-
declare function
|
|
217
|
+
declare function AnthropicChatCompletionInner(props: AnthropicChatCompletionProps, { render, logger, tracer, getContext }: RenderContext): AsyncGenerator<string, void, unknown>;
|
|
200
218
|
|
|
201
219
|
type ClaudeImageBlockBase64 = {
|
|
202
220
|
data: string;
|
|
@@ -208,4 +226,4 @@ type ClaudeImageBlockUrl = {
|
|
|
208
226
|
type ClaudeImageBlockProps = ClaudeImageBlockBase64 | ClaudeImageBlockUrl;
|
|
209
227
|
declare const ClaudeImageBlock: (props: ClaudeImageBlockProps) => Promise<JSX.Element>;
|
|
210
228
|
|
|
211
|
-
export { AIComponent, AINode, AISpanProcessor, AnthropicChatCompletion, type AnthropicChatCompletionRequest, AnthropicClientContext, ChatCompletionClientAndProvider, ClaudeImageBlock, ContentTypeImage, Context, type CostFn, EnrichingSpanProcessor, EvaluatorFn, EvaluatorResult, FunctionChain, LogImplementation, NotAsyncGenerator, OpenAIChatCompletion, type OpenAIChatCompletionRequest, type OpenAIChatMessage, OpenAIClientContext, OpenAIVisionChatCompletion, ParseVariablesError, Prompt, PromptInvalidOutputError, PromptParsed, ReadableSpan, RenderContext, SpanAttributes, SpanExporter, SpanProcessor, StreamChain, Trace, Tracer, type ValidAnthropicChatModel, type ValidOpenAIChatModel, type ValidOpenAIVisionModel, createFunctionChain, createPrompt, createRenderContext, createStreamChain, evaluatePrompt, tokenCountForOpenAIMessage, tokenCountForOpenAIVisionMessage, tokenLimitForChatModel, tokenizer, tracing };
|
|
229
|
+
export { AIComponent, AINode, AISpanProcessor, AnthropicChatCompletion, AnthropicChatCompletionInner, type AnthropicChatCompletionRequest, AnthropicClientContext, ChatCompletionClientAndProvider, ClaudeImageBlock, ContentTypeImage, Context, type CostFn, DefaultMaxRetriesContext, EnrichingSpanProcessor, EvaluatorFn, EvaluatorResult, FunctionChain, LogImplementation, NotAsyncGenerator, OpenAIChatCompletion, type OpenAIChatCompletionRequest, type OpenAIChatMessage, OpenAIClientContext, OpenAIVisionChatCompletion, OpenAIVisionChatCompletionInner, ParseVariablesError, Prompt, PromptInvalidOutputError, PromptParsed, ReadableSpan, RenderContext, Retry, RetryCountContext, SpanAttributes, SpanExporter, SpanProcessor, StreamChain, Trace, Tracer, type ValidAnthropicChatModel, type ValidOpenAIChatModel, type ValidOpenAIVisionModel, createFunctionChain, createPrompt, createRenderContext, createStreamChain, evaluatePrompt, tokenCountForOpenAIMessage, tokenCountForOpenAIVisionMessage, tokenLimitForChatModel, tokenizer, tracing };
|
package/dist/index.d.ts
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
import { R as RenderContext, L as LogImplementation, S as SpanProcessor, C as ContextValues,
|
|
2
|
-
export { I as AIElement, k as AIFragment, o as AssistantMessage, B as BoundLogger, r as ChatCompletionError, s as ChatCompletionRequestPayloads, m as ChatCompletionRole, z as CombinedLogger, y as ConsoleLogger, D as Literal, t as LogChatCompletionRequest, u as LogChatCompletionResponse, v as LogLevel, w as Logger, x as NoopLogImplementation, _ as OutputParser, M as PropsOfAIComponent, G as RenderResult, K as Renderable, p as RenderedConversationMessage, V as Span, O as SpanContext, W as SpanEvent, Q as SpanStatus, n as SystemMessage, Y as TracingContext, X as TracingContextKey, Z as TracingContextManager, U as UserMessage, H as attachedContextSymbol, q as computeUsage, j as createAIElement, l as createContext } from './jsx-dev-runtime-
|
|
1
|
+
import { R as RenderContext, L as LogImplementation, S as SpanProcessor, C as ContextValues, a as Context, A as AINode, T as Tracer, b as ReadableSpan, c as SpanExporter, d as AIComponent, e as SpanAttributes, E as EvaluatorFn, P as PromptParsed, f as Prompt, N as NotAsyncGenerator, F as FunctionChain, g as StreamChain, h as EvaluatorResult, i as ChatCompletionClientAndProvider, J as JSX } from './jsx-dev-runtime-n4ojN2eR.js';
|
|
2
|
+
export { I as AIElement, k as AIFragment, o as AssistantMessage, B as BoundLogger, r as ChatCompletionError, s as ChatCompletionRequestPayloads, m as ChatCompletionRole, z as CombinedLogger, y as ConsoleLogger, D as Literal, t as LogChatCompletionRequest, u as LogChatCompletionResponse, v as LogLevel, w as Logger, x as NoopLogImplementation, _ as OutputParser, M as PropsOfAIComponent, G as RenderResult, K as Renderable, p as RenderedConversationMessage, V as Span, O as SpanContext, W as SpanEvent, Q as SpanStatus, n as SystemMessage, Y as TracingContext, X as TracingContextKey, Z as TracingContextManager, U as UserMessage, H as attachedContextSymbol, q as computeUsage, j as createAIElement, l as createContext } from './jsx-dev-runtime-n4ojN2eR.js';
|
|
3
3
|
import { ZodObject, ZodRawShape, ZodTypeAny, ZodString, z } from 'zod';
|
|
4
4
|
import { OpenAI } from 'openai';
|
|
5
5
|
export { OpenAI as OpenAIClient } from 'openai';
|
|
@@ -17,6 +17,16 @@ type CreateRenderContextOptions = {
|
|
|
17
17
|
};
|
|
18
18
|
declare function createRenderContext({ logger, traceId, processor, contextValues, }?: CreateRenderContextOptions): RenderContext;
|
|
19
19
|
|
|
20
|
+
declare const RetryCountContext: Context<number>;
|
|
21
|
+
declare const DefaultMaxRetriesContext: Context<number>;
|
|
22
|
+
type RetryProps = {
|
|
23
|
+
shouldRetry: (error: Error) => boolean;
|
|
24
|
+
retries?: number;
|
|
25
|
+
maxRetries?: number;
|
|
26
|
+
children: AINode;
|
|
27
|
+
};
|
|
28
|
+
declare function Retry({ shouldRetry, retries, maxRetries, children }: RetryProps, ctx: RenderContext): AsyncGenerator<string, void, unknown>;
|
|
29
|
+
|
|
20
30
|
declare class AITraceAPI {
|
|
21
31
|
private static _instance?;
|
|
22
32
|
private enabled;
|
|
@@ -141,10 +151,12 @@ type OpenAIChatCompletionProps = {
|
|
|
141
151
|
model: ValidOpenAIChatModel;
|
|
142
152
|
maxTokens?: number;
|
|
143
153
|
temperature?: number;
|
|
154
|
+
stop?: string | string[];
|
|
144
155
|
responseFormat?: ChatCompletionCreateParams.ResponseFormat['type'];
|
|
156
|
+
maxRetries?: number;
|
|
145
157
|
children: AINode;
|
|
146
158
|
};
|
|
147
|
-
declare function OpenAIChatCompletion(props: OpenAIChatCompletionProps,
|
|
159
|
+
declare function OpenAIChatCompletion(props: OpenAIChatCompletionProps, ctx: RenderContext): AINode;
|
|
148
160
|
|
|
149
161
|
type ValidOpenAIVisionModel = 'gpt-4-vision-preview';
|
|
150
162
|
declare const ContentTypeImage: (_props: {
|
|
@@ -158,10 +170,13 @@ declare const ContentTypeImage: (_props: {
|
|
|
158
170
|
type OpenAIVisionChatCompletionProps = {
|
|
159
171
|
model?: ValidOpenAIVisionModel;
|
|
160
172
|
maxTokens?: number;
|
|
173
|
+
stop?: string | string[];
|
|
161
174
|
temperature?: number;
|
|
175
|
+
maxRetries?: number;
|
|
162
176
|
children: AINode;
|
|
163
177
|
};
|
|
164
|
-
declare function OpenAIVisionChatCompletion(props: OpenAIVisionChatCompletionProps,
|
|
178
|
+
declare function OpenAIVisionChatCompletion(props: OpenAIVisionChatCompletionProps, ctx: RenderContext): AINode;
|
|
179
|
+
declare function OpenAIVisionChatCompletionInner(props: OpenAIVisionChatCompletionProps, { logger, render, tracer, getContext }: RenderContext): AsyncGenerator<string, void, unknown>;
|
|
165
180
|
|
|
166
181
|
declare const tokenizer: {
|
|
167
182
|
encode: (text: string) => number[];
|
|
@@ -187,8 +202,11 @@ type AnthropicChatCompletionProps = {
|
|
|
187
202
|
model: ValidAnthropicChatModel;
|
|
188
203
|
maxTokens?: number;
|
|
189
204
|
temperature?: number;
|
|
205
|
+
stop?: string | string[];
|
|
206
|
+
maxRetries?: number;
|
|
190
207
|
children: AINode;
|
|
191
208
|
};
|
|
209
|
+
declare function AnthropicChatCompletion(props: AnthropicChatCompletionProps, ctx: RenderContext): JSX.Element;
|
|
192
210
|
/**
|
|
193
211
|
* An AI.JSX component that invokes an Anthropic Large Language Model.
|
|
194
212
|
* @param children The children to render.
|
|
@@ -196,7 +214,7 @@ type AnthropicChatCompletionProps = {
|
|
|
196
214
|
* @param completionModel The completion model to use.
|
|
197
215
|
* @param client The Anthropic client.
|
|
198
216
|
*/
|
|
199
|
-
declare function
|
|
217
|
+
declare function AnthropicChatCompletionInner(props: AnthropicChatCompletionProps, { render, logger, tracer, getContext }: RenderContext): AsyncGenerator<string, void, unknown>;
|
|
200
218
|
|
|
201
219
|
type ClaudeImageBlockBase64 = {
|
|
202
220
|
data: string;
|
|
@@ -208,4 +226,4 @@ type ClaudeImageBlockUrl = {
|
|
|
208
226
|
type ClaudeImageBlockProps = ClaudeImageBlockBase64 | ClaudeImageBlockUrl;
|
|
209
227
|
declare const ClaudeImageBlock: (props: ClaudeImageBlockProps) => Promise<JSX.Element>;
|
|
210
228
|
|
|
211
|
-
export { AIComponent, AINode, AISpanProcessor, AnthropicChatCompletion, type AnthropicChatCompletionRequest, AnthropicClientContext, ChatCompletionClientAndProvider, ClaudeImageBlock, ContentTypeImage, Context, type CostFn, EnrichingSpanProcessor, EvaluatorFn, EvaluatorResult, FunctionChain, LogImplementation, NotAsyncGenerator, OpenAIChatCompletion, type OpenAIChatCompletionRequest, type OpenAIChatMessage, OpenAIClientContext, OpenAIVisionChatCompletion, ParseVariablesError, Prompt, PromptInvalidOutputError, PromptParsed, ReadableSpan, RenderContext, SpanAttributes, SpanExporter, SpanProcessor, StreamChain, Trace, Tracer, type ValidAnthropicChatModel, type ValidOpenAIChatModel, type ValidOpenAIVisionModel, createFunctionChain, createPrompt, createRenderContext, createStreamChain, evaluatePrompt, tokenCountForOpenAIMessage, tokenCountForOpenAIVisionMessage, tokenLimitForChatModel, tokenizer, tracing };
|
|
229
|
+
export { AIComponent, AINode, AISpanProcessor, AnthropicChatCompletion, AnthropicChatCompletionInner, type AnthropicChatCompletionRequest, AnthropicClientContext, ChatCompletionClientAndProvider, ClaudeImageBlock, ContentTypeImage, Context, type CostFn, DefaultMaxRetriesContext, EnrichingSpanProcessor, EvaluatorFn, EvaluatorResult, FunctionChain, LogImplementation, NotAsyncGenerator, OpenAIChatCompletion, type OpenAIChatCompletionRequest, type OpenAIChatMessage, OpenAIClientContext, OpenAIVisionChatCompletion, OpenAIVisionChatCompletionInner, ParseVariablesError, Prompt, PromptInvalidOutputError, PromptParsed, ReadableSpan, RenderContext, Retry, RetryCountContext, SpanAttributes, SpanExporter, SpanProcessor, StreamChain, Trace, Tracer, type ValidAnthropicChatModel, type ValidOpenAIChatModel, type ValidOpenAIVisionModel, createFunctionChain, createPrompt, createRenderContext, createStreamChain, evaluatePrompt, tokenCountForOpenAIMessage, tokenCountForOpenAIVisionMessage, tokenLimitForChatModel, tokenizer, tracing };
|