@langchain/angular 0.3.0 → 0.3.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/context.cjs +6 -3
- package/dist/context.cjs.map +1 -1
- package/dist/context.d.cts +54 -25
- package/dist/context.d.cts.map +1 -1
- package/dist/context.d.ts +50 -21
- package/dist/context.d.ts.map +1 -1
- package/dist/context.js +6 -3
- package/dist/context.js.map +1 -1
- package/dist/index.cjs +3 -156
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +7 -90
- package/dist/index.d.cts.map +1 -1
- package/dist/index.d.ts +6 -89
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +3 -156
- package/dist/index.js.map +1 -1
- package/dist/stream-service-instance.d.cts +70 -0
- package/dist/stream-service-instance.d.cts.map +1 -0
- package/dist/stream-service-instance.d.ts +70 -0
- package/dist/stream-service-instance.d.ts.map +1 -0
- package/dist/stream.custom.cjs +7 -5
- package/dist/stream.custom.cjs.map +1 -1
- package/dist/stream.custom.d.cts +5 -5
- package/dist/stream.custom.d.cts.map +1 -1
- package/dist/stream.custom.d.ts +2 -2
- package/dist/stream.custom.d.ts.map +1 -1
- package/dist/stream.custom.js +7 -5
- package/dist/stream.custom.js.map +1 -1
- package/dist/stream.lgp.cjs +163 -0
- package/dist/stream.lgp.cjs.map +1 -0
- package/dist/stream.lgp.d.cts +57 -0
- package/dist/stream.lgp.d.cts.map +1 -0
- package/dist/stream.lgp.d.ts +57 -0
- package/dist/stream.lgp.d.ts.map +1 -0
- package/dist/stream.lgp.js +163 -0
- package/dist/stream.lgp.js.map +1 -0
- package/package.json +3 -3
- package/dist/subagent-types.d.cts +0 -15
- package/dist/subagent-types.d.cts.map +0 -1
- package/dist/subagent-types.d.ts +0 -15
- package/dist/subagent-types.d.ts.map +0 -1
package/dist/context.cjs
CHANGED
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
const
|
|
1
|
+
const require_stream_custom = require("./stream.custom.cjs");
|
|
2
|
+
const require_stream_lgp = require("./stream.lgp.cjs");
|
|
2
3
|
let _angular_core = require("@angular/core");
|
|
3
4
|
//#region src/context.ts
|
|
4
5
|
/**
|
|
@@ -76,12 +77,14 @@ function provideStream(options) {
|
|
|
76
77
|
provide: STREAM_INSTANCE,
|
|
77
78
|
useFactory: () => {
|
|
78
79
|
const defaults = (0, _angular_core.inject)(STREAM_DEFAULTS, { optional: true });
|
|
79
|
-
|
|
80
|
+
const merged = {
|
|
80
81
|
...defaults ?? {},
|
|
81
82
|
...options,
|
|
82
83
|
apiUrl: options.apiUrl ?? defaults?.apiUrl,
|
|
83
84
|
client: options.client ?? defaults?.client
|
|
84
|
-
}
|
|
85
|
+
};
|
|
86
|
+
if ("transport" in merged) return require_stream_custom.injectStreamCustom(merged);
|
|
87
|
+
return require_stream_lgp.useStreamLGP(merged);
|
|
85
88
|
}
|
|
86
89
|
};
|
|
87
90
|
}
|
package/dist/context.cjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"context.cjs","names":["InjectionToken","
|
|
1
|
+
{"version":3,"file":"context.cjs","names":["InjectionToken","injectStreamCustom","useStreamLGP"],"sources":["../src/context.ts"],"sourcesContent":["import {\n InjectionToken,\n inject as angularInject,\n type EnvironmentProviders,\n makeEnvironmentProviders,\n} from \"@angular/core\";\nimport type { BagTemplate } from \"@langchain/langgraph-sdk\";\nimport { Client } from \"@langchain/langgraph-sdk\";\nimport type {\n ResolveStreamOptions,\n InferBag,\n InferStateType,\n UseStreamCustomOptions,\n} from \"@langchain/langgraph-sdk/ui\";\nimport type { StreamServiceInstance } from \"./stream-service-instance.js\";\nimport { useStreamLGP } from \"./stream.lgp.js\";\nimport { injectStreamCustom } from \"./stream.custom.js\";\n\n/**\n * Configuration defaults for `useStream` and `injectStream` calls.\n */\nexport interface StreamDefaults {\n /** Base URL of the LangGraph API. */\n apiUrl?: string;\n /** API key for authenticating with the LangGraph API. */\n apiKey?: string;\n /** Pre-configured Client instance. */\n client?: Client;\n}\n\n/**\n * Injection token for stream default configuration.\n * Provide via `provideStreamDefaults()` in your application config.\n */\nexport const STREAM_DEFAULTS = new InjectionToken<StreamDefaults>(\n \"LANGCHAIN_STREAM_DEFAULTS\",\n);\n\n/**\n * Injection token for a shared stream instance.\n * Provide via `provideStream()` at the component level.\n */\nexport const STREAM_INSTANCE = new InjectionToken<StreamServiceInstance>(\n \"LANGCHAIN_STREAM_INSTANCE\",\n);\n\n/**\n * Provides default LangGraph configuration at the application level.\n *\n * Use this in your application's `providers` array to set defaults like\n * `apiUrl` that will be used by all `useStream` and `injectStream` calls.\n *\n * @example\n * ```typescript\n * // app.config.ts\n * import { ApplicationConfig } from \"@angular/core\";\n * import { provideStreamDefaults } from \"@langchain/angular\";\n *\n * export const appConfig: ApplicationConfig = {\n * providers: [\n * provideStreamDefaults({\n * apiUrl: \"http://localhost:2024\",\n * }),\n * ],\n * };\n * ```\n */\nexport function provideStreamDefaults(\n defaults: StreamDefaults,\n): EnvironmentProviders {\n return makeEnvironmentProviders([\n { provide: STREAM_DEFAULTS, useValue: defaults },\n ]);\n}\n\n/**\n * Creates a provider for a shared `useStream` instance at the component level.\n *\n * Add the returned provider to a component's `providers` array so that all\n * child components can access the same stream via `injectStream()`.\n *\n * @example\n * ```typescript\n * import { Component } from \"@angular/core\";\n * import { provideStream, injectStream } from \"@langchain/angular\";\n *\n * @Component({\n * providers: [provideStream({ assistantId: \"agent\" })],\n * template: `\n * <app-message-list />\n * <app-message-input />\n * `,\n * })\n * export class ChatContainer {}\n *\n * // In child components:\n * @Component({\n * template: `\n * @for (msg of stream.messages(); track msg.id) {\n * <div>{{ msg.content }}</div>\n * }\n * `,\n * })\n * export class MessageListComponent {\n * stream = injectStream();\n * }\n * ```\n */\nexport function provideStream<\n T = Record<string, unknown>,\n Bag extends BagTemplate = BagTemplate,\n>(\n options:\n | ResolveStreamOptions<T, InferBag<T, Bag>>\n | UseStreamCustomOptions<InferStateType<T>, InferBag<T, Bag>>,\n) {\n return {\n provide: STREAM_INSTANCE,\n useFactory: () => {\n const defaults = angularInject(STREAM_DEFAULTS, { optional: true });\n const merged = {\n ...(defaults ?? {}),\n ...options,\n apiUrl: (options as Record<string, unknown>).apiUrl ?? defaults?.apiUrl,\n client: (options as Record<string, unknown>).client ?? defaults?.client,\n };\n if (\"transport\" in merged) {\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n return injectStreamCustom(merged as any);\n }\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n return useStreamLGP(merged as any);\n },\n };\n}\n"],"mappings":";;;;;;;;AAkCA,MAAa,kBAAkB,IAAIA,cAAAA,eACjC,4BACD;;;;;AAMD,MAAa,kBAAkB,IAAIA,cAAAA,eACjC,4BACD;;;;;;;;;;;;;;;;;;;;;;AAuBD,SAAgB,sBACd,UACsB;AACtB,SAAA,GAAA,cAAA,0BAAgC,CAC9B;EAAE,SAAS;EAAiB,UAAU;EAAU,CACjD,CAAC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAoCJ,SAAgB,cAId,SAGA;AACA,QAAO;EACL,SAAS;EACT,kBAAkB;GAChB,MAAM,YAAA,GAAA,cAAA,QAAyB,iBAAiB,EAAE,UAAU,MAAM,CAAC;GACnE,MAAM,SAAS;IACb,GAAI,YAAY,EAAE;IAClB,GAAG;IACH,QAAS,QAAoC,UAAU,UAAU;IACjE,QAAS,QAAoC,UAAU,UAAU;IAClE;AACD,OAAI,eAAe,OAEjB,QAAOC,sBAAAA,mBAAmB,OAAc;AAG1C,UAAOC,mBAAAA,aAAa,OAAc;;EAErC"}
|
package/dist/context.d.cts
CHANGED
|
@@ -1,11 +1,12 @@
|
|
|
1
|
-
import { BaseMessage as BaseMessage$1 } from "./node_modules/.pnpm/langchain@1.2.30_@langchain_core@1.1.31_@opentelemetry_api@1.9.0_openai@6.27.0_ws@8.19._0520fb05d9e85da5f9e061dfe28cdbc8/node_modules/langchain/dist/index.cjs";
|
|
2
|
-
import
|
|
3
|
-
import
|
|
1
|
+
import { BaseMessage as BaseMessage$1, BaseMessageChunk } from "./node_modules/.pnpm/langchain@1.2.30_@langchain_core@1.1.31_@opentelemetry_api@1.9.0_openai@6.27.0_ws@8.19._0520fb05d9e85da5f9e061dfe28cdbc8/node_modules/langchain/dist/index.cjs";
|
|
2
|
+
import { StreamServiceInstance } from "./stream-service-instance.cjs";
|
|
3
|
+
import * as _angular_core0 from "@angular/core";
|
|
4
|
+
import { EnvironmentProviders, InjectionToken } from "@angular/core";
|
|
4
5
|
import * as _langchain_core_messages0 from "@langchain/core/messages";
|
|
5
6
|
import * as _langchain_langgraph_sdk_ui0 from "@langchain/langgraph-sdk/ui";
|
|
6
7
|
import { InferBag, InferStateType, ResolveStreamOptions, UseStreamCustomOptions } from "@langchain/langgraph-sdk/ui";
|
|
7
|
-
import * as
|
|
8
|
-
import {
|
|
8
|
+
import * as _langchain_langgraph_sdk0 from "@langchain/langgraph-sdk";
|
|
9
|
+
import { BagTemplate, Client } from "@langchain/langgraph-sdk";
|
|
9
10
|
|
|
10
11
|
//#region src/context.d.ts
|
|
11
12
|
/**
|
|
@@ -28,7 +29,7 @@ declare const STREAM_DEFAULTS: InjectionToken<StreamDefaults>;
|
|
|
28
29
|
* Injection token for a shared stream instance.
|
|
29
30
|
* Provide via `provideStream()` at the component level.
|
|
30
31
|
*/
|
|
31
|
-
declare const STREAM_INSTANCE: InjectionToken<
|
|
32
|
+
declare const STREAM_INSTANCE: InjectionToken<StreamServiceInstance<Record<string, unknown>, BagTemplate>>;
|
|
32
33
|
/**
|
|
33
34
|
* Provides default LangGraph configuration at the application level.
|
|
34
35
|
*
|
|
@@ -85,22 +86,27 @@ declare function provideStreamDefaults(defaults: StreamDefaults): EnvironmentPro
|
|
|
85
86
|
* ```
|
|
86
87
|
*/
|
|
87
88
|
declare function provideStream<T = Record<string, unknown>, Bag extends BagTemplate = BagTemplate>(options: ResolveStreamOptions<T, InferBag<T, Bag>> | UseStreamCustomOptions<InferStateType<T>, InferBag<T, Bag>>): {
|
|
88
|
-
provide: InjectionToken<
|
|
89
|
+
provide: InjectionToken<StreamServiceInstance<Record<string, unknown>, BagTemplate>>;
|
|
89
90
|
useFactory: () => {
|
|
91
|
+
assistantId: string;
|
|
92
|
+
client: Client<_langchain_langgraph_sdk0.DefaultValues, _langchain_langgraph_sdk0.DefaultValues, unknown>;
|
|
90
93
|
values: _angular_core0.Signal<Record<string, unknown>>;
|
|
91
94
|
error: _angular_core0.Signal<unknown>;
|
|
92
95
|
isLoading: _angular_core0.WritableSignal<boolean>;
|
|
93
|
-
|
|
96
|
+
branch: _angular_core0.WritableSignal<string>;
|
|
97
|
+
setBranch(value: string): void;
|
|
98
|
+
messages: _angular_core0.Signal<(BaseMessage$1<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType> | BaseMessageChunk<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType>)[]>;
|
|
99
|
+
toolCalls: _angular_core0.Signal<_langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[]>;
|
|
100
|
+
getToolCalls(message: _langchain_langgraph_sdk0.Message): _langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[];
|
|
94
101
|
interrupt: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown> | undefined>;
|
|
95
102
|
interrupts: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown>[]>;
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
joinStream: (runId: string, lastEventId?: string | undefined, options?: {
|
|
103
|
+
history: _angular_core0.Signal<_langchain_langgraph_sdk0.ThreadState<any>[]>;
|
|
104
|
+
isThreadLoading: _angular_core0.Signal<boolean>;
|
|
105
|
+
experimental_branchTree: _angular_core0.Signal<_langchain_langgraph_sdk_ui0.Sequence<any>>;
|
|
106
|
+
getMessagesMetadata(message: _langchain_langgraph_sdk0.Message, index?: number | undefined): _langchain_langgraph_sdk_ui0.MessageMetadata<Record<string, unknown>> | undefined;
|
|
107
|
+
submit: (values: Record<string, unknown>, submitOptions?: _langchain_langgraph_sdk_ui0.SubmitOptions<Record<string, unknown>, Record<string, unknown>> | undefined) => Promise<void | Promise<void>>;
|
|
108
|
+
stop: () => void;
|
|
109
|
+
joinStream: (runId: string, lastEventId?: string | undefined, joinOptions?: {
|
|
104
110
|
streamMode?: _langchain_langgraph_sdk0.StreamMode[] | _langchain_langgraph_sdk0.StreamMode | undefined;
|
|
105
111
|
filter?: ((event: {
|
|
106
112
|
id?: string | undefined;
|
|
@@ -108,21 +114,44 @@ declare function provideStream<T = Record<string, unknown>, Bag extends BagTempl
|
|
|
108
114
|
data: unknown;
|
|
109
115
|
}) => boolean) | undefined;
|
|
110
116
|
} | undefined) => Promise<void>;
|
|
111
|
-
switchThread: (newThreadId: string | null) => void;
|
|
112
117
|
queue: {
|
|
113
118
|
entries: _angular_core0.WritableSignal<readonly _langchain_langgraph_sdk_ui0.QueueEntry<Record<string, unknown>, _langchain_langgraph_sdk_ui0.SubmitOptions<Record<string, unknown>, Record<string, unknown>>>[]>;
|
|
114
119
|
size: _angular_core0.WritableSignal<number>;
|
|
115
120
|
cancel: (id: string) => Promise<boolean>;
|
|
116
121
|
clear: () => Promise<void>;
|
|
117
122
|
};
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
|
|
123
|
+
switchThread(newThreadId: string | null): void;
|
|
124
|
+
subagents: _angular_core0.Signal<ReadonlyMap<string, _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>>>;
|
|
125
|
+
activeSubagents: _angular_core0.Signal<readonly _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[]>;
|
|
126
|
+
getSubagent(toolCallId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string> | undefined;
|
|
127
|
+
getSubagentsByType(type: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
128
|
+
getSubagentsByMessage(messageId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
129
|
+
} | {
|
|
130
|
+
values: _angular_core0.Signal<Record<string, unknown>>;
|
|
131
|
+
error: _angular_core0.Signal<unknown>;
|
|
132
|
+
isLoading: _angular_core0.WritableSignal<boolean>;
|
|
133
|
+
stop: () => void;
|
|
134
|
+
submit(values: Partial<Record<string, unknown>> | null | undefined, submitOptions?: _langchain_langgraph_sdk_ui0.CustomSubmitOptions<Record<string, unknown>, Record<string, unknown>> | undefined): Promise<void>;
|
|
135
|
+
switchThread(newThreadId: string | null): void;
|
|
136
|
+
branch: _angular_core0.WritableSignal<string>;
|
|
137
|
+
setBranch(value: string): void;
|
|
138
|
+
getMessagesMetadata(message: _langchain_langgraph_sdk0.Message<_langchain_langgraph_sdk0.DefaultToolCall>, index?: number | undefined): _langchain_langgraph_sdk_ui0.MessageMetadata<Record<string, unknown>> | undefined;
|
|
139
|
+
queue: {
|
|
140
|
+
entries: _angular_core0.WritableSignal<never[]>;
|
|
141
|
+
size: _angular_core0.WritableSignal<number>;
|
|
142
|
+
cancel(): Promise<boolean>;
|
|
143
|
+
clear(): Promise<void>;
|
|
144
|
+
};
|
|
145
|
+
interrupts: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown>[]>;
|
|
146
|
+
interrupt: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown> | undefined>;
|
|
147
|
+
messages: _angular_core0.Signal<(BaseMessage$1<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType> | BaseMessageChunk<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType>)[]>;
|
|
148
|
+
toolCalls: _angular_core0.Signal<_langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[]>;
|
|
149
|
+
getToolCalls(message: _langchain_langgraph_sdk0.Message<_langchain_langgraph_sdk0.DefaultToolCall>): _langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[];
|
|
150
|
+
subagents: _angular_core0.Signal<ReadonlyMap<string, _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>>>;
|
|
151
|
+
activeSubagents: _angular_core0.Signal<readonly _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[]>;
|
|
152
|
+
getSubagent(toolCallId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string> | undefined;
|
|
153
|
+
getSubagentsByType(type: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
154
|
+
getSubagentsByMessage(messageId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
126
155
|
};
|
|
127
156
|
};
|
|
128
157
|
//#endregion
|
package/dist/context.d.cts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"context.d.cts","names":[],"sources":["../src/context.ts"],"mappings":"
|
|
1
|
+
{"version":3,"file":"context.d.cts","names":[],"sources":["../src/context.ts"],"mappings":";;;;;;;;;;;;;;UAqBiB,cAAA;;EAEf,MAAA;EAFe;EAIf,MAAA;;EAEA,MAAA,GAAS,MAAA;AAAA;;;;;cAOE,eAAA,EAAe,cAAA,CAAA,cAAA;AAA5B;;;;AAAA,cAQa,eAAA,EAAe,cAAA,CAAA,qBAAA,CAAA,MAAA,mBAAA,WAAA;AAA5B;;;;;;;;;;;;;;;AAyBA;;;;;;AAzBA,iBAyBgB,qBAAA,CACd,QAAA,EAAU,cAAA,GACT,oBAAA;;;AAuCH;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;iBAAgB,aAAA,KACV,MAAA,+BACQ,WAAA,GAAc,WAAA,CAAA,CAE1B,OAAA,EACI,oBAAA,CAAqB,CAAA,EAAG,QAAA,CAAS,CAAA,EAAG,GAAA,KACpC,sBAAA,CAAuB,cAAA,CAAe,CAAA,GAAI,QAAA,CAAS,CAAA,EAAG,GAAA"}
|
package/dist/context.d.ts
CHANGED
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
import { BaseMessage as BaseMessage$1 } from "./node_modules/.pnpm/langchain@1.2.30_@langchain_core@1.1.31_@opentelemetry_api@1.9.0_openai@6.27.0_ws@8.19._0520fb05d9e85da5f9e061dfe28cdbc8/node_modules/langchain/dist/index.js";
|
|
1
|
+
import { BaseMessage as BaseMessage$1, BaseMessageChunk } from "./node_modules/.pnpm/langchain@1.2.30_@langchain_core@1.1.31_@opentelemetry_api@1.9.0_openai@6.27.0_ws@8.19._0520fb05d9e85da5f9e061dfe28cdbc8/node_modules/langchain/dist/index.js";
|
|
2
|
+
import { StreamServiceInstance } from "./stream-service-instance.js";
|
|
2
3
|
import * as _angular_core0 from "@angular/core";
|
|
3
4
|
import { EnvironmentProviders, InjectionToken } from "@angular/core";
|
|
4
5
|
import * as _langchain_langgraph_sdk_ui0 from "@langchain/langgraph-sdk/ui";
|
|
@@ -28,7 +29,7 @@ declare const STREAM_DEFAULTS: InjectionToken<StreamDefaults>;
|
|
|
28
29
|
* Injection token for a shared stream instance.
|
|
29
30
|
* Provide via `provideStream()` at the component level.
|
|
30
31
|
*/
|
|
31
|
-
declare const STREAM_INSTANCE: InjectionToken<
|
|
32
|
+
declare const STREAM_INSTANCE: InjectionToken<StreamServiceInstance<Record<string, unknown>, BagTemplate>>;
|
|
32
33
|
/**
|
|
33
34
|
* Provides default LangGraph configuration at the application level.
|
|
34
35
|
*
|
|
@@ -85,22 +86,27 @@ declare function provideStreamDefaults(defaults: StreamDefaults): EnvironmentPro
|
|
|
85
86
|
* ```
|
|
86
87
|
*/
|
|
87
88
|
declare function provideStream<T = Record<string, unknown>, Bag extends BagTemplate = BagTemplate>(options: ResolveStreamOptions<T, InferBag<T, Bag>> | UseStreamCustomOptions<InferStateType<T>, InferBag<T, Bag>>): {
|
|
88
|
-
provide: InjectionToken<
|
|
89
|
+
provide: InjectionToken<StreamServiceInstance<Record<string, unknown>, BagTemplate>>;
|
|
89
90
|
useFactory: () => {
|
|
91
|
+
assistantId: string;
|
|
92
|
+
client: Client<_langchain_langgraph_sdk0.DefaultValues, _langchain_langgraph_sdk0.DefaultValues, unknown>;
|
|
90
93
|
values: _angular_core0.Signal<Record<string, unknown>>;
|
|
91
94
|
error: _angular_core0.Signal<unknown>;
|
|
92
95
|
isLoading: _angular_core0.WritableSignal<boolean>;
|
|
93
|
-
|
|
96
|
+
branch: _angular_core0.WritableSignal<string>;
|
|
97
|
+
setBranch(value: string): void;
|
|
98
|
+
messages: _angular_core0.Signal<(BaseMessage$1<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType> | BaseMessageChunk<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType>)[]>;
|
|
99
|
+
toolCalls: _angular_core0.Signal<_langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[]>;
|
|
100
|
+
getToolCalls(message: _langchain_langgraph_sdk0.Message): _langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[];
|
|
94
101
|
interrupt: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown> | undefined>;
|
|
95
102
|
interrupts: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown>[]>;
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
joinStream: (runId: string, lastEventId?: string | undefined, options?: {
|
|
103
|
+
history: _angular_core0.Signal<_langchain_langgraph_sdk0.ThreadState<any>[]>;
|
|
104
|
+
isThreadLoading: _angular_core0.Signal<boolean>;
|
|
105
|
+
experimental_branchTree: _angular_core0.Signal<_langchain_langgraph_sdk_ui0.Sequence<any>>;
|
|
106
|
+
getMessagesMetadata(message: _langchain_langgraph_sdk0.Message, index?: number | undefined): _langchain_langgraph_sdk_ui0.MessageMetadata<Record<string, unknown>> | undefined;
|
|
107
|
+
submit: (values: Record<string, unknown>, submitOptions?: _langchain_langgraph_sdk_ui0.SubmitOptions<Record<string, unknown>, Record<string, unknown>> | undefined) => Promise<void | Promise<void>>;
|
|
108
|
+
stop: () => void;
|
|
109
|
+
joinStream: (runId: string, lastEventId?: string | undefined, joinOptions?: {
|
|
104
110
|
streamMode?: _langchain_langgraph_sdk0.StreamMode[] | _langchain_langgraph_sdk0.StreamMode | undefined;
|
|
105
111
|
filter?: ((event: {
|
|
106
112
|
id?: string | undefined;
|
|
@@ -108,21 +114,44 @@ declare function provideStream<T = Record<string, unknown>, Bag extends BagTempl
|
|
|
108
114
|
data: unknown;
|
|
109
115
|
}) => boolean) | undefined;
|
|
110
116
|
} | undefined) => Promise<void>;
|
|
111
|
-
switchThread: (newThreadId: string | null) => void;
|
|
112
117
|
queue: {
|
|
113
118
|
entries: _angular_core0.WritableSignal<readonly _langchain_langgraph_sdk_ui0.QueueEntry<Record<string, unknown>, _langchain_langgraph_sdk_ui0.SubmitOptions<Record<string, unknown>, Record<string, unknown>>>[]>;
|
|
114
119
|
size: _angular_core0.WritableSignal<number>;
|
|
115
120
|
cancel: (id: string) => Promise<boolean>;
|
|
116
121
|
clear: () => Promise<void>;
|
|
117
122
|
};
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
|
|
123
|
+
switchThread(newThreadId: string | null): void;
|
|
124
|
+
subagents: _angular_core0.Signal<ReadonlyMap<string, _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>>>;
|
|
125
|
+
activeSubagents: _angular_core0.Signal<readonly _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[]>;
|
|
126
|
+
getSubagent(toolCallId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string> | undefined;
|
|
127
|
+
getSubagentsByType(type: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
128
|
+
getSubagentsByMessage(messageId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
129
|
+
} | {
|
|
130
|
+
values: _angular_core0.Signal<Record<string, unknown>>;
|
|
131
|
+
error: _angular_core0.Signal<unknown>;
|
|
132
|
+
isLoading: _angular_core0.WritableSignal<boolean>;
|
|
133
|
+
stop: () => void;
|
|
134
|
+
submit(values: Partial<Record<string, unknown>> | null | undefined, submitOptions?: _langchain_langgraph_sdk_ui0.CustomSubmitOptions<Record<string, unknown>, Record<string, unknown>> | undefined): Promise<void>;
|
|
135
|
+
switchThread(newThreadId: string | null): void;
|
|
136
|
+
branch: _angular_core0.WritableSignal<string>;
|
|
137
|
+
setBranch(value: string): void;
|
|
138
|
+
getMessagesMetadata(message: _langchain_langgraph_sdk0.Message<_langchain_langgraph_sdk0.DefaultToolCall>, index?: number | undefined): _langchain_langgraph_sdk_ui0.MessageMetadata<Record<string, unknown>> | undefined;
|
|
139
|
+
queue: {
|
|
140
|
+
entries: _angular_core0.WritableSignal<never[]>;
|
|
141
|
+
size: _angular_core0.WritableSignal<number>;
|
|
142
|
+
cancel(): Promise<boolean>;
|
|
143
|
+
clear(): Promise<void>;
|
|
144
|
+
};
|
|
145
|
+
interrupts: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown>[]>;
|
|
146
|
+
interrupt: _angular_core0.Signal<_langchain_langgraph_sdk0.Interrupt<unknown> | undefined>;
|
|
147
|
+
messages: _angular_core0.Signal<(BaseMessage$1<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType> | BaseMessageChunk<_langchain_core_messages0.MessageStructure<_langchain_core_messages0.MessageToolSet>, _langchain_core_messages0.MessageType>)[]>;
|
|
148
|
+
toolCalls: _angular_core0.Signal<_langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[]>;
|
|
149
|
+
getToolCalls(message: _langchain_langgraph_sdk0.Message<_langchain_langgraph_sdk0.DefaultToolCall>): _langchain_langgraph_sdk0.ToolCallWithResult<_langchain_langgraph_sdk0.DefaultToolCall>[];
|
|
150
|
+
subagents: _angular_core0.Signal<ReadonlyMap<string, _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>>>;
|
|
151
|
+
activeSubagents: _angular_core0.Signal<readonly _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[]>;
|
|
152
|
+
getSubagent(toolCallId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string> | undefined;
|
|
153
|
+
getSubagentsByType(type: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
154
|
+
getSubagentsByMessage(messageId: string): _langchain_langgraph_sdk_ui0.SubagentStreamInterface<Record<string, unknown>, _langchain_langgraph_sdk0.DefaultToolCall, string>[];
|
|
126
155
|
};
|
|
127
156
|
};
|
|
128
157
|
//#endregion
|
package/dist/context.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"context.d.ts","names":[],"sources":["../src/context.ts"],"mappings":"
|
|
1
|
+
{"version":3,"file":"context.d.ts","names":[],"sources":["../src/context.ts"],"mappings":";;;;;;;;;;;;;;UAqBiB,cAAA;;EAEf,MAAA;EAFe;EAIf,MAAA;;EAEA,MAAA,GAAS,MAAA;AAAA;;;;;cAOE,eAAA,EAAe,cAAA,CAAA,cAAA;AAA5B;;;;AAAA,cAQa,eAAA,EAAe,cAAA,CAAA,qBAAA,CAAA,MAAA,mBAAA,WAAA;AAA5B;;;;;;;;;;;;;;;AAyBA;;;;;;AAzBA,iBAyBgB,qBAAA,CACd,QAAA,EAAU,cAAA,GACT,oBAAA;;;AAuCH;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;iBAAgB,aAAA,KACV,MAAA,+BACQ,WAAA,GAAc,WAAA,CAAA,CAE1B,OAAA,EACI,oBAAA,CAAqB,CAAA,EAAG,QAAA,CAAS,CAAA,EAAG,GAAA,KACpC,sBAAA,CAAuB,cAAA,CAAe,CAAA,GAAI,QAAA,CAAS,CAAA,EAAG,GAAA"}
|
package/dist/context.js
CHANGED
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
import {
|
|
1
|
+
import { injectStreamCustom } from "./stream.custom.js";
|
|
2
|
+
import { useStreamLGP } from "./stream.lgp.js";
|
|
2
3
|
import { InjectionToken, inject, makeEnvironmentProviders } from "@angular/core";
|
|
3
4
|
//#region src/context.ts
|
|
4
5
|
/**
|
|
@@ -76,12 +77,14 @@ function provideStream(options) {
|
|
|
76
77
|
provide: STREAM_INSTANCE,
|
|
77
78
|
useFactory: () => {
|
|
78
79
|
const defaults = inject(STREAM_DEFAULTS, { optional: true });
|
|
79
|
-
|
|
80
|
+
const merged = {
|
|
80
81
|
...defaults ?? {},
|
|
81
82
|
...options,
|
|
82
83
|
apiUrl: options.apiUrl ?? defaults?.apiUrl,
|
|
83
84
|
client: options.client ?? defaults?.client
|
|
84
|
-
}
|
|
85
|
+
};
|
|
86
|
+
if ("transport" in merged) return injectStreamCustom(merged);
|
|
87
|
+
return useStreamLGP(merged);
|
|
85
88
|
}
|
|
86
89
|
};
|
|
87
90
|
}
|
package/dist/context.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"context.js","names":["angularInject"],"sources":["../src/context.ts"],"sourcesContent":["import {\n InjectionToken,\n inject as angularInject,\n type EnvironmentProviders,\n makeEnvironmentProviders,\n} from \"@angular/core\";\nimport type { BagTemplate } from \"@langchain/langgraph-sdk\";\nimport { Client } from \"@langchain/langgraph-sdk\";\nimport type {\n ResolveStreamOptions,\n InferBag,\n InferStateType,\n UseStreamCustomOptions,\n} from \"@langchain/langgraph-sdk/ui\";\nimport {
|
|
1
|
+
{"version":3,"file":"context.js","names":["angularInject"],"sources":["../src/context.ts"],"sourcesContent":["import {\n InjectionToken,\n inject as angularInject,\n type EnvironmentProviders,\n makeEnvironmentProviders,\n} from \"@angular/core\";\nimport type { BagTemplate } from \"@langchain/langgraph-sdk\";\nimport { Client } from \"@langchain/langgraph-sdk\";\nimport type {\n ResolveStreamOptions,\n InferBag,\n InferStateType,\n UseStreamCustomOptions,\n} from \"@langchain/langgraph-sdk/ui\";\nimport type { StreamServiceInstance } from \"./stream-service-instance.js\";\nimport { useStreamLGP } from \"./stream.lgp.js\";\nimport { injectStreamCustom } from \"./stream.custom.js\";\n\n/**\n * Configuration defaults for `useStream` and `injectStream` calls.\n */\nexport interface StreamDefaults {\n /** Base URL of the LangGraph API. */\n apiUrl?: string;\n /** API key for authenticating with the LangGraph API. */\n apiKey?: string;\n /** Pre-configured Client instance. */\n client?: Client;\n}\n\n/**\n * Injection token for stream default configuration.\n * Provide via `provideStreamDefaults()` in your application config.\n */\nexport const STREAM_DEFAULTS = new InjectionToken<StreamDefaults>(\n \"LANGCHAIN_STREAM_DEFAULTS\",\n);\n\n/**\n * Injection token for a shared stream instance.\n * Provide via `provideStream()` at the component level.\n */\nexport const STREAM_INSTANCE = new InjectionToken<StreamServiceInstance>(\n \"LANGCHAIN_STREAM_INSTANCE\",\n);\n\n/**\n * Provides default LangGraph configuration at the application level.\n *\n * Use this in your application's `providers` array to set defaults like\n * `apiUrl` that will be used by all `useStream` and `injectStream` calls.\n *\n * @example\n * ```typescript\n * // app.config.ts\n * import { ApplicationConfig } from \"@angular/core\";\n * import { provideStreamDefaults } from \"@langchain/angular\";\n *\n * export const appConfig: ApplicationConfig = {\n * providers: [\n * provideStreamDefaults({\n * apiUrl: \"http://localhost:2024\",\n * }),\n * ],\n * };\n * ```\n */\nexport function provideStreamDefaults(\n defaults: StreamDefaults,\n): EnvironmentProviders {\n return makeEnvironmentProviders([\n { provide: STREAM_DEFAULTS, useValue: defaults },\n ]);\n}\n\n/**\n * Creates a provider for a shared `useStream` instance at the component level.\n *\n * Add the returned provider to a component's `providers` array so that all\n * child components can access the same stream via `injectStream()`.\n *\n * @example\n * ```typescript\n * import { Component } from \"@angular/core\";\n * import { provideStream, injectStream } from \"@langchain/angular\";\n *\n * @Component({\n * providers: [provideStream({ assistantId: \"agent\" })],\n * template: `\n * <app-message-list />\n * <app-message-input />\n * `,\n * })\n * export class ChatContainer {}\n *\n * // In child components:\n * @Component({\n * template: `\n * @for (msg of stream.messages(); track msg.id) {\n * <div>{{ msg.content }}</div>\n * }\n * `,\n * })\n * export class MessageListComponent {\n * stream = injectStream();\n * }\n * ```\n */\nexport function provideStream<\n T = Record<string, unknown>,\n Bag extends BagTemplate = BagTemplate,\n>(\n options:\n | ResolveStreamOptions<T, InferBag<T, Bag>>\n | UseStreamCustomOptions<InferStateType<T>, InferBag<T, Bag>>,\n) {\n return {\n provide: STREAM_INSTANCE,\n useFactory: () => {\n const defaults = angularInject(STREAM_DEFAULTS, { optional: true });\n const merged = {\n ...(defaults ?? {}),\n ...options,\n apiUrl: (options as Record<string, unknown>).apiUrl ?? defaults?.apiUrl,\n client: (options as Record<string, unknown>).client ?? defaults?.client,\n };\n if (\"transport\" in merged) {\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n return injectStreamCustom(merged as any);\n }\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n return useStreamLGP(merged as any);\n },\n };\n}\n"],"mappings":";;;;;;;;AAkCA,MAAa,kBAAkB,IAAI,eACjC,4BACD;;;;;AAMD,MAAa,kBAAkB,IAAI,eACjC,4BACD;;;;;;;;;;;;;;;;;;;;;;AAuBD,SAAgB,sBACd,UACsB;AACtB,QAAO,yBAAyB,CAC9B;EAAE,SAAS;EAAiB,UAAU;EAAU,CACjD,CAAC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAoCJ,SAAgB,cAId,SAGA;AACA,QAAO;EACL,SAAS;EACT,kBAAkB;GAChB,MAAM,WAAWA,OAAc,iBAAiB,EAAE,UAAU,MAAM,CAAC;GACnE,MAAM,SAAS;IACb,GAAI,YAAY,EAAE;IAClB,GAAG;IACH,QAAS,QAAoC,UAAU,UAAU;IACjE,QAAS,QAAoC,UAAU,UAAU;IAClE;AACD,OAAI,eAAe,OAEjB,QAAO,mBAAmB,OAAc;AAG1C,UAAO,aAAa,OAAc;;EAErC"}
|
package/dist/index.cjs
CHANGED
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
Object.defineProperty(exports, Symbol.toStringTag, { value: "Module" });
|
|
2
2
|
const require_stream_custom = require("./stream.custom.cjs");
|
|
3
|
+
const require_stream_lgp = require("./stream.lgp.cjs");
|
|
3
4
|
const require_context = require("./context.cjs");
|
|
4
5
|
const require_decorate = require("./_virtual/_@oxc-project_runtime@0.115.0/helpers/decorate.cjs");
|
|
5
6
|
let _angular_core = require("@angular/core");
|
|
6
7
|
let _langchain_langgraph_sdk_ui = require("@langchain/langgraph-sdk/ui");
|
|
7
|
-
let _langchain_langgraph_sdk = require("@langchain/langgraph-sdk");
|
|
8
8
|
//#region src/index.ts
|
|
9
9
|
/**
|
|
10
10
|
* @internal Merges DI, LangGraph Platform, and custom transport overloads.
|
|
@@ -16,7 +16,7 @@ function injectStream(options) {
|
|
|
16
16
|
return instance;
|
|
17
17
|
}
|
|
18
18
|
if ("transport" in options) return require_stream_custom.injectStreamCustom(options);
|
|
19
|
-
return useStreamLGP(options);
|
|
19
|
+
return require_stream_lgp.useStreamLGP(options);
|
|
20
20
|
}
|
|
21
21
|
/**
|
|
22
22
|
* @deprecated Use `injectStream` instead. `useStream` will be removed in a
|
|
@@ -24,159 +24,6 @@ function injectStream(options) {
|
|
|
24
24
|
* convention for injection-based patterns.
|
|
25
25
|
*/
|
|
26
26
|
const useStream = injectStream;
|
|
27
|
-
function useStreamLGP(options) {
|
|
28
|
-
const client = options.client ?? new _langchain_langgraph_sdk.Client({ apiUrl: options.apiUrl });
|
|
29
|
-
const orchestrator = new _langchain_langgraph_sdk_ui.StreamOrchestrator(options, {
|
|
30
|
-
getClient: () => client,
|
|
31
|
-
getAssistantId: () => options.assistantId,
|
|
32
|
-
getMessagesKey: () => options.messagesKey ?? "messages"
|
|
33
|
-
});
|
|
34
|
-
orchestrator.initThreadId(options.threadId ?? void 0);
|
|
35
|
-
const version = (0, _angular_core.signal)(0);
|
|
36
|
-
const subagentVersion = (0, _angular_core.signal)(0);
|
|
37
|
-
(0, _angular_core.effect)((onCleanup) => {
|
|
38
|
-
const unsubscribe = orchestrator.subscribe(() => {
|
|
39
|
-
version.update((v) => v + 1);
|
|
40
|
-
subagentVersion.update((v) => v + 1);
|
|
41
|
-
});
|
|
42
|
-
onCleanup(() => unsubscribe());
|
|
43
|
-
});
|
|
44
|
-
(0, _angular_core.effect)((onCleanup) => {
|
|
45
|
-
version();
|
|
46
|
-
const hvMessages = orchestrator.messages;
|
|
47
|
-
if (options.filterSubagentMessages && !orchestrator.isLoading && !orchestrator.historyData.isLoading && hvMessages.length > 0) {
|
|
48
|
-
const controller = orchestrator.reconstructSubagentsIfNeeded();
|
|
49
|
-
if (controller) onCleanup(() => controller.abort());
|
|
50
|
-
}
|
|
51
|
-
});
|
|
52
|
-
const isLoadingForDrain = (0, _angular_core.computed)(() => {
|
|
53
|
-
version();
|
|
54
|
-
return orchestrator.isLoading;
|
|
55
|
-
});
|
|
56
|
-
(0, _angular_core.effect)(() => {
|
|
57
|
-
isLoadingForDrain();
|
|
58
|
-
orchestrator.drainQueue();
|
|
59
|
-
});
|
|
60
|
-
const { shouldReconnect } = orchestrator;
|
|
61
|
-
let hasReconnected = false;
|
|
62
|
-
(0, _angular_core.effect)(() => {
|
|
63
|
-
version();
|
|
64
|
-
const tid = orchestrator.threadId;
|
|
65
|
-
if (!hasReconnected && shouldReconnect && tid && !orchestrator.isLoading) {
|
|
66
|
-
if (orchestrator.tryReconnect()) hasReconnected = true;
|
|
67
|
-
}
|
|
68
|
-
});
|
|
69
|
-
const values = (0, _angular_core.computed)(() => {
|
|
70
|
-
version();
|
|
71
|
-
orchestrator.trackStreamMode("values");
|
|
72
|
-
return orchestrator.values;
|
|
73
|
-
});
|
|
74
|
-
const error = (0, _angular_core.computed)(() => {
|
|
75
|
-
version();
|
|
76
|
-
return orchestrator.error;
|
|
77
|
-
});
|
|
78
|
-
const isLoading = (0, _angular_core.signal)(orchestrator.isLoading);
|
|
79
|
-
(0, _angular_core.effect)(() => {
|
|
80
|
-
version();
|
|
81
|
-
isLoading.set(orchestrator.isLoading);
|
|
82
|
-
});
|
|
83
|
-
const branch = (0, _angular_core.signal)("");
|
|
84
|
-
(0, _angular_core.effect)(() => {
|
|
85
|
-
version();
|
|
86
|
-
const b = orchestrator.branch;
|
|
87
|
-
if (branch() !== b) branch.set(b);
|
|
88
|
-
});
|
|
89
|
-
const messages = (0, _angular_core.computed)(() => {
|
|
90
|
-
version();
|
|
91
|
-
orchestrator.trackStreamMode("messages-tuple", "values");
|
|
92
|
-
return (0, _langchain_langgraph_sdk_ui.ensureMessageInstances)(orchestrator.messages);
|
|
93
|
-
});
|
|
94
|
-
const toolCalls = (0, _angular_core.computed)(() => {
|
|
95
|
-
version();
|
|
96
|
-
orchestrator.trackStreamMode("messages-tuple", "values");
|
|
97
|
-
return orchestrator.toolCalls;
|
|
98
|
-
});
|
|
99
|
-
const interrupt = (0, _angular_core.computed)(() => {
|
|
100
|
-
version();
|
|
101
|
-
return orchestrator.interrupt;
|
|
102
|
-
});
|
|
103
|
-
const interrupts = (0, _angular_core.computed)(() => {
|
|
104
|
-
version();
|
|
105
|
-
return orchestrator.interrupts;
|
|
106
|
-
});
|
|
107
|
-
const historyList = (0, _angular_core.computed)(() => {
|
|
108
|
-
version();
|
|
109
|
-
return orchestrator.flatHistory;
|
|
110
|
-
});
|
|
111
|
-
const isThreadLoading = (0, _angular_core.computed)(() => {
|
|
112
|
-
version();
|
|
113
|
-
return orchestrator.isThreadLoading;
|
|
114
|
-
});
|
|
115
|
-
const experimentalBranchTree = (0, _angular_core.computed)(() => {
|
|
116
|
-
version();
|
|
117
|
-
return orchestrator.experimental_branchTree;
|
|
118
|
-
});
|
|
119
|
-
const queueEntries = (0, _angular_core.signal)(orchestrator.queueEntries);
|
|
120
|
-
const queueSize = (0, _angular_core.signal)(orchestrator.queueSize);
|
|
121
|
-
(0, _angular_core.effect)(() => {
|
|
122
|
-
version();
|
|
123
|
-
queueEntries.set(orchestrator.queueEntries);
|
|
124
|
-
queueSize.set(orchestrator.queueSize);
|
|
125
|
-
});
|
|
126
|
-
return {
|
|
127
|
-
assistantId: options.assistantId,
|
|
128
|
-
client,
|
|
129
|
-
values,
|
|
130
|
-
error,
|
|
131
|
-
isLoading,
|
|
132
|
-
branch,
|
|
133
|
-
setBranch(value) {
|
|
134
|
-
orchestrator.setBranch(value);
|
|
135
|
-
},
|
|
136
|
-
messages,
|
|
137
|
-
toolCalls,
|
|
138
|
-
getToolCalls(message) {
|
|
139
|
-
return orchestrator.getToolCalls(message);
|
|
140
|
-
},
|
|
141
|
-
interrupt,
|
|
142
|
-
interrupts,
|
|
143
|
-
history: historyList,
|
|
144
|
-
isThreadLoading,
|
|
145
|
-
experimental_branchTree: experimentalBranchTree,
|
|
146
|
-
getMessagesMetadata(message, index) {
|
|
147
|
-
return orchestrator.getMessagesMetadata(message, index);
|
|
148
|
-
},
|
|
149
|
-
submit: (values, submitOptions) => orchestrator.submit(values, submitOptions),
|
|
150
|
-
stop: () => orchestrator.stop(),
|
|
151
|
-
joinStream: (...args) => orchestrator.joinStream(...args),
|
|
152
|
-
queue: {
|
|
153
|
-
entries: queueEntries,
|
|
154
|
-
size: queueSize,
|
|
155
|
-
cancel: (id) => orchestrator.cancelQueueItem(id),
|
|
156
|
-
clear: () => orchestrator.clearQueue()
|
|
157
|
-
},
|
|
158
|
-
switchThread(newThreadId) {
|
|
159
|
-
orchestrator.switchThread(newThreadId);
|
|
160
|
-
},
|
|
161
|
-
get subagents() {
|
|
162
|
-
subagentVersion();
|
|
163
|
-
return orchestrator.subagents;
|
|
164
|
-
},
|
|
165
|
-
get activeSubagents() {
|
|
166
|
-
subagentVersion();
|
|
167
|
-
return orchestrator.activeSubagents;
|
|
168
|
-
},
|
|
169
|
-
getSubagent(toolCallId) {
|
|
170
|
-
return orchestrator.getSubagent(toolCallId);
|
|
171
|
-
},
|
|
172
|
-
getSubagentsByType(type) {
|
|
173
|
-
return orchestrator.getSubagentsByType(type);
|
|
174
|
-
},
|
|
175
|
-
getSubagentsByMessage(messageId) {
|
|
176
|
-
return orchestrator.getSubagentsByMessage(messageId);
|
|
177
|
-
}
|
|
178
|
-
};
|
|
179
|
-
}
|
|
180
27
|
let StreamService = class StreamService {
|
|
181
28
|
_stream;
|
|
182
29
|
constructor(options) {
|
|
@@ -313,6 +160,6 @@ exports.provideStream = require_context.provideStream;
|
|
|
313
160
|
exports.provideStreamDefaults = require_context.provideStreamDefaults;
|
|
314
161
|
exports.useStream = useStream;
|
|
315
162
|
exports.useStreamCustom = require_stream_custom.useStreamCustom;
|
|
316
|
-
exports.useStreamLGP = useStreamLGP;
|
|
163
|
+
exports.useStreamLGP = require_stream_lgp.useStreamLGP;
|
|
317
164
|
|
|
318
165
|
//# sourceMappingURL=index.cjs.map
|