@aigne/example-workflow-concurrency 1.16.88-beta.9 → 1.16.88
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +5 -5
- package/package.json +6 -6
package/README.md
CHANGED
|
@@ -49,7 +49,7 @@ class aggregator processing
|
|
|
49
49
|
npx -y @aigne/example-workflow-concurrency
|
|
50
50
|
|
|
51
51
|
# Run in interactive chat mode
|
|
52
|
-
npx -y @aigne/example-workflow-concurrency --
|
|
52
|
+
npx -y @aigne/example-workflow-concurrency --interactive
|
|
53
53
|
|
|
54
54
|
# Use pipeline input
|
|
55
55
|
echo "Analyze product: Smart home assistant with voice control and AI learning capabilities" | npx -y @aigne/example-workflow-concurrency
|
|
@@ -57,7 +57,7 @@ echo "Analyze product: Smart home assistant with voice control and AI learning c
|
|
|
57
57
|
|
|
58
58
|
### Connect to an AI Model
|
|
59
59
|
|
|
60
|
-
As an example, running `npx -y @aigne/example-workflow-concurrency --
|
|
60
|
+
As an example, running `npx -y @aigne/example-workflow-concurrency --interactive` requires an AI model. If this is your first run, you need to connect one.
|
|
61
61
|
|
|
62
62
|

|
|
63
63
|
|
|
@@ -118,7 +118,7 @@ pnpm install
|
|
|
118
118
|
pnpm start # Run in one-shot mode (default)
|
|
119
119
|
|
|
120
120
|
# Run in interactive chat mode
|
|
121
|
-
pnpm start -- --
|
|
121
|
+
pnpm start -- --interactive
|
|
122
122
|
|
|
123
123
|
# Use pipeline input
|
|
124
124
|
echo "Analyze product: Smart home assistant with voice control and AI learning capabilities" | pnpm start
|
|
@@ -130,7 +130,7 @@ The example supports the following command-line parameters:
|
|
|
130
130
|
|
|
131
131
|
| Parameter | Description | Default |
|
|
132
132
|
|-----------|-------------|---------|
|
|
133
|
-
| `--
|
|
133
|
+
| `--interactive` | Run in interactive chat mode | Disabled (one-shot mode) |
|
|
134
134
|
| `--model <provider[:model]>` | AI model to use in format 'provider\[:model]' where model is optional. Examples: 'openai' or 'openai:gpt-4o-mini' | openai |
|
|
135
135
|
| `--temperature <value>` | Temperature for model generation | Provider default |
|
|
136
136
|
| `--top-p <value>` | Top-p sampling value | Provider default |
|
|
@@ -143,7 +143,7 @@ The example supports the following command-line parameters:
|
|
|
143
143
|
|
|
144
144
|
```bash
|
|
145
145
|
# Run in chat mode (interactive)
|
|
146
|
-
pnpm start -- --
|
|
146
|
+
pnpm start -- --interactive
|
|
147
147
|
|
|
148
148
|
# Set logging level
|
|
149
149
|
pnpm start -- --log-level DEBUG
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@aigne/example-workflow-concurrency",
|
|
3
|
-
"version": "1.16.88
|
|
3
|
+
"version": "1.16.88",
|
|
4
4
|
"description": "A demonstration of using AIGNE Framework to build a concurrency workflow",
|
|
5
5
|
"author": "Arcblock <blocklet@arcblock.io> https://github.com/blocklet",
|
|
6
6
|
"homepage": "https://github.com/AIGNE-io/aigne-framework/tree/main/examples/workflow-concurrency",
|
|
@@ -16,14 +16,14 @@
|
|
|
16
16
|
"README.md"
|
|
17
17
|
],
|
|
18
18
|
"dependencies": {
|
|
19
|
-
"@aigne/agent-library": "^1.24.0
|
|
20
|
-
"@aigne/
|
|
21
|
-
"@aigne/
|
|
22
|
-
"@aigne/openai": "^0.16.16
|
|
19
|
+
"@aigne/agent-library": "^1.24.0",
|
|
20
|
+
"@aigne/cli": "^1.59.0",
|
|
21
|
+
"@aigne/core": "^1.72.0",
|
|
22
|
+
"@aigne/openai": "^0.16.16"
|
|
23
23
|
},
|
|
24
24
|
"devDependencies": {
|
|
25
25
|
"@types/bun": "^1.2.22",
|
|
26
|
-
"@aigne/test-utils": "^0.5.69
|
|
26
|
+
"@aigne/test-utils": "^0.5.69"
|
|
27
27
|
},
|
|
28
28
|
"scripts": {
|
|
29
29
|
"start": "bun run index.ts",
|