@aigne/ollama 0.3.11 → 0.4.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,22 @@
1
1
  # Changelog
2
2
 
3
+ ## [0.4.0](https://github.com/AIGNE-io/aigne-framework/compare/ollama-v0.3.11...ollama-v0.4.0) (2025-07-01)
4
+
5
+
6
+ ### Features
7
+
8
+ * **example:** use AIGNE cli to run chat-bot example ([#198](https://github.com/AIGNE-io/aigne-framework/issues/198)) ([7085541](https://github.com/AIGNE-io/aigne-framework/commit/708554100692f2a557f7329ea78e46c3c870ce10))
9
+
10
+
11
+ ### Dependencies
12
+
13
+ * The following workspace dependencies were updated
14
+ * dependencies
15
+ * @aigne/openai bumped to 0.6.0
16
+ * devDependencies
17
+ * @aigne/core bumped to 1.28.0
18
+ * @aigne/test-utils bumped to 0.4.12
19
+
3
20
  ## [0.3.11](https://github.com/AIGNE-io/aigne-framework/compare/ollama-v0.3.10...ollama-v0.3.11) (2025-07-01)
4
21
 
5
22
 
package/README.md CHANGED
@@ -80,6 +80,7 @@ console.log(result);
80
80
  ## Streaming Responses
81
81
 
82
82
  ```typescript file="test/ollama-chat-model.test.ts" region="example-ollama-chat-model-streaming"
83
+ import { isAgentResponseDelta } from "@aigne/core";
83
84
  import { OllamaChatModel } from "@aigne/ollama";
84
85
 
85
86
  const model = new OllamaChatModel({
@@ -91,7 +92,6 @@ const stream = await model.invoke(
91
92
  {
92
93
  messages: [{ role: "user", content: "Tell me what model you're using" }],
93
94
  },
94
- undefined,
95
95
  { streaming: true },
96
96
  );
97
97
 
@@ -99,9 +99,11 @@ let fullText = "";
99
99
  const json = {};
100
100
 
101
101
  for await (const chunk of stream) {
102
- const text = chunk.delta.text?.text;
103
- if (text) fullText += text;
104
- if (chunk.delta.json) Object.assign(json, chunk.delta.json);
102
+ if (isAgentResponseDelta(chunk)) {
103
+ const text = chunk.delta.text?.text;
104
+ if (text) fullText += text;
105
+ if (chunk.delta.json) Object.assign(json, chunk.delta.json);
106
+ }
105
107
  }
106
108
 
107
109
  console.log(fullText); // Output: "I'm an AI assistant running on Ollama with the llama3 model."
package/README.zh.md CHANGED
@@ -80,6 +80,7 @@ console.log(result);
80
80
  ## 流式响应
81
81
 
82
82
  ```typescript file="test/ollama-chat-model.test.ts" region="example-ollama-chat-model-streaming"
83
+ import { isAgentResponseDelta } from "@aigne/core";
83
84
  import { OllamaChatModel } from "@aigne/ollama";
84
85
 
85
86
  const model = new OllamaChatModel({
@@ -91,7 +92,6 @@ const stream = await model.invoke(
91
92
  {
92
93
  messages: [{ role: "user", content: "Tell me what model you're using" }],
93
94
  },
94
- undefined,
95
95
  { streaming: true },
96
96
  );
97
97
 
@@ -99,9 +99,11 @@ let fullText = "";
99
99
  const json = {};
100
100
 
101
101
  for await (const chunk of stream) {
102
- const text = chunk.delta.text?.text;
103
- if (text) fullText += text;
104
- if (chunk.delta.json) Object.assign(json, chunk.delta.json);
102
+ if (isAgentResponseDelta(chunk)) {
103
+ const text = chunk.delta.text?.text;
104
+ if (text) fullText += text;
105
+ if (chunk.delta.json) Object.assign(json, chunk.delta.json);
106
+ }
105
107
  }
106
108
 
107
109
  console.log(fullText); // Output: "I'm an AI assistant running on Ollama with the llama3 model."
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@aigne/ollama",
3
- "version": "0.3.11",
3
+ "version": "0.4.0",
4
4
  "description": "AIGNE Ollama SDK for integrating with locally hosted AI models via Ollama",
5
5
  "publishConfig": {
6
6
  "access": "public"
@@ -32,7 +32,7 @@
32
32
  }
33
33
  },
34
34
  "dependencies": {
35
- "@aigne/openai": "^0.5.0"
35
+ "@aigne/openai": "^0.6.0"
36
36
  },
37
37
  "devDependencies": {
38
38
  "@types/bun": "^1.2.12",
@@ -40,8 +40,8 @@
40
40
  "npm-run-all": "^4.1.5",
41
41
  "rimraf": "^6.0.1",
42
42
  "typescript": "^5.8.3",
43
- "@aigne/core": "^1.27.0",
44
- "@aigne/test-utils": "^0.4.11"
43
+ "@aigne/test-utils": "^0.4.12",
44
+ "@aigne/core": "^1.28.0"
45
45
  },
46
46
  "scripts": {
47
47
  "lint": "tsc --noEmit",