llmist 2.1.0 → 2.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -30,7 +30,7 @@ import {
30
30
  init_strategy,
31
31
  init_stream_processor,
32
32
  resolveHintTemplate
33
- } from "./chunk-PDYVT3FI.js";
33
+ } from "./chunk-GANXNBIZ.js";
34
34
 
35
35
  // src/index.ts
36
36
  init_builder();
@@ -978,4 +978,4 @@ export {
978
978
  Gadget,
979
979
  z
980
980
  };
981
- //# sourceMappingURL=chunk-LSCCBXS7.js.map
981
+ //# sourceMappingURL=chunk-ZDNV7DDO.js.map
package/dist/cli.cjs CHANGED
@@ -3835,6 +3835,23 @@ var init_agent = __esm({
3835
3835
  maxIterations: this.maxIterations
3836
3836
  });
3837
3837
  while (currentIteration < this.maxIterations) {
3838
+ if (this.signal?.aborted) {
3839
+ this.logger.info("Agent loop terminated by abort signal", {
3840
+ iteration: currentIteration,
3841
+ reason: this.signal.reason
3842
+ });
3843
+ await this.safeObserve(async () => {
3844
+ if (this.hooks.observers?.onAbort) {
3845
+ const context = {
3846
+ iteration: currentIteration,
3847
+ reason: this.signal?.reason,
3848
+ logger: this.logger
3849
+ };
3850
+ await this.hooks.observers.onAbort(context);
3851
+ }
3852
+ });
3853
+ return;
3854
+ }
3838
3855
  this.logger.debug("Starting iteration", { iteration: currentIteration });
3839
3856
  try {
3840
3857
  if (this.compactionManager) {
@@ -7028,8 +7045,8 @@ var import_commander2 = require("commander");
7028
7045
  // package.json
7029
7046
  var package_default = {
7030
7047
  name: "llmist",
7031
- version: "2.0.0",
7032
- description: "Universal TypeScript LLM client with streaming-first agent framework. Works with any model - no structured outputs or native tool calling required. Implements its own flexible grammar for function calling.",
7048
+ version: "2.2.0",
7049
+ description: "TypeScript LLM client with streaming tool execution. Tools fire mid-stream. Built-in function calling works with any model\u2014no structured outputs or native tool support required.",
7033
7050
  type: "module",
7034
7051
  main: "dist/index.cjs",
7035
7052
  module: "dist/index.js",
@@ -7103,9 +7120,16 @@ var package_default = {
7103
7120
  "universal-client",
7104
7121
  "multi-provider",
7105
7122
  "hooks",
7106
- "gadgets"
7123
+ "gadgets",
7124
+ "chatbot",
7125
+ "chatgpt",
7126
+ "agentic",
7127
+ "language-model",
7128
+ "generative-ai",
7129
+ "bun",
7130
+ "nodejs"
7107
7131
  ],
7108
- author: "",
7132
+ author: "Zbigniew Sobiecki <zbigniew@sobiecki.name>",
7109
7133
  license: "MIT",
7110
7134
  dependencies: {
7111
7135
  "@anthropic-ai/sdk": "^0.69.0",