@yourgpt/llm-sdk 1.2.0 → 1.2.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -3452,7 +3452,13 @@ var Runtime = class {
3452
3452
  options
3453
3453
  );
3454
3454
  }
3455
- const generator = useAgentLoop ? this.processChatWithLoop(body, signal, void 0, void 0, request) : this.processChat(body, signal);
3455
+ const generator = this.processChatWithLoop(
3456
+ body,
3457
+ signal,
3458
+ void 0,
3459
+ void 0,
3460
+ request
3461
+ );
3456
3462
  const wrappedGenerator = this.wrapGeneratorWithOnFinish(
3457
3463
  generator,
3458
3464
  body.threadId,
@@ -3499,15 +3505,15 @@ var Runtime = class {
3499
3505
  /**
3500
3506
  * Handle non-streaming request - returns JSON instead of SSE
3501
3507
  */
3502
- async handleNonStreamingRequest(body, signal, useAgentLoop, httpRequest, options) {
3508
+ async handleNonStreamingRequest(body, signal, _useAgentLoop, httpRequest, options) {
3503
3509
  try {
3504
- const generator = useAgentLoop ? this.processChatWithLoop(
3510
+ const generator = this.processChatWithLoop(
3505
3511
  body,
3506
3512
  signal,
3507
3513
  void 0,
3508
3514
  void 0,
3509
3515
  httpRequest
3510
- ) : this.processChat(body, signal);
3516
+ );
3511
3517
  const events = [];
3512
3518
  let content = "";
3513
3519
  const toolCalls = [];