graphlit-client 1.0.20250714001 → 1.0.20250716001

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1505,7 +1505,11 @@ export declare enum CerebrasModels {
1505
1505
  /** LLaMA 3.1 8b */
1506
1506
  Llama_3_1_8B = "LLAMA_3_1_8B",
1507
1507
  /** LLaMA 3.3 70b */
1508
- Llama_3_3_70B = "LLAMA_3_3_70B"
1508
+ Llama_3_3_70B = "LLAMA_3_3_70B",
1509
+ /** LLaMA 4 Scout 17b */
1510
+ Llama_4Scout_17B = "LLAMA_4_SCOUT_17B",
1511
+ /** Qwen 3 32b */
1512
+ Qwen_3_32B = "QWEN_3_32B"
1509
1513
  }
1510
1514
  /** Represents a classification workflow job. */
1511
1515
  export type ClassificationWorkflowJob = {
@@ -4471,10 +4475,7 @@ export declare enum FeedTypes {
4471
4475
  Twitter = "TWITTER",
4472
4476
  /** Web feed */
4473
4477
  Web = "WEB",
4474
- /**
4475
- * YouTube audio feed
4476
- * @deprecated No longer supported. We suggest using your own YouTube downloader and then ingest the video or audio files.
4477
- */
4478
+ /** YouTube audio feed */
4478
4479
  YouTube = "YOU_TUBE",
4479
4480
  /** Zendesk articles feed */
4480
4481
  Zendesk = "ZENDESK"
@@ -5290,6 +5291,8 @@ export declare enum GroqModels {
5290
5291
  Custom = "CUSTOM",
5291
5292
  /** Deepseek R1 Distill Llama 70b Preview */
5292
5293
  DeepseekR1Llama_70BPreview = "DEEPSEEK_R1_LLAMA_70B_PREVIEW",
5294
+ /** Kimi K2 32b */
5295
+ KimiK2_32B = "KIMI_K2_32B",
5293
5296
  /** LLaMA 3.1 8b */
5294
5297
  Llama_3_1_8B = "LLAMA_3_1_8B",
5295
5298
  /**
@@ -5323,7 +5326,9 @@ export declare enum GroqModels {
5323
5326
  /** LLaMA 4 Scout 17b */
5324
5327
  Llama_4Scout_17B = "LLAMA_4_SCOUT_17B",
5325
5328
  /** Mixtral 8x7b Instruct */
5326
- Mixtral_8X7BInstruct = "MIXTRAL_8X7B_INSTRUCT"
5329
+ Mixtral_8X7BInstruct = "MIXTRAL_8X7B_INSTRUCT",
5330
+ /** Qwen 3 32b */
5331
+ Qwen_3_32B = "QWEN_3_32B"
5327
5332
  }
5328
5333
  /** Represents an H3 index. */
5329
5334
  export type H3 = {
@@ -279,6 +279,10 @@ export var CerebrasModels;
279
279
  CerebrasModels["Llama_3_1_8B"] = "LLAMA_3_1_8B";
280
280
  /** LLaMA 3.3 70b */
281
281
  CerebrasModels["Llama_3_3_70B"] = "LLAMA_3_3_70B";
282
+ /** LLaMA 4 Scout 17b */
283
+ CerebrasModels["Llama_4Scout_17B"] = "LLAMA_4_SCOUT_17B";
284
+ /** Qwen 3 32b */
285
+ CerebrasModels["Qwen_3_32B"] = "QWEN_3_32B";
282
286
  })(CerebrasModels || (CerebrasModels = {}));
283
287
  /** Cohere model type */
284
288
  export var CohereModels;
@@ -917,10 +921,7 @@ export var FeedTypes;
917
921
  FeedTypes["Twitter"] = "TWITTER";
918
922
  /** Web feed */
919
923
  FeedTypes["Web"] = "WEB";
920
- /**
921
- * YouTube audio feed
922
- * @deprecated No longer supported. We suggest using your own YouTube downloader and then ingest the video or audio files.
923
- */
924
+ /** YouTube audio feed */
924
925
  FeedTypes["YouTube"] = "YOU_TUBE";
925
926
  /** Zendesk articles feed */
926
927
  FeedTypes["Zendesk"] = "ZENDESK";
@@ -1082,6 +1083,8 @@ export var GroqModels;
1082
1083
  GroqModels["Custom"] = "CUSTOM";
1083
1084
  /** Deepseek R1 Distill Llama 70b Preview */
1084
1085
  GroqModels["DeepseekR1Llama_70BPreview"] = "DEEPSEEK_R1_LLAMA_70B_PREVIEW";
1086
+ /** Kimi K2 32b */
1087
+ GroqModels["KimiK2_32B"] = "KIMI_K2_32B";
1085
1088
  /** LLaMA 3.1 8b */
1086
1089
  GroqModels["Llama_3_1_8B"] = "LLAMA_3_1_8B";
1087
1090
  /**
@@ -1116,6 +1119,8 @@ export var GroqModels;
1116
1119
  GroqModels["Llama_4Scout_17B"] = "LLAMA_4_SCOUT_17B";
1117
1120
  /** Mixtral 8x7b Instruct */
1118
1121
  GroqModels["Mixtral_8X7BInstruct"] = "MIXTRAL_8X7B_INSTRUCT";
1122
+ /** Qwen 3 32b */
1123
+ GroqModels["Qwen_3_32B"] = "QWEN_3_32B";
1119
1124
  })(GroqModels || (GroqModels = {}));
1120
1125
  /** H3 index resolution types */
1121
1126
  export var H3ResolutionTypes;
@@ -962,13 +962,6 @@ onEvent, onComplete, abortSignal) {
962
962
  const result = await chat.sendMessageStream(prompt);
963
963
  for await (const chunk of result.stream) {
964
964
  const text = chunk.text();
965
- // Debug log chunk details
966
- if (process.env.DEBUG_GRAPHLIT_SDK_STREAMING) {
967
- console.log(`[Google] Raw chunk:`, JSON.stringify(chunk, null, 2));
968
- if (text) {
969
- console.log(`[Google] Text delta: "${text}" (${text.length} chars)`);
970
- }
971
- }
972
965
  if (text) {
973
966
  fullMessage += text;
974
967
  tokenCount++;
@@ -1103,18 +1096,12 @@ onEvent, onComplete, abortSignal) {
1103
1096
  }
1104
1097
  // Only add if it's not already included in fullMessage
1105
1098
  if (!fullMessage.includes(finalText) && !fullMessage.endsWith(finalText)) {
1106
- if (process.env.DEBUG_GRAPHLIT_SDK_STREAMING) {
1107
- console.log(`[Google] Adding final text: ${finalText.length} chars`);
1108
- }
1109
1099
  fullMessage += finalText;
1110
1100
  onEvent({
1111
1101
  type: "token",
1112
1102
  token: finalText,
1113
1103
  });
1114
1104
  }
1115
- else if (process.env.DEBUG_GRAPHLIT_SDK_STREAMING) {
1116
- console.log(`[Google] Skipping final text (already in message): ${finalText.length} chars`);
1117
- }
1118
1105
  }
1119
1106
  // Check for function calls
1120
1107
  if (part.functionCall &&
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "graphlit-client",
3
- "version": "1.0.20250714001",
3
+ "version": "1.0.20250716001",
4
4
  "description": "Graphlit API Client for TypeScript",
5
5
  "type": "module",
6
6
  "main": "./dist/client.js",