@chainfuse/types 4.2.16 → 4.2.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -435,8 +435,12 @@ export declare const azureCatalog: readonly [{
435
435
  readonly name: "gpt-4.1-nano";
436
436
  }, {
437
437
  readonly name: "gpt-4o";
438
+ readonly inputTokenCost: 0.00000275;
439
+ readonly outputTokenCost: 0.000011;
438
440
  }, {
439
441
  readonly name: "gpt-4o-mini";
442
+ readonly inputTokenCost: 1.65e-7;
443
+ readonly outputTokenCost: 6.6e-7;
440
444
  }, {
441
445
  readonly name: "gpt-5";
442
446
  }, {
@@ -445,10 +449,14 @@ export declare const azureCatalog: readonly [{
445
449
  readonly name: "gpt-5-nano";
446
450
  }, {
447
451
  readonly name: "o1";
452
+ readonly inputTokenCost: 0.0000165;
453
+ readonly outputTokenCost: 0.000066;
448
454
  }, {
449
455
  readonly name: "o3";
450
456
  }, {
451
457
  readonly name: "o3-mini";
458
+ readonly inputTokenCost: 0.00000121;
459
+ readonly outputTokenCost: 0.00000484;
452
460
  }, {
453
461
  readonly name: "o4-mini";
454
462
  }];
@@ -823,6 +831,7 @@ export declare const azureCatalog: readonly [{
823
831
  readonly imageModelAvailability: readonly [];
824
832
  readonly textEmbeddingModelAvailability: readonly [{
825
833
  readonly name: "text-embedding-3-large";
834
+ readonly tokenCost: 1.72e-7;
826
835
  }];
827
836
  }, {
828
837
  readonly id: "OpenAi-UK-London";
@@ -583,9 +583,13 @@ export const azureCatalog = [
583
583
  },
584
584
  {
585
585
  name: 'gpt-4o',
586
+ inputTokenCost: 0.00000275,
587
+ outputTokenCost: 0.000011,
586
588
  },
587
589
  {
588
590
  name: 'gpt-4o-mini',
591
+ inputTokenCost: 1.65e-7,
592
+ outputTokenCost: 6.6e-7,
589
593
  },
590
594
  {
591
595
  name: 'gpt-5',
@@ -598,12 +602,16 @@ export const azureCatalog = [
598
602
  },
599
603
  {
600
604
  name: 'o1',
605
+ inputTokenCost: 0.0000165,
606
+ outputTokenCost: 0.000066,
601
607
  },
602
608
  {
603
609
  name: 'o3',
604
610
  },
605
611
  {
606
612
  name: 'o3-mini',
613
+ inputTokenCost: 0.00000121,
614
+ outputTokenCost: 0.00000484,
607
615
  },
608
616
  {
609
617
  name: 'o4-mini',
@@ -1112,6 +1120,7 @@ export const azureCatalog = [
1112
1120
  textEmbeddingModelAvailability: [
1113
1121
  {
1114
1122
  name: 'text-embedding-3-large',
1123
+ tokenCost: 1.72e-7,
1115
1124
  },
1116
1125
  ],
1117
1126
  },
@@ -1,3 +1,3 @@
1
- import type { azureCatalog } from './catalog.js';
2
- export type AzureChatModels = (typeof azureCatalog)[number]['languageModelAvailability'][number]['name'];
3
- export type AzureEmbeddingModels = (typeof azureCatalog)[number]['textEmbeddingModelAvailability'][number]['name'];
1
+ export declare const AzureChatModels: ("o1" | "o3" | "gpt-4.1" | "gpt-4.1-mini" | "gpt-4.1-nano" | "gpt-4o" | "gpt-4o-mini" | "gpt-5" | "gpt-5-mini" | "gpt-5-nano" | "o3-mini" | "o4-mini" | "gpt-5.1" | "o1-mini")[];
2
+ export declare const AzureImageModels: "gpt-image-1"[];
3
+ export declare const AzureEmbeddingModels: ("text-embedding-3-large" | "text-embedding-3-small")[];
@@ -1 +1,4 @@
1
- export {};
1
+ import { azureCatalog } from './catalog.js';
2
+ export const AzureChatModels = azureCatalog.map(({ languageModelAvailability }) => languageModelAvailability.map(({ name }) => name)).flat();
3
+ export const AzureImageModels = azureCatalog.map(({ imageModelAvailability }) => imageModelAvailability.map(({ name }) => name)).flat();
4
+ export const AzureEmbeddingModels = azureCatalog.map(({ textEmbeddingModelAvailability }) => textEmbeddingModelAvailability.map(({ name }) => name)).flat();
@@ -104,6 +104,7 @@ export declare const ZodLanguageModelValues: zm.ZodMiniEnum<{
104
104
  "workersai:@hf/thebloke/mistral-7b-instruct-v0.1-awq": "workersai:@hf/thebloke/mistral-7b-instruct-v0.1-awq";
105
105
  "workersai:@cf/qwen/qwen1.5-7b-chat-awq": "workersai:@cf/qwen/qwen1.5-7b-chat-awq";
106
106
  "workersai:@cf/meta/llama-3.2-1b-instruct": "workersai:@cf/meta/llama-3.2-1b-instruct";
107
+ "workersai:@cf/zai-org/glm-4.7-flash": "workersai:@cf/zai-org/glm-4.7-flash";
107
108
  "workersai:@hf/thebloke/llama-2-13b-chat-awq": "workersai:@hf/thebloke/llama-2-13b-chat-awq";
108
109
  "workersai:@hf/thebloke/deepseek-coder-6.7b-base-awq": "workersai:@hf/thebloke/deepseek-coder-6.7b-base-awq";
109
110
  "workersai:@cf/meta-llama/llama-2-7b-chat-hf-lora": "workersai:@cf/meta-llama/llama-2-7b-chat-hf-lora";
@@ -188,6 +189,7 @@ export declare const ZodLanguageModelValues4: z4.ZodEnum<{
188
189
  "workersai:@hf/thebloke/mistral-7b-instruct-v0.1-awq": "workersai:@hf/thebloke/mistral-7b-instruct-v0.1-awq";
189
190
  "workersai:@cf/qwen/qwen1.5-7b-chat-awq": "workersai:@cf/qwen/qwen1.5-7b-chat-awq";
190
191
  "workersai:@cf/meta/llama-3.2-1b-instruct": "workersai:@cf/meta/llama-3.2-1b-instruct";
192
+ "workersai:@cf/zai-org/glm-4.7-flash": "workersai:@cf/zai-org/glm-4.7-flash";
191
193
  "workersai:@hf/thebloke/llama-2-13b-chat-awq": "workersai:@hf/thebloke/llama-2-13b-chat-awq";
192
194
  "workersai:@hf/thebloke/deepseek-coder-6.7b-base-awq": "workersai:@hf/thebloke/deepseek-coder-6.7b-base-awq";
193
195
  "workersai:@cf/meta-llama/llama-2-7b-chat-hf-lora": "workersai:@cf/meta-llama/llama-2-7b-chat-hf-lora";
@@ -361,6 +361,26 @@ export declare const workersAiCatalog: {
361
361
  readonly context_window: 60000;
362
362
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
363
363
  };
364
+ }, {
365
+ readonly id: "86b3e51a-4b05-43fa-a403-0f27821919d2";
366
+ readonly source: 1;
367
+ readonly name: "@cf/zai-org/glm-4.7-flash";
368
+ readonly description: "GLM-4.7-Flash is a fast and efficient multilingual text generation model with a 131,072 token context window. Optimized for dialogue, instruction-following, and multi-turn tool calling across 100+ languages.";
369
+ readonly created_at: "2026-01-28 16:04:39.346";
370
+ readonly tags: readonly [];
371
+ readonly properties: {
372
+ readonly context_window: 131072;
373
+ readonly price: readonly [{
374
+ readonly unit: "per M input tokens";
375
+ readonly price: 0.06;
376
+ readonly currency: "USD";
377
+ }, {
378
+ readonly unit: "per M output tokens";
379
+ readonly price: 0.4;
380
+ readonly currency: "USD";
381
+ }];
382
+ readonly function_calling: true;
383
+ };
364
384
  }, {
365
385
  readonly id: "85c5a3c6-24b0-45e7-b23a-023182578822";
366
386
  readonly source: 2;
@@ -414,6 +414,30 @@ export const workersAiCatalog = {
414
414
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE',
415
415
  },
416
416
  },
417
+ {
418
+ id: '86b3e51a-4b05-43fa-a403-0f27821919d2',
419
+ source: 1,
420
+ name: '@cf/zai-org/glm-4.7-flash',
421
+ description: 'GLM-4.7-Flash is a fast and efficient multilingual text generation model with a 131,072 token context window. Optimized for dialogue, instruction-following, and multi-turn tool calling across 100+ languages.',
422
+ created_at: '2026-01-28 16:04:39.346',
423
+ tags: [],
424
+ properties: {
425
+ context_window: 131072,
426
+ price: [
427
+ {
428
+ unit: 'per M input tokens',
429
+ price: 0.06,
430
+ currency: 'USD',
431
+ },
432
+ {
433
+ unit: 'per M output tokens',
434
+ price: 0.4,
435
+ currency: 'USD',
436
+ },
437
+ ],
438
+ function_calling: true,
439
+ },
440
+ },
417
441
  {
418
442
  id: '85c5a3c6-24b0-45e7-b23a-023182578822',
419
443
  source: 2,
@@ -11,9 +11,11 @@ type cloudflareFilteredModelPossibilitiesRaw<M extends cloudflareModelTypes = cl
11
11
  } ? Model : never) : never;
12
12
  type cloudflareFilteredModelPossibilities<M extends cloudflareModelTypes = cloudflareModelTypes, K extends cloudflareModelPossibilitiesProperties<M> = cloudflareModelPossibilitiesProperties<M>, V extends cloudflareModelPossibilitiesRaw<M>['properties'][K] = any> = cloudflareFilteredModelPossibilitiesRaw<M, K, V>['name'];
13
13
  export declare const enabledCloudflareLlmEmbeddingProviders: cloudflareModelPossibilities<'Text Embeddings'>[];
14
- export declare const enabledCloudflareLlmProviders: cloudflareModelPossibilities<'Text Generation'>[];
15
- export declare const enabledCloudflareLlmFunctionProviders: ("@cf/meta/llama-3.3-70b-instruct-fp8-fast" | "@cf/ibm-granite/granite-4.0-h-micro" | "@hf/nousresearch/hermes-2-pro-mistral-7b" | "@cf/qwen/qwen3-30b-a3b-fp8" | "@cf/mistralai/mistral-small-3.1-24b-instruct" | "@cf/meta/llama-4-scout-17b-16e-instruct")[];
16
14
  export declare const enabledCloudflareLlmImageProviders: cloudflareModelPossibilities<'Text-to-Image'>[];
15
+ export declare const enabledCloudflareLlmProviders: cloudflareModelPossibilities<'Text Generation'>[];
16
+ export declare const enabledCloudflareLlmFunctionProviders: ("@cf/zai-org/glm-4.7-flash" | "@cf/meta/llama-3.3-70b-instruct-fp8-fast" | "@cf/ibm-granite/granite-4.0-h-micro" | "@hf/nousresearch/hermes-2-pro-mistral-7b" | "@cf/qwen/qwen3-30b-a3b-fp8" | "@cf/mistralai/mistral-small-3.1-24b-instruct" | "@cf/meta/llama-4-scout-17b-16e-instruct")[];
17
+ export declare const enabledCloudflareLlmSpeechProviders: ("@cf/myshell-ai/melotts" | "@cf/deepgram/aura-2-es" | "@cf/deepgram/aura-1" | "@cf/deepgram/aura-2-en")[];
18
+ export declare const enabledCloudflareLlmTranscriptionProviders: ("@cf/openai/whisper" | "@cf/deepgram/flux" | "@cf/deepgram/nova-3" | "@cf/openai/whisper-tiny-en" | "@cf/openai/whisper-large-v3-turbo")[];
17
19
  export type CloudflareModelsEnum<M extends cloudflareModelTypes = cloudflareModelTypes> = {
18
20
  [K in cloudflareModelPossibilities<M>]: `workersai:${K}`;
19
21
  };
@@ -2,6 +2,8 @@ import { workersAiCatalog } from './catalog.js';
2
2
  // const enabledCloudflareLlmSummaryProviders: cloudflareModelPossibilities<'Summarization'>[] = workersAiCatalog.modelGroups.Summarization.models.map((model) => model.name);
3
3
  // const enabledCloudflareLlmClassificationProviders: cloudflareModelPossibilities<'Text Classification'>[] = workersAiCatalog.modelGroups['Text Classification'].models.map((model) => model.name);
4
4
  export const enabledCloudflareLlmEmbeddingProviders = workersAiCatalog.modelGroups['Text Embeddings'].models.map((model) => model.name);
5
+ export const enabledCloudflareLlmImageProviders = workersAiCatalog.modelGroups['Text-to-Image'].models.map((model) => model.name);
5
6
  export const enabledCloudflareLlmProviders = workersAiCatalog.modelGroups['Text Generation'].models.map((model) => model.name);
6
7
  export const enabledCloudflareLlmFunctionProviders = workersAiCatalog.modelGroups['Text Generation'].models.filter((model) => 'function_calling' in model.properties && model.properties.function_calling).map((model) => model.name);
7
- export const enabledCloudflareLlmImageProviders = workersAiCatalog.modelGroups['Text-to-Image'].models.map((model) => model.name);
8
+ export const enabledCloudflareLlmSpeechProviders = workersAiCatalog.modelGroups['Text-to-Speech'].models.map((model) => model.name);
9
+ export const enabledCloudflareLlmTranscriptionProviders = workersAiCatalog.modelGroups['Automatic Speech Recognition'].models.map((model) => model.name);
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@chainfuse/types",
3
- "version": "4.2.16",
3
+ "version": "4.2.17",
4
4
  "description": "",
5
5
  "author": "ChainFuse",
6
6
  "homepage": "https://github.com/ChainFuse/packages/tree/main/packages/types#readme",
@@ -98,8 +98,8 @@
98
98
  "zod": "^4.3.6"
99
99
  },
100
100
  "devDependencies": {
101
- "@cloudflare/workers-types": "^4.20260212.0",
101
+ "@cloudflare/workers-types": "^4.20260214.0",
102
102
  "@types/validator": "^13.15.10"
103
103
  },
104
- "gitHead": "516f3c7d628c7c7793c9dbf228a35afbf22b3e69"
104
+ "gitHead": "688aa912c143b60b0e663b8125531b0e8d649666"
105
105
  }