@chainfuse/types 2.9.1 → 2.10.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -367,6 +367,7 @@ export declare const workersAiCatalog: {
367
367
  readonly created_at: "2024-12-06 17:09:18.338";
368
368
  readonly tags: readonly [];
369
369
  readonly properties: {
370
+ readonly async_queue: true;
370
371
  readonly context_window: 24000;
371
372
  readonly price: readonly [{
372
373
  readonly unit: "per M input tokens";
@@ -403,6 +404,26 @@ export declare const workersAiCatalog: {
403
404
  readonly context_window: 4096;
404
405
  readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ";
405
406
  };
407
+ }, {
408
+ readonly id: "51b71d5b-8bc0-4489-a107-95e542b69914";
409
+ readonly source: 1;
410
+ readonly name: "@cf/qwen/qwen2.5-coder-32b-instruct";
411
+ readonly description: "Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). As of now, Qwen2.5-Coder has covered six mainstream model sizes, 0.5, 1.5, 3, 7, 14, 32 billion parameters, to meet the needs of different developers. Qwen2.5-Coder brings the following improvements upon CodeQwen1.5:";
412
+ readonly created_at: "2025-02-27 00:31:43.829";
413
+ readonly tags: readonly [];
414
+ readonly properties: {
415
+ readonly context_window: 32768;
416
+ readonly price: readonly [{
417
+ readonly unit: "per M input tokens";
418
+ readonly price: 0.66;
419
+ readonly currency: "USD";
420
+ }, {
421
+ readonly unit: "per M output tokens";
422
+ readonly price: 1;
423
+ readonly currency: "USD";
424
+ }];
425
+ readonly lora: true;
426
+ };
406
427
  }, {
407
428
  readonly id: "4c3a544e-da47-4336-9cea-c7cbfab33f16";
408
429
  readonly source: 1;
@@ -517,6 +538,25 @@ export declare const workersAiCatalog: {
517
538
  readonly context_window: 32000;
518
539
  readonly info: "https://huggingface.co/qwen/qwen1.5-1.8b-chat";
519
540
  };
541
+ }, {
542
+ readonly id: "31690291-ebdc-4f98-bcfc-a44844e215b7";
543
+ readonly source: 1;
544
+ readonly name: "@cf/mistralai/mistral-small-3.1-24b-instruct";
545
+ readonly description: "Building upon Mistral Small 3 (2501), Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance. With 24 billion parameters, this model achieves top-tier capabilities in both text and vision tasks.";
546
+ readonly created_at: "2025-03-18 03:28:37.890";
547
+ readonly tags: readonly [];
548
+ readonly properties: {
549
+ readonly context_window: 128000;
550
+ readonly price: readonly [{
551
+ readonly unit: "per M input tokens";
552
+ readonly price: 0.35;
553
+ readonly currency: "USD";
554
+ }, {
555
+ readonly unit: "per M output tokens";
556
+ readonly price: 0.56;
557
+ readonly currency: "USD";
558
+ }];
559
+ };
520
560
  }, {
521
561
  readonly id: "31097538-a3ff-4e6e-bb56-ad0e1f428b61";
522
562
  readonly source: 1;
@@ -651,6 +691,46 @@ export declare const workersAiCatalog: {
651
691
  }];
652
692
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama4/LICENSE";
653
693
  };
694
+ }, {
695
+ readonly id: "053d5ac0-861b-4d3b-8501-e58d00417ef8";
696
+ readonly source: 1;
697
+ readonly name: "@cf/google/gemma-3-12b-it";
698
+ readonly description: "Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Gemma 3 models are multimodal, handling text and image input and generating text output, with a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions.";
699
+ readonly created_at: "2025-03-18 03:58:02.423";
700
+ readonly tags: readonly [];
701
+ readonly properties: {
702
+ readonly context_window: 1280000;
703
+ readonly price: readonly [{
704
+ readonly unit: "per M input tokens";
705
+ readonly price: 0.35;
706
+ readonly currency: "USD";
707
+ }, {
708
+ readonly unit: "per M output tokens";
709
+ readonly price: 0.56;
710
+ readonly currency: "USD";
711
+ }];
712
+ readonly lora: true;
713
+ };
714
+ }, {
715
+ readonly id: "02c16efa-29f5-4304-8e6c-3d188889f875";
716
+ readonly source: 1;
717
+ readonly name: "@cf/qwen/qwq-32b";
718
+ readonly description: "QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.";
719
+ readonly created_at: "2025-03-05 21:52:40.974";
720
+ readonly tags: readonly [];
721
+ readonly properties: {
722
+ readonly context_window: 24000;
723
+ readonly price: readonly [{
724
+ readonly unit: "per M input tokens";
725
+ readonly price: 0.66;
726
+ readonly currency: "USD";
727
+ }, {
728
+ readonly unit: "per M output tokens";
729
+ readonly price: 1;
730
+ readonly currency: "USD";
731
+ }];
732
+ readonly lora: true;
733
+ };
654
734
  }];
655
735
  };
656
736
  readonly 'Text Embeddings': {
@@ -416,6 +416,7 @@ export const workersAiCatalog = {
416
416
  created_at: '2024-12-06 17:09:18.338',
417
417
  tags: [],
418
418
  properties: {
419
+ async_queue: true,
419
420
  context_window: 24000,
420
421
  price: [
421
422
  {
@@ -458,6 +459,30 @@ export const workersAiCatalog = {
458
459
  terms: 'https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ',
459
460
  },
460
461
  },
462
+ {
463
+ id: '51b71d5b-8bc0-4489-a107-95e542b69914',
464
+ source: 1,
465
+ name: '@cf/qwen/qwen2.5-coder-32b-instruct',
466
+ description: 'Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). As of now, Qwen2.5-Coder has covered six mainstream model sizes, 0.5, 1.5, 3, 7, 14, 32 billion parameters, to meet the needs of different developers. Qwen2.5-Coder brings the following improvements upon CodeQwen1.5:',
467
+ created_at: '2025-02-27 00:31:43.829',
468
+ tags: [],
469
+ properties: {
470
+ context_window: 32768,
471
+ price: [
472
+ {
473
+ unit: 'per M input tokens',
474
+ price: 0.66,
475
+ currency: 'USD',
476
+ },
477
+ {
478
+ unit: 'per M output tokens',
479
+ price: 1,
480
+ currency: 'USD',
481
+ },
482
+ ],
483
+ lora: true,
484
+ },
485
+ },
461
486
  {
462
487
  id: '4c3a544e-da47-4336-9cea-c7cbfab33f16',
463
488
  source: 1,
@@ -586,6 +611,29 @@ export const workersAiCatalog = {
586
611
  info: 'https://huggingface.co/qwen/qwen1.5-1.8b-chat',
587
612
  },
588
613
  },
614
+ {
615
+ id: '31690291-ebdc-4f98-bcfc-a44844e215b7',
616
+ source: 1,
617
+ name: '@cf/mistralai/mistral-small-3.1-24b-instruct',
618
+ description: 'Building upon Mistral Small 3 (2501), Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance. With 24 billion parameters, this model achieves top-tier capabilities in both text and vision tasks.',
619
+ created_at: '2025-03-18 03:28:37.890',
620
+ tags: [],
621
+ properties: {
622
+ context_window: 128000,
623
+ price: [
624
+ {
625
+ unit: 'per M input tokens',
626
+ price: 0.35,
627
+ currency: 'USD',
628
+ },
629
+ {
630
+ unit: 'per M output tokens',
631
+ price: 0.56,
632
+ currency: 'USD',
633
+ },
634
+ ],
635
+ },
636
+ },
589
637
  {
590
638
  id: '31097538-a3ff-4e6e-bb56-ad0e1f428b61',
591
639
  source: 1,
@@ -738,6 +786,54 @@ export const workersAiCatalog = {
738
786
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama4/LICENSE',
739
787
  },
740
788
  },
789
+ {
790
+ id: '053d5ac0-861b-4d3b-8501-e58d00417ef8',
791
+ source: 1,
792
+ name: '@cf/google/gemma-3-12b-it',
793
+ description: 'Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Gemma 3 models are multimodal, handling text and image input and generating text output, with a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions.',
794
+ created_at: '2025-03-18 03:58:02.423',
795
+ tags: [],
796
+ properties: {
797
+ context_window: 1280000,
798
+ price: [
799
+ {
800
+ unit: 'per M input tokens',
801
+ price: 0.35,
802
+ currency: 'USD',
803
+ },
804
+ {
805
+ unit: 'per M output tokens',
806
+ price: 0.56,
807
+ currency: 'USD',
808
+ },
809
+ ],
810
+ lora: true,
811
+ },
812
+ },
813
+ {
814
+ id: '02c16efa-29f5-4304-8e6c-3d188889f875',
815
+ source: 1,
816
+ name: '@cf/qwen/qwq-32b',
817
+ description: 'QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.',
818
+ created_at: '2025-03-05 21:52:40.974',
819
+ tags: [],
820
+ properties: {
821
+ context_window: 24000,
822
+ price: [
823
+ {
824
+ unit: 'per M input tokens',
825
+ price: 0.66,
826
+ currency: 'USD',
827
+ },
828
+ {
829
+ unit: 'per M output tokens',
830
+ price: 1,
831
+ currency: 'USD',
832
+ },
833
+ ],
834
+ lora: true,
835
+ },
836
+ },
741
837
  ],
742
838
  },
743
839
  'Text Embeddings': {
package/dist/index.d.ts CHANGED
@@ -50,7 +50,7 @@ export interface ExternallyResolvablePromise<T> {
50
50
  export type UndefinedProperties<T extends object> = {
51
51
  [P in keyof T]: undefined;
52
52
  };
53
- export type CustomLogCallback = (message?: any, ...optionalParams: any[]) => void;
53
+ export type CustomLogCallback = (message?: any, ...optionalParams: any[]) => PromiseLike<void> | void;
54
54
  export type CustomLoging = boolean | CustomLogCallback;
55
55
  /**
56
56
  * @link https://developers.cloudflare.com/durable-objects/reference/data-location/#restrict-durable-objects-to-a-jurisdiction
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@chainfuse/types",
3
- "version": "2.9.1",
3
+ "version": "2.10.0",
4
4
  "description": "",
5
5
  "author": "ChainFuse",
6
6
  "homepage": "https://github.com/ChainFuse/packages/tree/main/packages/types#readme",
@@ -97,5 +97,5 @@
97
97
  "@cloudflare/workers-types": "^4.20250410.0",
98
98
  "@types/validator": "^13.12.3"
99
99
  },
100
- "gitHead": "e7ec0b382add48d991bacd096563691bc74f9a46"
100
+ "gitHead": "c480462d073b3d89c915eb003b0ba789fe79d414"
101
101
  }