@chainfuse/types 1.6.8 → 1.6.9

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -80,6 +80,13 @@ export declare const workersAiCatalog: {
80
80
  readonly beta: true;
81
81
  readonly context_window: 4096;
82
82
  };
83
+ }, {
84
+ readonly id: "cc80437b-9a8d-4f1a-9c77-9aaf0d226922";
85
+ readonly source: 1;
86
+ readonly name: "@cf/meta/llama-guard-3-8b";
87
+ readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
88
+ readonly tags: readonly [];
89
+ readonly properties: {};
83
90
  }, {
84
91
  readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
85
92
  readonly source: 1;
@@ -88,6 +88,14 @@ export const workersAiCatalog = {
88
88
  context_window: 4096,
89
89
  },
90
90
  },
91
+ {
92
+ id: 'cc80437b-9a8d-4f1a-9c77-9aaf0d226922',
93
+ source: 1,
94
+ name: '@cf/meta/llama-guard-3-8b',
95
+ description: 'Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.',
96
+ tags: [],
97
+ properties: {},
98
+ },
91
99
  {
92
100
  id: 'ca54bcd6-0d98-4739-9b3b-5c8b4402193d',
93
101
  source: 1,
@@ -196,6 +196,13 @@ export declare const possibilities_mc_safety: readonly [...{
196
196
  readonly beta: true;
197
197
  readonly context_window: 4096;
198
198
  };
199
+ }, {
200
+ readonly id: "cc80437b-9a8d-4f1a-9c77-9aaf0d226922";
201
+ readonly source: 1;
202
+ readonly name: "@cf/meta/llama-guard-3-8b";
203
+ readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
204
+ readonly tags: readonly [];
205
+ readonly properties: {};
199
206
  }, {
200
207
  readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
201
208
  readonly source: 1;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@chainfuse/types",
3
- "version": "1.6.8",
3
+ "version": "1.6.9",
4
4
  "description": "",
5
5
  "author": "ChainFuse",
6
6
  "homepage": "https://github.com/ChainFuse/packages/tree/main/packages/types#readme",
@@ -89,5 +89,5 @@
89
89
  "@cloudflare/workers-types": "^4.20250224.0",
90
90
  "@types/json-schema": "^7.0.15"
91
91
  },
92
- "gitHead": "ab7a78396cb7871e1b4b78f9ffd7a0973839964e"
92
+ "gitHead": "9d842a11f77e62dab485f3965c724827b44f688f"
93
93
  }