@hyperspaceng/neural-ai 0.65.1 → 0.65.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3015,7 +3015,7 @@ export const MODELS = {
3015
3015
  },
3016
3016
  "gpt-5.4-mini": {
3017
3017
  id: "gpt-5.4-mini",
3018
- name: "GPT-5.4 mini",
3018
+ name: "GPT-5.4 Mini",
3019
3019
  api: "openai-responses",
3020
3020
  provider: "github-copilot",
3021
3021
  baseUrl: "https://api.individual.githubcopilot.com",
@@ -3477,8 +3477,8 @@ export const MODELS = {
3477
3477
  contextWindow: 131072,
3478
3478
  maxTokens: 8192,
3479
3479
  },
3480
- "gemma-4-26b": {
3481
- id: "gemma-4-26b",
3480
+ "gemma-4-26b-it": {
3481
+ id: "gemma-4-26b-it",
3482
3482
  name: "Gemma 4 26B",
3483
3483
  api: "google-generative-ai",
3484
3484
  provider: "google",
@@ -3494,8 +3494,8 @@ export const MODELS = {
3494
3494
  contextWindow: 256000,
3495
3495
  maxTokens: 8192,
3496
3496
  },
3497
- "gemma-4-31b": {
3498
- id: "gemma-4-31b",
3497
+ "gemma-4-31b-it": {
3498
+ id: "gemma-4-31b-it",
3499
3499
  name: "Gemma 4 31B",
3500
3500
  api: "google-generative-ai",
3501
3501
  provider: "google",
@@ -13943,9 +13943,9 @@ export const MODELS = {
13943
13943
  reasoning: true,
13944
13944
  input: ["text"],
13945
13945
  cost: {
13946
- input: 0.6,
13947
- output: 2.2,
13948
- cacheRead: 0.11,
13946
+ input: 0,
13947
+ output: 0,
13948
+ cacheRead: 0,
13949
13949
  cacheWrite: 0,
13950
13950
  },
13951
13951
  contextWindow: 131072,
@@ -13961,9 +13961,9 @@ export const MODELS = {
13961
13961
  reasoning: true,
13962
13962
  input: ["text"],
13963
13963
  cost: {
13964
- input: 0.2,
13965
- output: 1.1,
13966
- cacheRead: 0.03,
13964
+ input: 0,
13965
+ output: 0,
13966
+ cacheRead: 0,
13967
13967
  cacheWrite: 0,
13968
13968
  },
13969
13969
  contextWindow: 131072,
@@ -13997,8 +13997,8 @@ export const MODELS = {
13997
13997
  reasoning: true,
13998
13998
  input: ["text", "image"],
13999
13999
  cost: {
14000
- input: 0.6,
14001
- output: 1.8,
14000
+ input: 0,
14001
+ output: 0,
14002
14002
  cacheRead: 0,
14003
14003
  cacheWrite: 0,
14004
14004
  },
@@ -14015,9 +14015,9 @@ export const MODELS = {
14015
14015
  reasoning: true,
14016
14016
  input: ["text"],
14017
14017
  cost: {
14018
- input: 0.6,
14019
- output: 2.2,
14020
- cacheRead: 0.11,
14018
+ input: 0,
14019
+ output: 0,
14020
+ cacheRead: 0,
14021
14021
  cacheWrite: 0,
14022
14022
  },
14023
14023
  contextWindow: 204800,
@@ -14033,8 +14033,8 @@ export const MODELS = {
14033
14033
  reasoning: true,
14034
14034
  input: ["text", "image"],
14035
14035
  cost: {
14036
- input: 0.3,
14037
- output: 0.9,
14036
+ input: 0,
14037
+ output: 0,
14038
14038
  cacheRead: 0,
14039
14039
  cacheWrite: 0,
14040
14040
  },
@@ -14051,9 +14051,9 @@ export const MODELS = {
14051
14051
  reasoning: true,
14052
14052
  input: ["text"],
14053
14053
  cost: {
14054
- input: 0.6,
14055
- output: 2.2,
14056
- cacheRead: 0.11,
14054
+ input: 0,
14055
+ output: 0,
14056
+ cacheRead: 0,
14057
14057
  cacheWrite: 0,
14058
14058
  },
14059
14059
  contextWindow: 204800,
@@ -14105,9 +14105,9 @@ export const MODELS = {
14105
14105
  reasoning: true,
14106
14106
  input: ["text"],
14107
14107
  cost: {
14108
- input: 1,
14109
- output: 3.2,
14110
- cacheRead: 0.2,
14108
+ input: 0,
14109
+ output: 0,
14110
+ cacheRead: 0,
14111
14111
  cacheWrite: 0,
14112
14112
  },
14113
14113
  contextWindow: 204800,
@@ -14123,9 +14123,27 @@ export const MODELS = {
14123
14123
  reasoning: true,
14124
14124
  input: ["text"],
14125
14125
  cost: {
14126
- input: 1.2,
14127
- output: 4,
14128
- cacheRead: 0.24,
14126
+ input: 0,
14127
+ output: 0,
14128
+ cacheRead: 0,
14129
+ cacheWrite: 0,
14130
+ },
14131
+ contextWindow: 200000,
14132
+ maxTokens: 131072,
14133
+ },
14134
+ "glm-5.1": {
14135
+ id: "glm-5.1",
14136
+ name: "GLM-5.1",
14137
+ api: "openai-completions",
14138
+ provider: "zai",
14139
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
14140
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14141
+ reasoning: true,
14142
+ input: ["text"],
14143
+ cost: {
14144
+ input: 0,
14145
+ output: 0,
14146
+ cacheRead: 0,
14129
14147
  cacheWrite: 0,
14130
14148
  },
14131
14149
  contextWindow: 200000,
@@ -14141,9 +14159,9 @@ export const MODELS = {
14141
14159
  reasoning: true,
14142
14160
  input: ["text", "image"],
14143
14161
  cost: {
14144
- input: 1.2,
14145
- output: 4,
14146
- cacheRead: 0.24,
14162
+ input: 0,
14163
+ output: 0,
14164
+ cacheRead: 0,
14147
14165
  cacheWrite: 0,
14148
14166
  },
14149
14167
  contextWindow: 200000,