hedgequantx 2.5.27 → 2.5.28

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "hedgequantx",
3
- "version": "2.5.27",
3
+ "version": "2.5.28",
4
4
  "description": "HedgeQuantX - Prop Futures Trading CLI",
5
5
  "main": "src/app.js",
6
6
  "bin": {
@@ -1,6 +1,9 @@
1
1
  /**
2
2
  * AI Providers Configuration
3
3
  * Each provider has connection options (API Key, Plans, etc.)
4
+ *
5
+ * IMPORTANT: models arrays are empty - models MUST be fetched from real APIs
6
+ * No hardcoded model lists allowed (see RULES.md)
4
7
  */
5
8
 
6
9
  const PROVIDERS = {
@@ -66,8 +69,8 @@ const PROVIDERS = {
66
69
  name: 'OPENAI (GPT-4)',
67
70
  description: 'Direct connection to GPT-4',
68
71
  category: 'direct',
69
- models: ['gpt-4o', 'gpt-4-turbo', 'gpt-4o-mini', 'gpt-3.5-turbo'],
70
- defaultModel: 'gpt-4o',
72
+ models: [], // Fetched from API at runtime
73
+ defaultModel: null, // Will use first model from API
71
74
  options: [
72
75
  {
73
76
  id: 'api_key',
@@ -110,8 +113,8 @@ const PROVIDERS = {
110
113
  name: 'DEEPSEEK',
111
114
  description: 'Very cheap & capable',
112
115
  category: 'direct',
113
- models: ['deepseek-chat', 'deepseek-coder'],
114
- defaultModel: 'deepseek-chat',
116
+ models: [], // Fetched from API at runtime
117
+ defaultModel: null, // Will use first model from API
115
118
  options: [
116
119
  {
117
120
  id: 'api_key',
@@ -132,8 +135,8 @@ const PROVIDERS = {
132
135
  name: 'GROQ',
133
136
  description: 'Ultra fast inference',
134
137
  category: 'direct',
135
- models: ['llama-3.3-70b-versatile', 'llama-3.1-8b-instant', 'mixtral-8x7b-32768'],
136
- defaultModel: 'llama-3.3-70b-versatile',
138
+ models: [], // Fetched from API at runtime
139
+ defaultModel: null, // Will use first model from API
137
140
  options: [
138
141
  {
139
142
  id: 'api_key',
@@ -155,8 +158,8 @@ const PROVIDERS = {
155
158
  name: 'GROK (XAI)',
156
159
  description: 'Elon Musk\'s Grok AI',
157
160
  category: 'direct',
158
- models: ['grok-beta', 'grok-2'],
159
- defaultModel: 'grok-beta',
161
+ models: [], // Fetched from API at runtime
162
+ defaultModel: null, // Will use first model from API
160
163
  options: [
161
164
  {
162
165
  id: 'api_key',
@@ -177,8 +180,8 @@ const PROVIDERS = {
177
180
  name: 'MISTRAL',
178
181
  description: 'European AI leader',
179
182
  category: 'direct',
180
- models: ['mistral-large-latest', 'mistral-medium', 'mistral-small'],
181
- defaultModel: 'mistral-large-latest',
183
+ models: [], // Fetched from API at runtime
184
+ defaultModel: null, // Will use first model from API
182
185
  options: [
183
186
  {
184
187
  id: 'api_key',
@@ -199,8 +202,8 @@ const PROVIDERS = {
199
202
  name: 'PERPLEXITY',
200
203
  description: 'Real-time web search AI',
201
204
  category: 'direct',
202
- models: ['llama-3.1-sonar-large-128k-online', 'llama-3.1-sonar-small-128k-online', 'llama-3.1-sonar-huge-128k-online'],
203
- defaultModel: 'llama-3.1-sonar-large-128k-online',
205
+ models: [], // Fetched from API at runtime
206
+ defaultModel: null, // Will use first model from API
204
207
  options: [
205
208
  {
206
209
  id: 'api_key',
@@ -222,8 +225,8 @@ const PROVIDERS = {
222
225
  name: 'TOGETHER AI',
223
226
  description: 'Open source models, fast & cheap',
224
227
  category: 'direct',
225
- models: ['meta-llama/Llama-3.3-70B-Instruct-Turbo', 'mistralai/Mixtral-8x22B-Instruct-v0.1', 'Qwen/Qwen2.5-72B-Instruct-Turbo'],
226
- defaultModel: 'meta-llama/Llama-3.3-70B-Instruct-Turbo',
228
+ models: [], // Fetched from API at runtime
229
+ defaultModel: null, // Will use first model from API
227
230
  options: [
228
231
  {
229
232
  id: 'api_key',
@@ -246,8 +249,8 @@ const PROVIDERS = {
246
249
  name: 'QWEN (ALIBABA)',
247
250
  description: 'Alibaba\'s top AI model',
248
251
  category: 'direct',
249
- models: ['qwen-turbo', 'qwen-plus', 'qwen-max'],
250
- defaultModel: 'qwen-plus',
252
+ models: [], // Fetched from API at runtime
253
+ defaultModel: null, // Will use first model from API
251
254
  options: [
252
255
  {
253
256
  id: 'api_key',
@@ -269,8 +272,8 @@ const PROVIDERS = {
269
272
  name: 'MOONSHOT (KIMI)',
270
273
  description: '200K context window',
271
274
  category: 'direct',
272
- models: ['moonshot-v1-8k', 'moonshot-v1-32k', 'moonshot-v1-128k'],
273
- defaultModel: 'moonshot-v1-32k',
275
+ models: [], // Fetched from API at runtime
276
+ defaultModel: null, // Will use first model from API
274
277
  options: [
275
278
  {
276
279
  id: 'api_key',
@@ -292,8 +295,8 @@ const PROVIDERS = {
292
295
  name: '01.AI (YI)',
293
296
  description: 'Yi models by Kai-Fu Lee',
294
297
  category: 'direct',
295
- models: ['yi-large', 'yi-medium', 'yi-spark'],
296
- defaultModel: 'yi-large',
298
+ models: [], // Fetched from API at runtime
299
+ defaultModel: null, // Will use first model from API
297
300
  options: [
298
301
  {
299
302
  id: 'api_key',
@@ -315,8 +318,8 @@ const PROVIDERS = {
315
318
  name: 'ZHIPU AI (GLM)',
316
319
  description: 'ChatGLM models',
317
320
  category: 'direct',
318
- models: ['glm-4-plus', 'glm-4', 'glm-4-flash'],
319
- defaultModel: 'glm-4',
321
+ models: [], // Fetched from API at runtime
322
+ defaultModel: null, // Will use first model from API
320
323
  options: [
321
324
  {
322
325
  id: 'api_key',
@@ -338,15 +341,15 @@ const PROVIDERS = {
338
341
  name: 'BAICHUAN',
339
342
  description: 'Multilingual AI model',
340
343
  category: 'direct',
341
- models: ['Baichuan4', 'Baichuan3-Turbo', 'Baichuan2-Turbo'],
342
- defaultModel: 'Baichuan4',
344
+ models: [], // Fetched from API at runtime
345
+ defaultModel: null, // Will use first model from API
343
346
  options: [
344
347
  {
345
348
  id: 'api_key',
346
349
  label: 'API KEY',
347
350
  description: [
348
351
  'Get key at platform.baichuan-ai.com',
349
- 'Best for Chinese content',
352
+ 'Strong multilingual support',
350
353
  'Competitive pricing'
351
354
  ],
352
355
  fields: ['apiKey'],
@@ -362,8 +365,8 @@ const PROVIDERS = {
362
365
  name: 'OLLAMA (LOCAL - FREE)',
363
366
  description: '100% free, runs locally',
364
367
  category: 'local',
365
- models: ['llama3', 'llama3.1', 'mistral', 'codellama', 'phi3', 'gemma2'],
366
- defaultModel: 'llama3.1',
368
+ models: [], // Fetched from local API at runtime
369
+ defaultModel: null, // Will use first model from API
367
370
  options: [
368
371
  {
369
372
  id: 'local',
@@ -385,8 +388,8 @@ const PROVIDERS = {
385
388
  name: 'LM STUDIO (LOCAL - FREE)',
386
389
  description: 'Local with GUI',
387
390
  category: 'local',
388
- models: [],
389
- defaultModel: '',
391
+ models: [], // Fetched from local API at runtime
392
+ defaultModel: null, // Will use first model from API
390
393
  options: [
391
394
  {
392
395
  id: 'local',
@@ -409,8 +412,8 @@ const PROVIDERS = {
409
412
  name: 'CUSTOM ENDPOINT',
410
413
  description: 'Any OpenAI-compatible API',
411
414
  category: 'custom',
412
- models: [],
413
- defaultModel: '',
415
+ models: [], // User must specify model
416
+ defaultModel: null,
414
417
  options: [
415
418
  {
416
419
  id: 'custom',