omegon 0.9.0 → 0.9.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (47) hide show
  1. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.d.ts +58 -105
  2. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.d.ts.map +1 -1
  3. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.js +98 -148
  4. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.js.map +1 -1
  5. package/node_modules/@styrene-lab/pi-ai/dist/utils/oauth/anthropic.d.ts.map +1 -1
  6. package/node_modules/@styrene-lab/pi-ai/dist/utils/oauth/anthropic.js +3 -7
  7. package/node_modules/@styrene-lab/pi-ai/dist/utils/oauth/anthropic.js.map +1 -1
  8. package/node_modules/@styrene-lab/pi-tui/dist/components/editor.d.ts.map +1 -1
  9. package/node_modules/@styrene-lab/pi-tui/dist/components/editor.js +4 -3
  10. package/node_modules/@styrene-lab/pi-tui/dist/components/editor.js.map +1 -1
  11. package/node_modules/@styrene-lab/pi-tui/dist/components/input.d.ts.map +1 -1
  12. package/node_modules/@styrene-lab/pi-tui/dist/components/input.js +9 -1
  13. package/node_modules/@styrene-lab/pi-tui/dist/components/input.js.map +1 -1
  14. package/node_modules/@styrene-lab/pi-tui/dist/index.d.ts +1 -1
  15. package/node_modules/@styrene-lab/pi-tui/dist/index.d.ts.map +1 -1
  16. package/node_modules/@styrene-lab/pi-tui/dist/index.js +1 -1
  17. package/node_modules/@styrene-lab/pi-tui/dist/index.js.map +1 -1
  18. package/node_modules/@styrene-lab/pi-tui/dist/keys.d.ts +2 -0
  19. package/node_modules/@styrene-lab/pi-tui/dist/keys.d.ts.map +1 -1
  20. package/node_modules/@styrene-lab/pi-tui/dist/keys.js +39 -0
  21. package/node_modules/@styrene-lab/pi-tui/dist/keys.js.map +1 -1
  22. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.d.ts +14 -0
  23. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.d.ts.map +1 -1
  24. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.js +21 -1
  25. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.js.map +1 -1
  26. package/node_modules/@styrene-lab/pi-tui/dist/terminal.d.ts +1 -0
  27. package/node_modules/@styrene-lab/pi-tui/dist/terminal.d.ts.map +1 -1
  28. package/node_modules/@styrene-lab/pi-tui/dist/terminal.js +40 -0
  29. package/node_modules/@styrene-lab/pi-tui/dist/terminal.js.map +1 -1
  30. package/package.json +5 -5
  31. package/scripts/check-vendor-dist.mjs +56 -0
  32. package/scripts/publish-pi-mono.sh +5 -22
  33. package/node_modules/cliui/CHANGELOG.md +0 -121
  34. package/node_modules/color-convert/CHANGELOG.md +0 -54
  35. package/node_modules/esprima/ChangeLog +0 -235
  36. package/node_modules/mz/HISTORY.md +0 -66
  37. package/node_modules/proper-lockfile/CHANGELOG.md +0 -108
  38. package/node_modules/source-map/CHANGELOG.md +0 -301
  39. package/node_modules/thenify/History.md +0 -11
  40. package/node_modules/thenify-all/History.md +0 -11
  41. package/node_modules/y18n/CHANGELOG.md +0 -100
  42. package/node_modules/yargs/CHANGELOG.md +0 -88
  43. package/node_modules/yargs-parser/CHANGELOG.md +0 -263
  44. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/LICENSE +0 -0
  45. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/README.md +0 -0
  46. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/index.d.ts +0 -0
  47. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/package.json +0 -0
@@ -5966,23 +5966,6 @@ export const MODELS = {
5966
5966
  contextWindow: 1048576,
5967
5967
  maxTokens: 65536,
5968
5968
  },
5969
- "gemini-3-pro": {
5970
- id: "gemini-3-pro",
5971
- name: "Gemini 3 Pro",
5972
- api: "google-generative-ai",
5973
- provider: "opencode",
5974
- baseUrl: "https://opencode.ai/zen/v1",
5975
- reasoning: true,
5976
- input: ["text", "image"],
5977
- cost: {
5978
- input: 2,
5979
- output: 12,
5980
- cacheRead: 0.2,
5981
- cacheWrite: 0,
5982
- },
5983
- contextWindow: 1048576,
5984
- maxTokens: 65536,
5985
- },
5986
5969
  "gemini-3.1-pro": {
5987
5970
  id: "gemini-3.1-pro",
5988
5971
  name: "Gemini 3.1 Pro Preview",
@@ -6000,40 +5983,6 @@ export const MODELS = {
6000
5983
  contextWindow: 1048576,
6001
5984
  maxTokens: 65536,
6002
5985
  },
6003
- "glm-4.6": {
6004
- id: "glm-4.6",
6005
- name: "GLM-4.6",
6006
- api: "openai-completions",
6007
- provider: "opencode",
6008
- baseUrl: "https://opencode.ai/zen/v1",
6009
- reasoning: true,
6010
- input: ["text"],
6011
- cost: {
6012
- input: 0.6,
6013
- output: 2.2,
6014
- cacheRead: 0.1,
6015
- cacheWrite: 0,
6016
- },
6017
- contextWindow: 204800,
6018
- maxTokens: 131072,
6019
- },
6020
- "glm-4.7": {
6021
- id: "glm-4.7",
6022
- name: "GLM-4.7",
6023
- api: "openai-completions",
6024
- provider: "opencode",
6025
- baseUrl: "https://opencode.ai/zen/v1",
6026
- reasoning: true,
6027
- input: ["text"],
6028
- cost: {
6029
- input: 0.6,
6030
- output: 2.2,
6031
- cacheRead: 0.1,
6032
- cacheWrite: 0,
6033
- },
6034
- contextWindow: 204800,
6035
- maxTokens: 131072,
6036
- },
6037
5986
  "glm-5": {
6038
5987
  id: "glm-5",
6039
5988
  name: "GLM-5",
@@ -6289,23 +6238,6 @@ export const MODELS = {
6289
6238
  contextWindow: 262144,
6290
6239
  maxTokens: 65536,
6291
6240
  },
6292
- "minimax-m2.1": {
6293
- id: "minimax-m2.1",
6294
- name: "MiniMax M2.1",
6295
- api: "openai-completions",
6296
- provider: "opencode",
6297
- baseUrl: "https://opencode.ai/zen/v1",
6298
- reasoning: true,
6299
- input: ["text"],
6300
- cost: {
6301
- input: 0.3,
6302
- output: 1.2,
6303
- cacheRead: 0.1,
6304
- cacheWrite: 0,
6305
- },
6306
- contextWindow: 204800,
6307
- maxTokens: 131072,
6308
- },
6309
6241
  "minimax-m2.5": {
6310
6242
  id: "minimax-m2.5",
6311
6243
  name: "MiniMax M2.5",
@@ -7959,6 +7891,23 @@ export const MODELS = {
7959
7891
  contextWindow: 32768,
7960
7892
  maxTokens: 16384,
7961
7893
  },
7894
+ "mistralai/mistral-small-2603": {
7895
+ id: "mistralai/mistral-small-2603",
7896
+ name: "Mistral: Mistral Small 4",
7897
+ api: "openai-completions",
7898
+ provider: "openrouter",
7899
+ baseUrl: "https://openrouter.ai/api/v1",
7900
+ reasoning: true,
7901
+ input: ["text", "image"],
7902
+ cost: {
7903
+ input: 0.15,
7904
+ output: 0.6,
7905
+ cacheRead: 0,
7906
+ cacheWrite: 0,
7907
+ },
7908
+ contextWindow: 262144,
7909
+ maxTokens: 4096,
7910
+ },
7962
7911
  "mistralai/mistral-small-3.1-24b-instruct:free": {
7963
7912
  id: "mistralai/mistral-small-3.1-24b-instruct:free",
7964
7913
  name: "Mistral: Mistral Small 3.1 24B (free)",
@@ -9053,7 +9002,7 @@ export const MODELS = {
9053
9002
  api: "openai-completions",
9054
9003
  provider: "openrouter",
9055
9004
  baseUrl: "https://openrouter.ai/api/v1",
9056
- reasoning: false,
9005
+ reasoning: true,
9057
9006
  input: ["text", "image"],
9058
9007
  cost: {
9059
9008
  input: 15,
@@ -9104,7 +9053,7 @@ export const MODELS = {
9104
9053
  api: "openai-completions",
9105
9054
  provider: "openrouter",
9106
9055
  baseUrl: "https://openrouter.ai/api/v1",
9107
- reasoning: false,
9056
+ reasoning: true,
9108
9057
  input: ["text"],
9109
9058
  cost: {
9110
9059
  input: 1.1,
@@ -9121,7 +9070,7 @@ export const MODELS = {
9121
9070
  api: "openai-completions",
9122
9071
  provider: "openrouter",
9123
9072
  baseUrl: "https://openrouter.ai/api/v1",
9124
- reasoning: false,
9073
+ reasoning: true,
9125
9074
  input: ["text"],
9126
9075
  cost: {
9127
9076
  input: 1.1,
@@ -9345,9 +9294,9 @@ export const MODELS = {
9345
9294
  reasoning: false,
9346
9295
  input: ["text"],
9347
9296
  cost: {
9348
- input: 0.39999999999999997,
9349
- output: 1.2,
9350
- cacheRead: 0.08,
9297
+ input: 0.26,
9298
+ output: 0.78,
9299
+ cacheRead: 0.052000000000000005,
9351
9300
  cacheWrite: 0,
9352
9301
  },
9353
9302
  contextWindow: 1000000,
@@ -9413,8 +9362,8 @@ export const MODELS = {
9413
9362
  reasoning: false,
9414
9363
  input: ["text", "image"],
9415
9364
  cost: {
9416
- input: 0.7999999999999999,
9417
- output: 3.1999999999999997,
9365
+ input: 0.52,
9366
+ output: 2.08,
9418
9367
  cacheRead: 0,
9419
9368
  cacheWrite: 0,
9420
9369
  },
@@ -9702,9 +9651,9 @@ export const MODELS = {
9702
9651
  reasoning: false,
9703
9652
  input: ["text"],
9704
9653
  cost: {
9705
- input: 1.2,
9706
- output: 6,
9707
- cacheRead: 0.24,
9654
+ input: 0.78,
9655
+ output: 3.9,
9656
+ cacheRead: 0.156,
9708
9657
  cacheWrite: 0,
9709
9658
  },
9710
9659
  contextWindow: 262144,
@@ -9991,8 +9940,8 @@ export const MODELS = {
9991
9940
  reasoning: true,
9992
9941
  input: ["text", "image"],
9993
9942
  cost: {
9994
- input: 0.09999999999999999,
9995
- output: 0.39999999999999997,
9943
+ input: 0.065,
9944
+ output: 0.26,
9996
9945
  cacheRead: 0,
9997
9946
  cacheWrite: 0,
9998
9947
  },
@@ -10554,7 +10503,7 @@ export const MODELS = {
10554
10503
  reasoning: true,
10555
10504
  input: ["text"],
10556
10505
  cost: {
10557
- input: 0.06,
10506
+ input: 0.12,
10558
10507
  output: 0.24,
10559
10508
  cacheRead: 0,
10560
10509
  cacheWrite: 0,
@@ -10607,7 +10556,7 @@ export const MODELS = {
10607
10556
  cost: {
10608
10557
  input: 0.29,
10609
10558
  output: 0.59,
10610
- cacheRead: 0,
10559
+ cacheRead: 0.145,
10611
10560
  cacheWrite: 0,
10612
10561
  },
10613
10562
  contextWindow: 131072,
@@ -10622,9 +10571,9 @@ export const MODELS = {
10622
10571
  reasoning: true,
10623
10572
  input: ["text", "image"],
10624
10573
  cost: {
10625
- input: 0.3,
10626
- output: 2.9000000000000004,
10627
- cacheRead: 0,
10574
+ input: 0.22999999999999998,
10575
+ output: 2.3,
10576
+ cacheRead: 0.19999999999999998,
10628
10577
  cacheWrite: 0,
10629
10578
  },
10630
10579
  contextWindow: 262114,
@@ -10641,7 +10590,7 @@ export const MODELS = {
10641
10590
  cost: {
10642
10591
  input: 0.39999999999999997,
10643
10592
  output: 1.5999999999999999,
10644
- cacheRead: 0,
10593
+ cacheRead: 0.022,
10645
10594
  cacheWrite: 0,
10646
10595
  },
10647
10596
  contextWindow: 262144,
@@ -11134,7 +11083,7 @@ export const MODELS = {
11134
11083
  cost: {
11135
11084
  input: 0.27,
11136
11085
  output: 1,
11137
- cacheRead: 0,
11086
+ cacheRead: 0.135,
11138
11087
  cacheWrite: 0,
11139
11088
  },
11140
11089
  contextWindow: 131072,
@@ -11149,9 +11098,9 @@ export const MODELS = {
11149
11098
  reasoning: false,
11150
11099
  input: ["text"],
11151
11100
  cost: {
11152
- input: 0.26,
11153
- output: 0.38,
11154
- cacheRead: 0.13,
11101
+ input: 0.28,
11102
+ output: 0.42,
11103
+ cacheRead: 0.028,
11155
11104
  cacheWrite: 0,
11156
11105
  },
11157
11106
  contextWindow: 128000,
@@ -11185,7 +11134,7 @@ export const MODELS = {
11185
11134
  cost: {
11186
11135
  input: 0.15,
11187
11136
  output: 0.6,
11188
- cacheRead: 0,
11137
+ cacheRead: 0.024999999999999998,
11189
11138
  cacheWrite: 0,
11190
11139
  },
11191
11140
  contextWindow: 1048576,
@@ -11202,7 +11151,7 @@ export const MODELS = {
11202
11151
  cost: {
11203
11152
  input: 0.075,
11204
11153
  output: 0.3,
11205
- cacheRead: 0,
11154
+ cacheRead: 0.02,
11206
11155
  cacheWrite: 0,
11207
11156
  },
11208
11157
  contextWindow: 1048576,
@@ -11242,40 +11191,6 @@ export const MODELS = {
11242
11191
  contextWindow: 1048576,
11243
11192
  maxTokens: 65536,
11244
11193
  },
11245
- "google/gemini-2.5-flash-lite-preview-09-2025": {
11246
- id: "google/gemini-2.5-flash-lite-preview-09-2025",
11247
- name: "Gemini 2.5 Flash Lite Preview 09-2025",
11248
- api: "anthropic-messages",
11249
- provider: "vercel-ai-gateway",
11250
- baseUrl: "https://ai-gateway.vercel.sh",
11251
- reasoning: true,
11252
- input: ["text", "image"],
11253
- cost: {
11254
- input: 0.09999999999999999,
11255
- output: 0.39999999999999997,
11256
- cacheRead: 0.01,
11257
- cacheWrite: 0,
11258
- },
11259
- contextWindow: 1048576,
11260
- maxTokens: 65536,
11261
- },
11262
- "google/gemini-2.5-flash-preview-09-2025": {
11263
- id: "google/gemini-2.5-flash-preview-09-2025",
11264
- name: "Gemini 2.5 Flash Preview 09-2025",
11265
- api: "anthropic-messages",
11266
- provider: "vercel-ai-gateway",
11267
- baseUrl: "https://ai-gateway.vercel.sh",
11268
- reasoning: true,
11269
- input: ["text", "image"],
11270
- cost: {
11271
- input: 0.3,
11272
- output: 2.5,
11273
- cacheRead: 0.03,
11274
- cacheWrite: 0,
11275
- },
11276
- contextWindow: 1000000,
11277
- maxTokens: 65536,
11278
- },
11279
11194
  "google/gemini-2.5-pro": {
11280
11195
  id: "google/gemini-2.5-pro",
11281
11196
  name: "Gemini 2.5 Pro",
@@ -11457,7 +11372,7 @@ export const MODELS = {
11457
11372
  cost: {
11458
11373
  input: 0.09999999999999999,
11459
11374
  output: 0.09999999999999999,
11460
- cacheRead: 0,
11375
+ cacheRead: 0.09999999999999999,
11461
11376
  cacheWrite: 0,
11462
11377
  },
11463
11378
  contextWindow: 128000,
@@ -11630,8 +11545,8 @@ export const MODELS = {
11630
11545
  cacheRead: 0.03,
11631
11546
  cacheWrite: 0.375,
11632
11547
  },
11633
- contextWindow: 4096,
11634
- maxTokens: 4096,
11548
+ contextWindow: 204800,
11549
+ maxTokens: 131000,
11635
11550
  },
11636
11551
  "mistral/codestral": {
11637
11552
  id: "mistral/codestral",
@@ -11659,8 +11574,8 @@ export const MODELS = {
11659
11574
  reasoning: false,
11660
11575
  input: ["text"],
11661
11576
  cost: {
11662
- input: 0,
11663
- output: 0,
11577
+ input: 0.39999999999999997,
11578
+ output: 2,
11664
11579
  cacheRead: 0,
11665
11580
  cacheWrite: 0,
11666
11581
  },
@@ -11693,8 +11608,8 @@ export const MODELS = {
11693
11608
  reasoning: false,
11694
11609
  input: ["text"],
11695
11610
  cost: {
11696
- input: 0,
11697
- output: 0,
11611
+ input: 0.09999999999999999,
11612
+ output: 0.3,
11698
11613
  cacheRead: 0,
11699
11614
  cacheWrite: 0,
11700
11615
  },
@@ -11710,8 +11625,8 @@ export const MODELS = {
11710
11625
  reasoning: false,
11711
11626
  input: ["text"],
11712
11627
  cost: {
11713
- input: 0.04,
11714
- output: 0.04,
11628
+ input: 0.09999999999999999,
11629
+ output: 0.09999999999999999,
11715
11630
  cacheRead: 0,
11716
11631
  cacheWrite: 0,
11717
11632
  },
@@ -11727,8 +11642,8 @@ export const MODELS = {
11727
11642
  reasoning: false,
11728
11643
  input: ["text"],
11729
11644
  cost: {
11730
- input: 0.09999999999999999,
11731
- output: 0.09999999999999999,
11645
+ input: 0.15,
11646
+ output: 0.15,
11732
11647
  cacheRead: 0,
11733
11648
  cacheWrite: 0,
11734
11649
  },
@@ -11814,7 +11729,7 @@ export const MODELS = {
11814
11729
  cost: {
11815
11730
  input: 0.6,
11816
11731
  output: 2.5,
11817
- cacheRead: 0,
11732
+ cacheRead: 0.15,
11818
11733
  cacheWrite: 0,
11819
11734
  },
11820
11735
  contextWindow: 131072,
@@ -11831,7 +11746,7 @@ export const MODELS = {
11831
11746
  cost: {
11832
11747
  input: 0.6,
11833
11748
  output: 2.5,
11834
- cacheRead: 0,
11749
+ cacheRead: 0.15,
11835
11750
  cacheWrite: 0,
11836
11751
  },
11837
11752
  contextWindow: 256000,
@@ -11880,9 +11795,9 @@ export const MODELS = {
11880
11795
  reasoning: false,
11881
11796
  input: ["text"],
11882
11797
  cost: {
11883
- input: 2.4,
11884
- output: 10,
11885
- cacheRead: 0,
11798
+ input: 1.15,
11799
+ output: 8,
11800
+ cacheRead: 0.15,
11886
11801
  cacheWrite: 0,
11887
11802
  },
11888
11803
  contextWindow: 256000,
@@ -12613,7 +12528,7 @@ export const MODELS = {
12613
12528
  cost: {
12614
12529
  input: 3,
12615
12530
  output: 15,
12616
- cacheRead: 0,
12531
+ cacheRead: 0.75,
12617
12532
  cacheWrite: 0,
12618
12533
  },
12619
12534
  contextWindow: 131072,
@@ -12630,7 +12545,7 @@ export const MODELS = {
12630
12545
  cost: {
12631
12546
  input: 5,
12632
12547
  output: 25,
12633
- cacheRead: 0,
12548
+ cacheRead: 1.25,
12634
12549
  cacheWrite: 0,
12635
12550
  },
12636
12551
  contextWindow: 131072,
@@ -12647,7 +12562,7 @@ export const MODELS = {
12647
12562
  cost: {
12648
12563
  input: 0.3,
12649
12564
  output: 0.5,
12650
- cacheRead: 0,
12565
+ cacheRead: 0.075,
12651
12566
  cacheWrite: 0,
12652
12567
  },
12653
12568
  contextWindow: 131072,
@@ -12900,8 +12815,8 @@ export const MODELS = {
12900
12815
  reasoning: true,
12901
12816
  input: ["text"],
12902
12817
  cost: {
12903
- input: 0.44999999999999996,
12904
- output: 1.7999999999999998,
12818
+ input: 0.6,
12819
+ output: 2.2,
12905
12820
  cacheRead: 0.11,
12906
12821
  cacheWrite: 0,
12907
12822
  },
@@ -13010,6 +12925,23 @@ export const MODELS = {
13010
12925
  contextWindow: 202800,
13011
12926
  maxTokens: 131100,
13012
12927
  },
12928
+ "zai/glm-5-turbo": {
12929
+ id: "zai/glm-5-turbo",
12930
+ name: "GLM 5 Turbo",
12931
+ api: "anthropic-messages",
12932
+ provider: "vercel-ai-gateway",
12933
+ baseUrl: "https://ai-gateway.vercel.sh",
12934
+ reasoning: true,
12935
+ input: ["text"],
12936
+ cost: {
12937
+ input: 1.2,
12938
+ output: 4,
12939
+ cacheRead: 0.24,
12940
+ cacheWrite: 0,
12941
+ },
12942
+ contextWindow: 202800,
12943
+ maxTokens: 131100,
12944
+ },
13013
12945
  },
13014
12946
  "xai": {
13015
12947
  "grok-2": {
@@ -13601,6 +13533,24 @@ export const MODELS = {
13601
13533
  contextWindow: 204800,
13602
13534
  maxTokens: 131072,
13603
13535
  },
13536
+ "glm-5-turbo": {
13537
+ id: "glm-5-turbo",
13538
+ name: "GLM-5-Turbo",
13539
+ api: "openai-completions",
13540
+ provider: "zai",
13541
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
13542
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
13543
+ reasoning: true,
13544
+ input: ["text"],
13545
+ cost: {
13546
+ input: 1.2,
13547
+ output: 4,
13548
+ cacheRead: 0.24,
13549
+ cacheWrite: 0,
13550
+ },
13551
+ contextWindow: 200000,
13552
+ maxTokens: 131072,
13553
+ },
13604
13554
  },
13605
13555
  };
13606
13556
  //# sourceMappingURL=models.generated.js.map