omegon 0.9.1 → 0.9.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (46) hide show
  1. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.d.ts +58 -105
  2. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.d.ts.map +1 -1
  3. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.js +98 -148
  4. package/node_modules/@styrene-lab/pi-ai/dist/models.generated.js.map +1 -1
  5. package/node_modules/@styrene-lab/pi-ai/dist/utils/oauth/anthropic.d.ts.map +1 -1
  6. package/node_modules/@styrene-lab/pi-ai/dist/utils/oauth/anthropic.js +3 -7
  7. package/node_modules/@styrene-lab/pi-ai/dist/utils/oauth/anthropic.js.map +1 -1
  8. package/node_modules/@styrene-lab/pi-tui/dist/components/editor.d.ts.map +1 -1
  9. package/node_modules/@styrene-lab/pi-tui/dist/components/editor.js +4 -3
  10. package/node_modules/@styrene-lab/pi-tui/dist/components/editor.js.map +1 -1
  11. package/node_modules/@styrene-lab/pi-tui/dist/components/input.d.ts.map +1 -1
  12. package/node_modules/@styrene-lab/pi-tui/dist/components/input.js +9 -1
  13. package/node_modules/@styrene-lab/pi-tui/dist/components/input.js.map +1 -1
  14. package/node_modules/@styrene-lab/pi-tui/dist/index.d.ts +1 -1
  15. package/node_modules/@styrene-lab/pi-tui/dist/index.d.ts.map +1 -1
  16. package/node_modules/@styrene-lab/pi-tui/dist/index.js +1 -1
  17. package/node_modules/@styrene-lab/pi-tui/dist/index.js.map +1 -1
  18. package/node_modules/@styrene-lab/pi-tui/dist/keys.d.ts +2 -0
  19. package/node_modules/@styrene-lab/pi-tui/dist/keys.d.ts.map +1 -1
  20. package/node_modules/@styrene-lab/pi-tui/dist/keys.js +39 -0
  21. package/node_modules/@styrene-lab/pi-tui/dist/keys.js.map +1 -1
  22. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.d.ts +14 -0
  23. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.d.ts.map +1 -1
  24. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.js +21 -1
  25. package/node_modules/@styrene-lab/pi-tui/dist/stdin-buffer.js.map +1 -1
  26. package/node_modules/@styrene-lab/pi-tui/dist/terminal.d.ts +1 -0
  27. package/node_modules/@styrene-lab/pi-tui/dist/terminal.d.ts.map +1 -1
  28. package/node_modules/@styrene-lab/pi-tui/dist/terminal.js +40 -0
  29. package/node_modules/@styrene-lab/pi-tui/dist/terminal.js.map +1 -1
  30. package/package.json +5 -5
  31. package/scripts/publish-pi-mono.sh +5 -22
  32. package/node_modules/cliui/CHANGELOG.md +0 -121
  33. package/node_modules/color-convert/CHANGELOG.md +0 -54
  34. package/node_modules/esprima/ChangeLog +0 -235
  35. package/node_modules/mz/HISTORY.md +0 -66
  36. package/node_modules/proper-lockfile/CHANGELOG.md +0 -108
  37. package/node_modules/source-map/CHANGELOG.md +0 -301
  38. package/node_modules/thenify/History.md +0 -11
  39. package/node_modules/thenify-all/History.md +0 -11
  40. package/node_modules/y18n/CHANGELOG.md +0 -100
  41. package/node_modules/yargs/CHANGELOG.md +0 -88
  42. package/node_modules/yargs-parser/CHANGELOG.md +0 -263
  43. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/LICENSE +0 -0
  44. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/README.md +0 -0
  45. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/index.d.ts +0 -0
  46. /package/node_modules/{@types → p-retry/node_modules/@types}/retry/package.json +0 -0
@@ -6143,23 +6143,6 @@ export declare const MODELS: {
6143
6143
  contextWindow: number;
6144
6144
  maxTokens: number;
6145
6145
  };
6146
- readonly "gemini-3-pro": {
6147
- id: string;
6148
- name: string;
6149
- api: "google-generative-ai";
6150
- provider: string;
6151
- baseUrl: string;
6152
- reasoning: true;
6153
- input: ("image" | "text")[];
6154
- cost: {
6155
- input: number;
6156
- output: number;
6157
- cacheRead: number;
6158
- cacheWrite: number;
6159
- };
6160
- contextWindow: number;
6161
- maxTokens: number;
6162
- };
6163
6146
  readonly "gemini-3.1-pro": {
6164
6147
  id: string;
6165
6148
  name: string;
@@ -6177,40 +6160,6 @@ export declare const MODELS: {
6177
6160
  contextWindow: number;
6178
6161
  maxTokens: number;
6179
6162
  };
6180
- readonly "glm-4.6": {
6181
- id: string;
6182
- name: string;
6183
- api: "openai-completions";
6184
- provider: string;
6185
- baseUrl: string;
6186
- reasoning: true;
6187
- input: "text"[];
6188
- cost: {
6189
- input: number;
6190
- output: number;
6191
- cacheRead: number;
6192
- cacheWrite: number;
6193
- };
6194
- contextWindow: number;
6195
- maxTokens: number;
6196
- };
6197
- readonly "glm-4.7": {
6198
- id: string;
6199
- name: string;
6200
- api: "openai-completions";
6201
- provider: string;
6202
- baseUrl: string;
6203
- reasoning: true;
6204
- input: "text"[];
6205
- cost: {
6206
- input: number;
6207
- output: number;
6208
- cacheRead: number;
6209
- cacheWrite: number;
6210
- };
6211
- contextWindow: number;
6212
- maxTokens: number;
6213
- };
6214
6163
  readonly "glm-5": {
6215
6164
  id: string;
6216
6165
  name: string;
@@ -6466,23 +6415,6 @@ export declare const MODELS: {
6466
6415
  contextWindow: number;
6467
6416
  maxTokens: number;
6468
6417
  };
6469
- readonly "minimax-m2.1": {
6470
- id: string;
6471
- name: string;
6472
- api: "openai-completions";
6473
- provider: string;
6474
- baseUrl: string;
6475
- reasoning: true;
6476
- input: "text"[];
6477
- cost: {
6478
- input: number;
6479
- output: number;
6480
- cacheRead: number;
6481
- cacheWrite: number;
6482
- };
6483
- contextWindow: number;
6484
- maxTokens: number;
6485
- };
6486
6418
  readonly "minimax-m2.5": {
6487
6419
  id: string;
6488
6420
  name: string;
@@ -8136,6 +8068,23 @@ export declare const MODELS: {
8136
8068
  contextWindow: number;
8137
8069
  maxTokens: number;
8138
8070
  };
8071
+ readonly "mistralai/mistral-small-2603": {
8072
+ id: string;
8073
+ name: string;
8074
+ api: "openai-completions";
8075
+ provider: string;
8076
+ baseUrl: string;
8077
+ reasoning: true;
8078
+ input: ("image" | "text")[];
8079
+ cost: {
8080
+ input: number;
8081
+ output: number;
8082
+ cacheRead: number;
8083
+ cacheWrite: number;
8084
+ };
8085
+ contextWindow: number;
8086
+ maxTokens: number;
8087
+ };
8139
8088
  readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
8140
8089
  id: string;
8141
8090
  name: string;
@@ -9230,7 +9179,7 @@ export declare const MODELS: {
9230
9179
  api: "openai-completions";
9231
9180
  provider: string;
9232
9181
  baseUrl: string;
9233
- reasoning: false;
9182
+ reasoning: true;
9234
9183
  input: ("image" | "text")[];
9235
9184
  cost: {
9236
9185
  input: number;
@@ -9281,7 +9230,7 @@ export declare const MODELS: {
9281
9230
  api: "openai-completions";
9282
9231
  provider: string;
9283
9232
  baseUrl: string;
9284
- reasoning: false;
9233
+ reasoning: true;
9285
9234
  input: "text"[];
9286
9235
  cost: {
9287
9236
  input: number;
@@ -9298,7 +9247,7 @@ export declare const MODELS: {
9298
9247
  api: "openai-completions";
9299
9248
  provider: string;
9300
9249
  baseUrl: string;
9301
- reasoning: false;
9250
+ reasoning: true;
9302
9251
  input: "text"[];
9303
9252
  cost: {
9304
9253
  input: number;
@@ -11419,40 +11368,6 @@ export declare const MODELS: {
11419
11368
  contextWindow: number;
11420
11369
  maxTokens: number;
11421
11370
  };
11422
- readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
11423
- id: string;
11424
- name: string;
11425
- api: "anthropic-messages";
11426
- provider: string;
11427
- baseUrl: string;
11428
- reasoning: true;
11429
- input: ("image" | "text")[];
11430
- cost: {
11431
- input: number;
11432
- output: number;
11433
- cacheRead: number;
11434
- cacheWrite: number;
11435
- };
11436
- contextWindow: number;
11437
- maxTokens: number;
11438
- };
11439
- readonly "google/gemini-2.5-flash-preview-09-2025": {
11440
- id: string;
11441
- name: string;
11442
- api: "anthropic-messages";
11443
- provider: string;
11444
- baseUrl: string;
11445
- reasoning: true;
11446
- input: ("image" | "text")[];
11447
- cost: {
11448
- input: number;
11449
- output: number;
11450
- cacheRead: number;
11451
- cacheWrite: number;
11452
- };
11453
- contextWindow: number;
11454
- maxTokens: number;
11455
- };
11456
11371
  readonly "google/gemini-2.5-pro": {
11457
11372
  id: string;
11458
11373
  name: string;
@@ -13187,6 +13102,23 @@ export declare const MODELS: {
13187
13102
  contextWindow: number;
13188
13103
  maxTokens: number;
13189
13104
  };
13105
+ readonly "zai/glm-5-turbo": {
13106
+ id: string;
13107
+ name: string;
13108
+ api: "anthropic-messages";
13109
+ provider: string;
13110
+ baseUrl: string;
13111
+ reasoning: true;
13112
+ input: "text"[];
13113
+ cost: {
13114
+ input: number;
13115
+ output: number;
13116
+ cacheRead: number;
13117
+ cacheWrite: number;
13118
+ };
13119
+ contextWindow: number;
13120
+ maxTokens: number;
13121
+ };
13190
13122
  };
13191
13123
  readonly xai: {
13192
13124
  readonly "grok-2": {
@@ -13805,6 +13737,27 @@ export declare const MODELS: {
13805
13737
  contextWindow: number;
13806
13738
  maxTokens: number;
13807
13739
  };
13740
+ readonly "glm-5-turbo": {
13741
+ id: string;
13742
+ name: string;
13743
+ api: "openai-completions";
13744
+ provider: string;
13745
+ baseUrl: string;
13746
+ compat: {
13747
+ supportsDeveloperRole: false;
13748
+ thinkingFormat: "zai";
13749
+ };
13750
+ reasoning: true;
13751
+ input: "text"[];
13752
+ cost: {
13753
+ input: number;
13754
+ output: number;
13755
+ cacheRead: number;
13756
+ cacheWrite: number;
13757
+ };
13758
+ contextWindow: number;
13759
+ maxTokens: number;
13760
+ };
13808
13761
  };
13809
13762
  };
13810
13763
  //# sourceMappingURL=models.generated.d.ts.map