@bonginkan/maria 3.1.7 → 3.1.9

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/cli.cjs CHANGED
@@ -23,7 +23,6 @@ var inquirer = require('inquirer');
23
23
  var commander = require('commander');
24
24
  var readline = require('readline/promises');
25
25
  var process$1 = require('process');
26
- require('url');
27
26
 
28
27
  function _interopDefault (e) { return e && e.__esModule ? e : { default: e }; }
29
28
 
@@ -1041,26 +1040,240 @@ var init_base_provider = __esm({
1041
1040
  };
1042
1041
  }
1043
1042
  });
1043
+ var packageVersion, parseList, DEFAULT_MEMORY_CONFIG, DEFAULT_ROUTER_CONFIG, DEFAULT_UI_CONFIG, DEFAULT_PATHS, DEFAULT_ENV_VARS, DEFAULT_DUMMY_VALUES, DEFAULT_HELP_TEXTS, DEFAULT_COMMAND_OUTPUTS, DEFAULT_FEATURE_FLAGS, DEFAULT_PROVIDER_PREFS, AI_PROVIDERS_CONFIG, DEFAULT_STARTUP, DEFAULT_HSR_CONFIG, DEFAULT_RL_CONFIG, APP_VERSION, DEFAULT_CONFIG;
1044
+ var init_defaults = __esm({
1045
+ "src/config/defaults.ts"() {
1046
+ packageVersion = "3.1.8";
1047
+ try {
1048
+ const possiblePaths = [
1049
+ path17.join(process.cwd(), "package.json"),
1050
+ path17.join(__dirname, "../../package.json")
1051
+ ];
1052
+ for (const path19 of possiblePaths) {
1053
+ try {
1054
+ const pkgContent = fs11.readFileSync(path19, "utf8");
1055
+ const pkg2 = JSON.parse(pkgContent);
1056
+ if (pkg2.version) {
1057
+ packageVersion = pkg2.version;
1058
+ break;
1059
+ }
1060
+ } catch {
1061
+ }
1062
+ }
1063
+ } catch {
1064
+ }
1065
+ parseList = (envVar, defaultList) => {
1066
+ if (!envVar) return defaultList;
1067
+ return envVar.split(",").map((s) => s.trim()).filter(Boolean);
1068
+ };
1069
+ DEFAULT_MEMORY_CONFIG = {
1070
+ system1: {
1071
+ maxKnowledgeNodes: 1e3,
1072
+ embeddingDimension: 1536,
1073
+ cacheSize: 100,
1074
+ compressionThreshold: 0.75,
1075
+ accessDecayRate: 0.03
1076
+ },
1077
+ system2: {
1078
+ maxReasoningTraces: 100,
1079
+ qualityThreshold: 0.75,
1080
+ reflectionFrequency: 12,
1081
+ enhancementEvaluationInterval: 6
1082
+ },
1083
+ coordinator: {
1084
+ syncInterval: 5e3,
1085
+ conflictResolutionStrategy: "balanced",
1086
+ learningRate: 0.15,
1087
+ adaptationThreshold: 0.7
1088
+ },
1089
+ performance: {
1090
+ targetLatency: 50,
1091
+ maxMemoryUsage: 256,
1092
+ cacheStrategy: "lru",
1093
+ preloadPriority: "medium",
1094
+ backgroundOptimization: true,
1095
+ batchSize: 10
1096
+ }
1097
+ };
1098
+ DEFAULT_ROUTER_CONFIG = {
1099
+ confidenceThreshold: 0.7,
1100
+ enableLearning: true,
1101
+ supportedLanguages: ["en", "ja", "zh", "ko", "vi"],
1102
+ // Fixed language codes
1103
+ enableConfirmation: false,
1104
+ maxSuggestions: 5,
1105
+ enableFuzzyMatch: true
1106
+ };
1107
+ DEFAULT_UI_CONFIG = {
1108
+ width: 120,
1109
+ borderColor: "white",
1110
+ promptColor: "cyan",
1111
+ textColor: "white",
1112
+ promptSymbol: ">",
1113
+ placeholder: "Type your command or question here...",
1114
+ enablePasteDetection: true,
1115
+ enableFileDrop: true,
1116
+ showModeIndicator: true,
1117
+ animationSpeed: 100
1118
+ };
1119
+ DEFAULT_PATHS = {
1120
+ avatar: "face_only_96x96_ramp.txt"
1121
+ // 相対パスに変更
1122
+ };
1123
+ DEFAULT_ENV_VARS = {
1124
+ OPENAIAPI_KEY: "OPENAI_API_KEY",
1125
+ ANTHROPICAPI_KEY: "ANTHROPIC_API_KEY",
1126
+ GOOGLEAI_API_KEY: "GOOGLE_AI_API_KEY"
1127
+ };
1128
+ DEFAULT_DUMMY_VALUES = {
1129
+ lintScore: 94,
1130
+ typeCoverage: 87,
1131
+ securityScore: 89,
1132
+ defaultModel: "gpt-4 (openai)",
1133
+ cognitiveModes: 58
1134
+ // Updated to actual count
1135
+ };
1136
+ DEFAULT_HELP_TEXTS = {
1137
+ priorityUsage: "Usage: /priority <privacy-first|performance|cost-effective|auto>",
1138
+ githubAppUrl: "https://github.com/apps/maria-ai-assistant",
1139
+ cognitiveModesCount: 58,
1140
+ // Updated to actual count
1141
+ priorityModes: ["privacy-first", "performance", "cost-effective", "auto"]
1142
+ };
1143
+ DEFAULT_COMMAND_OUTPUTS = {
1144
+ setupEnvVars: [
1145
+ "export OPENAI_API_KEY=",
1146
+ "export ANTHROPIC_API_KEY=",
1147
+ "export GOOGLE_AI_API_KEY="
1148
+ ],
1149
+ modelSelector: {
1150
+ currentModel: "gpt-4 (openai) - Default",
1151
+ provider: "openai"
1152
+ }
1153
+ };
1154
+ DEFAULT_FEATURE_FLAGS = {
1155
+ enableEnhancedInterface: true,
1156
+ enableStreaming: true,
1157
+ enableAutoRouting: true,
1158
+ enableBgOptimization: true,
1159
+ enableMemorySystem: true,
1160
+ enableRLEvolution: true,
1161
+ enableHSRSystem: true,
1162
+ enableVisionAnalysis: true,
1163
+ enableProactiveReporting: true
1164
+ };
1165
+ DEFAULT_PROVIDER_PREFS = {
1166
+ provider: process.env.AI_PROVIDER || "openai",
1167
+ model: process.env.OPENAI_MODEL || "gpt-5-mini-2025-08-07",
1168
+ offline: process.env.OFFLINE_MODE === "true",
1169
+ debug: process.env.DEBUG === "true",
1170
+ priority: "auto",
1171
+ maxTokens: parseInt(process.env.OPENAI_MAX_TOKENS || "2000", 10),
1172
+ temperature: parseFloat(process.env.OPENAI_TEMPERATURE || "0.7")
1173
+ };
1174
+ AI_PROVIDERS_CONFIG = {
1175
+ openai: {
1176
+ apiKey: process.env.OPENAI_API_KEY,
1177
+ model: process.env.OPENAI_MODEL || "gpt-5-mini-2025-08-07",
1178
+ availableModels: parseList(process.env.OPENAI_MODELS, [
1179
+ "gpt-5-mini-2025-08-07",
1180
+ "gpt-5-mini",
1181
+ "gpt-4",
1182
+ "gpt-4-turbo",
1183
+ "o1-preview",
1184
+ "o1-mini"
1185
+ ]),
1186
+ endpoint: process.env.OPENAI_ENDPOINT || "https://api.openai.com/v1/chat/completions",
1187
+ maxTokens: parseInt(process.env.OPENAI_MAX_TOKENS || "2000", 10),
1188
+ temperature: parseFloat(process.env.OPENAI_TEMPERATURE || "0.7"),
1189
+ timeout: parseInt(process.env.OPENAI_TIMEOUT || "30000", 10),
1190
+ maxRetries: parseInt(process.env.OPENAI_MAX_RETRIES || "3", 10)
1191
+ },
1192
+ anthropic: {
1193
+ apiKey: process.env.ANTHROPIC_API_KEY,
1194
+ model: process.env.ANTHROPIC_MODEL || "claude-3-sonnet-20240229",
1195
+ availableModels: parseList(process.env.ANTHROPIC_MODELS, [
1196
+ "claude-3-opus-20240229",
1197
+ "claude-3-sonnet-20240229",
1198
+ "claude-3-haiku-20240307"
1199
+ ]),
1200
+ endpoint: process.env.ANTHROPIC_ENDPOINT || "https://api.anthropic.com/v1/messages",
1201
+ maxTokens: parseInt(process.env.ANTHROPIC_MAX_TOKENS || "2000", 10),
1202
+ temperature: parseFloat(process.env.ANTHROPIC_TEMPERATURE || "0.7")
1203
+ },
1204
+ ollama: {
1205
+ enabled: process.env.OLLAMA_ENABLED === "true",
1206
+ apiUrl: process.env.OLLAMA_API_URL || "http://localhost:11434",
1207
+ defaultModel: process.env.OLLAMA_DEFAULT_MODEL || "llama3.2:3b",
1208
+ maxTokens: parseInt(process.env.OLLAMA_MAX_TOKENS || "4096", 10)
1209
+ },
1210
+ lmstudio: {
1211
+ enabled: process.env.LMSTUDIO_ENABLED === "true",
1212
+ apiUrl: process.env.LMSTUDIO_API_URL || "http://localhost:1234",
1213
+ defaultModel: process.env.LMSTUDIO_DEFAULT_MODEL || "gpt-oss-120b",
1214
+ maxTokens: parseInt(process.env.LMSTUDIO_MAX_TOKENS || "8192", 10)
1215
+ }
1216
+ };
1217
+ DEFAULT_STARTUP = {
1218
+ showLogo: true,
1219
+ totalProvidersHint: 8,
1220
+ startupTimeout: 1e4,
1221
+ showLoadingAnimation: true,
1222
+ checkForUpdates: false
1223
+ };
1224
+ DEFAULT_HSR_CONFIG = {
1225
+ enableHumanInterruption: true,
1226
+ interruptionCheckInterval: 100,
1227
+ autoSaveInterval: 3e4,
1228
+ maxOperationTime: 6e4,
1229
+ safetyLevel: "medium"
1230
+ };
1231
+ DEFAULT_RL_CONFIG = {
1232
+ learningRate: 1e-3,
1233
+ discountFactor: 0.99,
1234
+ epsilon: 0.1,
1235
+ replayBufferSize: 1e4,
1236
+ batchSize: 32,
1237
+ updateFrequency: 100,
1238
+ enablePPO: true,
1239
+ enableDPO: false
1240
+ };
1241
+ APP_VERSION = process.env.npm_package_version || packageVersion;
1242
+ DEFAULT_CONFIG = {
1243
+ version: APP_VERSION,
1244
+ memory: DEFAULT_MEMORY_CONFIG,
1245
+ router: DEFAULT_ROUTER_CONFIG,
1246
+ ui: DEFAULT_UI_CONFIG,
1247
+ paths: DEFAULT_PATHS,
1248
+ env: DEFAULT_ENV_VARS,
1249
+ values: DEFAULT_DUMMY_VALUES,
1250
+ help: DEFAULT_HELP_TEXTS,
1251
+ command: DEFAULT_COMMAND_OUTPUTS,
1252
+ flags: DEFAULT_FEATURE_FLAGS,
1253
+ provider: DEFAULT_PROVIDER_PREFS,
1254
+ providers: AI_PROVIDERS_CONFIG,
1255
+ startup: DEFAULT_STARTUP,
1256
+ hsr: DEFAULT_HSR_CONFIG,
1257
+ rl: DEFAULT_RL_CONFIG
1258
+ };
1259
+ }
1260
+ });
1044
1261
 
1045
1262
  // src/services/ai-response/providers/openai-provider.ts
1046
1263
  var OpenAIProvider;
1047
1264
  var init_openai_provider = __esm({
1048
1265
  "src/services/ai-response/providers/openai-provider.ts"() {
1049
1266
  init_base_provider();
1267
+ init_defaults();
1050
1268
  OpenAIProvider = class extends BaseAIProvider {
1051
1269
  name = "OpenAI";
1052
- apiEndpoint = "https://api.openai.com/v1/chat/completions";
1270
+ apiEndpoint = AI_PROVIDERS_CONFIG.openai.endpoint;
1053
1271
  async performInitialization() {
1054
1272
  if (!this.config.defaultModel) {
1055
- this.config.defaultModel = "gpt-3.5-turbo";
1273
+ this.config.defaultModel = AI_PROVIDERS_CONFIG.openai.model;
1056
1274
  }
1057
- this.availableModels = [
1058
- "gpt-4",
1059
- "gpt-4-turbo-preview",
1060
- "gpt-3.5-turbo",
1061
- "gpt-3.5-turbo-16k"
1062
- ];
1063
- if (!this.config.apiKey) {
1275
+ this.availableModels = AI_PROVIDERS_CONFIG.openai.availableModels;
1276
+ if (!this.config.apiKey && !AI_PROVIDERS_CONFIG.openai.apiKey) {
1064
1277
  console.warn("[OpenAI Provider] No API key provided - will use template responses");
1065
1278
  }
1066
1279
  }
@@ -1205,21 +1418,16 @@ var AnthropicProvider;
1205
1418
  var init_anthropic_provider = __esm({
1206
1419
  "src/services/ai-response/providers/anthropic-provider.ts"() {
1207
1420
  init_base_provider();
1421
+ init_defaults();
1208
1422
  AnthropicProvider = class extends BaseAIProvider {
1209
1423
  name = "Anthropic";
1210
- apiEndpoint = "https://api.anthropic.com/v1/messages";
1424
+ apiEndpoint = AI_PROVIDERS_CONFIG.anthropic.endpoint;
1211
1425
  async performInitialization() {
1212
1426
  if (!this.config.defaultModel) {
1213
- this.config.defaultModel = "claude-3-sonnet-20240229";
1427
+ this.config.defaultModel = AI_PROVIDERS_CONFIG.anthropic.model;
1214
1428
  }
1215
- this.availableModels = [
1216
- "claude-3-opus-20240229",
1217
- "claude-3-sonnet-20240229",
1218
- "claude-3-haiku-20240307",
1219
- "claude-2.1",
1220
- "claude-instant-1.2"
1221
- ];
1222
- if (!this.config.apiKey) {
1429
+ this.availableModels = AI_PROVIDERS_CONFIG.anthropic.availableModels;
1430
+ if (!this.config.apiKey && !AI_PROVIDERS_CONFIG.anthropic.apiKey) {
1223
1431
  console.warn("[Anthropic Provider] No API key provided - will use template responses");
1224
1432
  }
1225
1433
  }
@@ -11682,157 +11890,6 @@ var init_startup_display = __esm({
11682
11890
  }
11683
11891
  });
11684
11892
 
11685
- // src/config/defaults.ts
11686
- var DEFAULT_MEMORY_CONFIG, DEFAULT_ROUTER_CONFIG, DEFAULT_UI_CONFIG, DEFAULT_PATHS, DEFAULT_ENV_VARS, DEFAULT_DUMMY_VALUES, DEFAULT_HELP_TEXTS, DEFAULT_COMMAND_OUTPUTS, DEFAULT_FEATURE_FLAGS, DEFAULT_PROVIDER_PREFS, DEFAULT_STARTUP, DEFAULT_HSR_CONFIG, DEFAULT_RL_CONFIG, DEFAULT_CONFIG;
11687
- var init_defaults = __esm({
11688
- "src/config/defaults.ts"() {
11689
- DEFAULT_MEMORY_CONFIG = {
11690
- system1: {
11691
- maxKnowledgeNodes: 1e3,
11692
- embeddingDimension: 1536,
11693
- cacheSize: 100,
11694
- compressionThreshold: 0.75,
11695
- accessDecayRate: 0.03
11696
- },
11697
- system2: {
11698
- maxReasoningTraces: 100,
11699
- qualityThreshold: 0.75,
11700
- reflectionFrequency: 12,
11701
- enhancementEvaluationInterval: 6
11702
- },
11703
- coordinator: {
11704
- syncInterval: 5e3,
11705
- conflictResolutionStrategy: "balanced",
11706
- learningRate: 0.15,
11707
- adaptationThreshold: 0.7
11708
- },
11709
- performance: {
11710
- targetLatency: 50,
11711
- maxMemoryUsage: 256,
11712
- cacheStrategy: "lru",
11713
- preloadPriority: "medium",
11714
- backgroundOptimization: true,
11715
- batchSize: 10
11716
- }
11717
- };
11718
- DEFAULT_ROUTER_CONFIG = {
11719
- confidenceThreshold: 0.7,
11720
- enableLearning: true,
11721
- supportedLanguages: ["en", "ja", "zh", "ko", "vi"],
11722
- // Fixed language codes
11723
- enableConfirmation: false,
11724
- maxSuggestions: 5,
11725
- enableFuzzyMatch: true
11726
- };
11727
- DEFAULT_UI_CONFIG = {
11728
- width: 120,
11729
- borderColor: "white",
11730
- promptColor: "cyan",
11731
- textColor: "white",
11732
- promptSymbol: ">",
11733
- placeholder: "Type your command or question here...",
11734
- enablePasteDetection: true,
11735
- enableFileDrop: true,
11736
- showModeIndicator: true,
11737
- animationSpeed: 100
11738
- };
11739
- DEFAULT_PATHS = {
11740
- avatar: "face_only_96x96_ramp.txt"
11741
- // 相対パスに変更
11742
- };
11743
- DEFAULT_ENV_VARS = {
11744
- OPENAIAPI_KEY: "OPENAI_API_KEY",
11745
- ANTHROPICAPI_KEY: "ANTHROPIC_API_KEY",
11746
- GOOGLEAI_API_KEY: "GOOGLE_AI_API_KEY"
11747
- };
11748
- DEFAULT_DUMMY_VALUES = {
11749
- lintScore: 94,
11750
- typeCoverage: 87,
11751
- securityScore: 89,
11752
- defaultModel: "gpt-4 (openai)",
11753
- cognitiveModes: 58
11754
- // Updated to actual count
11755
- };
11756
- DEFAULT_HELP_TEXTS = {
11757
- priorityUsage: "Usage: /priority <privacy-first|performance|cost-effective|auto>",
11758
- githubAppUrl: "https://github.com/apps/maria-ai-assistant",
11759
- cognitiveModesCount: 58,
11760
- // Updated to actual count
11761
- priorityModes: ["privacy-first", "performance", "cost-effective", "auto"]
11762
- };
11763
- DEFAULT_COMMAND_OUTPUTS = {
11764
- setupEnvVars: [
11765
- "export OPENAI_API_KEY=",
11766
- "export ANTHROPIC_API_KEY=",
11767
- "export GOOGLE_AI_API_KEY="
11768
- ],
11769
- modelSelector: {
11770
- currentModel: "gpt-4 (openai) - Default",
11771
- provider: "openai"
11772
- }
11773
- };
11774
- DEFAULT_FEATURE_FLAGS = {
11775
- enableEnhancedInterface: true,
11776
- enableStreaming: true,
11777
- enableAutoRouting: true,
11778
- enableBgOptimization: true,
11779
- enableMemorySystem: true,
11780
- enableRLEvolution: true,
11781
- enableHSRSystem: true,
11782
- enableVisionAnalysis: true,
11783
- enableProactiveReporting: true
11784
- };
11785
- DEFAULT_PROVIDER_PREFS = {
11786
- provider: "openai",
11787
- model: "gpt-4",
11788
- offline: false,
11789
- debug: false,
11790
- priority: "auto",
11791
- maxTokens: 4096,
11792
- temperature: 0.7
11793
- };
11794
- DEFAULT_STARTUP = {
11795
- showLogo: true,
11796
- totalProvidersHint: 8,
11797
- startupTimeout: 1e4,
11798
- showLoadingAnimation: true,
11799
- checkForUpdates: false
11800
- };
11801
- DEFAULT_HSR_CONFIG = {
11802
- enableHumanInterruption: true,
11803
- interruptionCheckInterval: 100,
11804
- autoSaveInterval: 3e4,
11805
- maxOperationTime: 6e4,
11806
- safetyLevel: "medium"
11807
- };
11808
- DEFAULT_RL_CONFIG = {
11809
- learningRate: 1e-3,
11810
- discountFactor: 0.99,
11811
- epsilon: 0.1,
11812
- replayBufferSize: 1e4,
11813
- batchSize: 32,
11814
- updateFrequency: 100,
11815
- enablePPO: true,
11816
- enableDPO: false
11817
- };
11818
- DEFAULT_CONFIG = {
11819
- memory: DEFAULT_MEMORY_CONFIG,
11820
- router: DEFAULT_ROUTER_CONFIG,
11821
- ui: DEFAULT_UI_CONFIG,
11822
- paths: DEFAULT_PATHS,
11823
- env: DEFAULT_ENV_VARS,
11824
- values: DEFAULT_DUMMY_VALUES,
11825
- help: DEFAULT_HELP_TEXTS,
11826
- command: DEFAULT_COMMAND_OUTPUTS,
11827
- flags: DEFAULT_FEATURE_FLAGS,
11828
- provider: DEFAULT_PROVIDER_PREFS,
11829
- startup: DEFAULT_STARTUP,
11830
- hsr: DEFAULT_HSR_CONFIG,
11831
- rl: DEFAULT_RL_CONFIG
11832
- };
11833
- }
11834
- });
11835
-
11836
11893
  // src/config/loader.ts
11837
11894
  function loadCompleteConfig(options2 = {}) {
11838
11895
  const _config = { ...DEFAULT_CONFIG };
@@ -22207,16 +22264,8 @@ var init_cli = __esm({
22207
22264
  init_chat_context_fixed_service();
22208
22265
  init_ai_response_service();
22209
22266
  init_slash_command_handler();
22210
- packageJson2 = { version: "3.1.7" };
22211
- try {
22212
- const packagePath = path17.join(process.cwd(), "package.json");
22213
- const pkgContent = fs11.readFileSync(packagePath, "utf8");
22214
- const pkg2 = JSON.parse(pkgContent);
22215
- if (pkg2.version) {
22216
- packageJson2.version = pkg2.version;
22217
- }
22218
- } catch {
22219
- }
22267
+ init_defaults();
22268
+ packageJson2 = { version: APP_VERSION };
22220
22269
  sessionMemory = [];
22221
22270
  program = createCLI();
22222
22271
  program.parse(process.argv);