@bonginkan/maria 3.1.7 → 3.1.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +53 -27
- package/dist/bin/maria.cjs +133 -82
- package/dist/bin/maria.cjs.map +1 -1
- package/dist/cli.cjs +230 -181
- package/dist/cli.cjs.map +1 -1
- package/package.json +2 -2
package/dist/cli.cjs
CHANGED
|
@@ -23,7 +23,6 @@ var inquirer = require('inquirer');
|
|
|
23
23
|
var commander = require('commander');
|
|
24
24
|
var readline = require('readline/promises');
|
|
25
25
|
var process$1 = require('process');
|
|
26
|
-
require('url');
|
|
27
26
|
|
|
28
27
|
function _interopDefault (e) { return e && e.__esModule ? e : { default: e }; }
|
|
29
28
|
|
|
@@ -1041,26 +1040,240 @@ var init_base_provider = __esm({
|
|
|
1041
1040
|
};
|
|
1042
1041
|
}
|
|
1043
1042
|
});
|
|
1043
|
+
var packageVersion, parseList, DEFAULT_MEMORY_CONFIG, DEFAULT_ROUTER_CONFIG, DEFAULT_UI_CONFIG, DEFAULT_PATHS, DEFAULT_ENV_VARS, DEFAULT_DUMMY_VALUES, DEFAULT_HELP_TEXTS, DEFAULT_COMMAND_OUTPUTS, DEFAULT_FEATURE_FLAGS, DEFAULT_PROVIDER_PREFS, AI_PROVIDERS_CONFIG, DEFAULT_STARTUP, DEFAULT_HSR_CONFIG, DEFAULT_RL_CONFIG, APP_VERSION, DEFAULT_CONFIG;
|
|
1044
|
+
var init_defaults = __esm({
|
|
1045
|
+
"src/config/defaults.ts"() {
|
|
1046
|
+
packageVersion = "3.1.8";
|
|
1047
|
+
try {
|
|
1048
|
+
const possiblePaths = [
|
|
1049
|
+
path17.join(process.cwd(), "package.json"),
|
|
1050
|
+
path17.join(__dirname, "../../package.json")
|
|
1051
|
+
];
|
|
1052
|
+
for (const path19 of possiblePaths) {
|
|
1053
|
+
try {
|
|
1054
|
+
const pkgContent = fs11.readFileSync(path19, "utf8");
|
|
1055
|
+
const pkg2 = JSON.parse(pkgContent);
|
|
1056
|
+
if (pkg2.version) {
|
|
1057
|
+
packageVersion = pkg2.version;
|
|
1058
|
+
break;
|
|
1059
|
+
}
|
|
1060
|
+
} catch {
|
|
1061
|
+
}
|
|
1062
|
+
}
|
|
1063
|
+
} catch {
|
|
1064
|
+
}
|
|
1065
|
+
parseList = (envVar, defaultList) => {
|
|
1066
|
+
if (!envVar) return defaultList;
|
|
1067
|
+
return envVar.split(",").map((s) => s.trim()).filter(Boolean);
|
|
1068
|
+
};
|
|
1069
|
+
DEFAULT_MEMORY_CONFIG = {
|
|
1070
|
+
system1: {
|
|
1071
|
+
maxKnowledgeNodes: 1e3,
|
|
1072
|
+
embeddingDimension: 1536,
|
|
1073
|
+
cacheSize: 100,
|
|
1074
|
+
compressionThreshold: 0.75,
|
|
1075
|
+
accessDecayRate: 0.03
|
|
1076
|
+
},
|
|
1077
|
+
system2: {
|
|
1078
|
+
maxReasoningTraces: 100,
|
|
1079
|
+
qualityThreshold: 0.75,
|
|
1080
|
+
reflectionFrequency: 12,
|
|
1081
|
+
enhancementEvaluationInterval: 6
|
|
1082
|
+
},
|
|
1083
|
+
coordinator: {
|
|
1084
|
+
syncInterval: 5e3,
|
|
1085
|
+
conflictResolutionStrategy: "balanced",
|
|
1086
|
+
learningRate: 0.15,
|
|
1087
|
+
adaptationThreshold: 0.7
|
|
1088
|
+
},
|
|
1089
|
+
performance: {
|
|
1090
|
+
targetLatency: 50,
|
|
1091
|
+
maxMemoryUsage: 256,
|
|
1092
|
+
cacheStrategy: "lru",
|
|
1093
|
+
preloadPriority: "medium",
|
|
1094
|
+
backgroundOptimization: true,
|
|
1095
|
+
batchSize: 10
|
|
1096
|
+
}
|
|
1097
|
+
};
|
|
1098
|
+
DEFAULT_ROUTER_CONFIG = {
|
|
1099
|
+
confidenceThreshold: 0.7,
|
|
1100
|
+
enableLearning: true,
|
|
1101
|
+
supportedLanguages: ["en", "ja", "zh", "ko", "vi"],
|
|
1102
|
+
// Fixed language codes
|
|
1103
|
+
enableConfirmation: false,
|
|
1104
|
+
maxSuggestions: 5,
|
|
1105
|
+
enableFuzzyMatch: true
|
|
1106
|
+
};
|
|
1107
|
+
DEFAULT_UI_CONFIG = {
|
|
1108
|
+
width: 120,
|
|
1109
|
+
borderColor: "white",
|
|
1110
|
+
promptColor: "cyan",
|
|
1111
|
+
textColor: "white",
|
|
1112
|
+
promptSymbol: ">",
|
|
1113
|
+
placeholder: "Type your command or question here...",
|
|
1114
|
+
enablePasteDetection: true,
|
|
1115
|
+
enableFileDrop: true,
|
|
1116
|
+
showModeIndicator: true,
|
|
1117
|
+
animationSpeed: 100
|
|
1118
|
+
};
|
|
1119
|
+
DEFAULT_PATHS = {
|
|
1120
|
+
avatar: "face_only_96x96_ramp.txt"
|
|
1121
|
+
// 相対パスに変更
|
|
1122
|
+
};
|
|
1123
|
+
DEFAULT_ENV_VARS = {
|
|
1124
|
+
OPENAIAPI_KEY: "OPENAI_API_KEY",
|
|
1125
|
+
ANTHROPICAPI_KEY: "ANTHROPIC_API_KEY",
|
|
1126
|
+
GOOGLEAI_API_KEY: "GOOGLE_AI_API_KEY"
|
|
1127
|
+
};
|
|
1128
|
+
DEFAULT_DUMMY_VALUES = {
|
|
1129
|
+
lintScore: 94,
|
|
1130
|
+
typeCoverage: 87,
|
|
1131
|
+
securityScore: 89,
|
|
1132
|
+
defaultModel: "gpt-4 (openai)",
|
|
1133
|
+
cognitiveModes: 58
|
|
1134
|
+
// Updated to actual count
|
|
1135
|
+
};
|
|
1136
|
+
DEFAULT_HELP_TEXTS = {
|
|
1137
|
+
priorityUsage: "Usage: /priority <privacy-first|performance|cost-effective|auto>",
|
|
1138
|
+
githubAppUrl: "https://github.com/apps/maria-ai-assistant",
|
|
1139
|
+
cognitiveModesCount: 58,
|
|
1140
|
+
// Updated to actual count
|
|
1141
|
+
priorityModes: ["privacy-first", "performance", "cost-effective", "auto"]
|
|
1142
|
+
};
|
|
1143
|
+
DEFAULT_COMMAND_OUTPUTS = {
|
|
1144
|
+
setupEnvVars: [
|
|
1145
|
+
"export OPENAI_API_KEY=",
|
|
1146
|
+
"export ANTHROPIC_API_KEY=",
|
|
1147
|
+
"export GOOGLE_AI_API_KEY="
|
|
1148
|
+
],
|
|
1149
|
+
modelSelector: {
|
|
1150
|
+
currentModel: "gpt-4 (openai) - Default",
|
|
1151
|
+
provider: "openai"
|
|
1152
|
+
}
|
|
1153
|
+
};
|
|
1154
|
+
DEFAULT_FEATURE_FLAGS = {
|
|
1155
|
+
enableEnhancedInterface: true,
|
|
1156
|
+
enableStreaming: true,
|
|
1157
|
+
enableAutoRouting: true,
|
|
1158
|
+
enableBgOptimization: true,
|
|
1159
|
+
enableMemorySystem: true,
|
|
1160
|
+
enableRLEvolution: true,
|
|
1161
|
+
enableHSRSystem: true,
|
|
1162
|
+
enableVisionAnalysis: true,
|
|
1163
|
+
enableProactiveReporting: true
|
|
1164
|
+
};
|
|
1165
|
+
DEFAULT_PROVIDER_PREFS = {
|
|
1166
|
+
provider: process.env.AI_PROVIDER || "openai",
|
|
1167
|
+
model: process.env.OPENAI_MODEL || "gpt-5-mini-2025-08-07",
|
|
1168
|
+
offline: process.env.OFFLINE_MODE === "true",
|
|
1169
|
+
debug: process.env.DEBUG === "true",
|
|
1170
|
+
priority: "auto",
|
|
1171
|
+
maxTokens: parseInt(process.env.OPENAI_MAX_TOKENS || "2000", 10),
|
|
1172
|
+
temperature: parseFloat(process.env.OPENAI_TEMPERATURE || "0.7")
|
|
1173
|
+
};
|
|
1174
|
+
AI_PROVIDERS_CONFIG = {
|
|
1175
|
+
openai: {
|
|
1176
|
+
apiKey: process.env.OPENAI_API_KEY,
|
|
1177
|
+
model: process.env.OPENAI_MODEL || "gpt-5-mini-2025-08-07",
|
|
1178
|
+
availableModels: parseList(process.env.OPENAI_MODELS, [
|
|
1179
|
+
"gpt-5-mini-2025-08-07",
|
|
1180
|
+
"gpt-5-mini",
|
|
1181
|
+
"gpt-4",
|
|
1182
|
+
"gpt-4-turbo",
|
|
1183
|
+
"o1-preview",
|
|
1184
|
+
"o1-mini"
|
|
1185
|
+
]),
|
|
1186
|
+
endpoint: process.env.OPENAI_ENDPOINT || "https://api.openai.com/v1/chat/completions",
|
|
1187
|
+
maxTokens: parseInt(process.env.OPENAI_MAX_TOKENS || "2000", 10),
|
|
1188
|
+
temperature: parseFloat(process.env.OPENAI_TEMPERATURE || "0.7"),
|
|
1189
|
+
timeout: parseInt(process.env.OPENAI_TIMEOUT || "30000", 10),
|
|
1190
|
+
maxRetries: parseInt(process.env.OPENAI_MAX_RETRIES || "3", 10)
|
|
1191
|
+
},
|
|
1192
|
+
anthropic: {
|
|
1193
|
+
apiKey: process.env.ANTHROPIC_API_KEY,
|
|
1194
|
+
model: process.env.ANTHROPIC_MODEL || "claude-3-sonnet-20240229",
|
|
1195
|
+
availableModels: parseList(process.env.ANTHROPIC_MODELS, [
|
|
1196
|
+
"claude-3-opus-20240229",
|
|
1197
|
+
"claude-3-sonnet-20240229",
|
|
1198
|
+
"claude-3-haiku-20240307"
|
|
1199
|
+
]),
|
|
1200
|
+
endpoint: process.env.ANTHROPIC_ENDPOINT || "https://api.anthropic.com/v1/messages",
|
|
1201
|
+
maxTokens: parseInt(process.env.ANTHROPIC_MAX_TOKENS || "2000", 10),
|
|
1202
|
+
temperature: parseFloat(process.env.ANTHROPIC_TEMPERATURE || "0.7")
|
|
1203
|
+
},
|
|
1204
|
+
ollama: {
|
|
1205
|
+
enabled: process.env.OLLAMA_ENABLED === "true",
|
|
1206
|
+
apiUrl: process.env.OLLAMA_API_URL || "http://localhost:11434",
|
|
1207
|
+
defaultModel: process.env.OLLAMA_DEFAULT_MODEL || "llama3.2:3b",
|
|
1208
|
+
maxTokens: parseInt(process.env.OLLAMA_MAX_TOKENS || "4096", 10)
|
|
1209
|
+
},
|
|
1210
|
+
lmstudio: {
|
|
1211
|
+
enabled: process.env.LMSTUDIO_ENABLED === "true",
|
|
1212
|
+
apiUrl: process.env.LMSTUDIO_API_URL || "http://localhost:1234",
|
|
1213
|
+
defaultModel: process.env.LMSTUDIO_DEFAULT_MODEL || "gpt-oss-120b",
|
|
1214
|
+
maxTokens: parseInt(process.env.LMSTUDIO_MAX_TOKENS || "8192", 10)
|
|
1215
|
+
}
|
|
1216
|
+
};
|
|
1217
|
+
DEFAULT_STARTUP = {
|
|
1218
|
+
showLogo: true,
|
|
1219
|
+
totalProvidersHint: 8,
|
|
1220
|
+
startupTimeout: 1e4,
|
|
1221
|
+
showLoadingAnimation: true,
|
|
1222
|
+
checkForUpdates: false
|
|
1223
|
+
};
|
|
1224
|
+
DEFAULT_HSR_CONFIG = {
|
|
1225
|
+
enableHumanInterruption: true,
|
|
1226
|
+
interruptionCheckInterval: 100,
|
|
1227
|
+
autoSaveInterval: 3e4,
|
|
1228
|
+
maxOperationTime: 6e4,
|
|
1229
|
+
safetyLevel: "medium"
|
|
1230
|
+
};
|
|
1231
|
+
DEFAULT_RL_CONFIG = {
|
|
1232
|
+
learningRate: 1e-3,
|
|
1233
|
+
discountFactor: 0.99,
|
|
1234
|
+
epsilon: 0.1,
|
|
1235
|
+
replayBufferSize: 1e4,
|
|
1236
|
+
batchSize: 32,
|
|
1237
|
+
updateFrequency: 100,
|
|
1238
|
+
enablePPO: true,
|
|
1239
|
+
enableDPO: false
|
|
1240
|
+
};
|
|
1241
|
+
APP_VERSION = process.env.npm_package_version || packageVersion;
|
|
1242
|
+
DEFAULT_CONFIG = {
|
|
1243
|
+
version: APP_VERSION,
|
|
1244
|
+
memory: DEFAULT_MEMORY_CONFIG,
|
|
1245
|
+
router: DEFAULT_ROUTER_CONFIG,
|
|
1246
|
+
ui: DEFAULT_UI_CONFIG,
|
|
1247
|
+
paths: DEFAULT_PATHS,
|
|
1248
|
+
env: DEFAULT_ENV_VARS,
|
|
1249
|
+
values: DEFAULT_DUMMY_VALUES,
|
|
1250
|
+
help: DEFAULT_HELP_TEXTS,
|
|
1251
|
+
command: DEFAULT_COMMAND_OUTPUTS,
|
|
1252
|
+
flags: DEFAULT_FEATURE_FLAGS,
|
|
1253
|
+
provider: DEFAULT_PROVIDER_PREFS,
|
|
1254
|
+
providers: AI_PROVIDERS_CONFIG,
|
|
1255
|
+
startup: DEFAULT_STARTUP,
|
|
1256
|
+
hsr: DEFAULT_HSR_CONFIG,
|
|
1257
|
+
rl: DEFAULT_RL_CONFIG
|
|
1258
|
+
};
|
|
1259
|
+
}
|
|
1260
|
+
});
|
|
1044
1261
|
|
|
1045
1262
|
// src/services/ai-response/providers/openai-provider.ts
|
|
1046
1263
|
var OpenAIProvider;
|
|
1047
1264
|
var init_openai_provider = __esm({
|
|
1048
1265
|
"src/services/ai-response/providers/openai-provider.ts"() {
|
|
1049
1266
|
init_base_provider();
|
|
1267
|
+
init_defaults();
|
|
1050
1268
|
OpenAIProvider = class extends BaseAIProvider {
|
|
1051
1269
|
name = "OpenAI";
|
|
1052
|
-
apiEndpoint =
|
|
1270
|
+
apiEndpoint = AI_PROVIDERS_CONFIG.openai.endpoint;
|
|
1053
1271
|
async performInitialization() {
|
|
1054
1272
|
if (!this.config.defaultModel) {
|
|
1055
|
-
this.config.defaultModel =
|
|
1273
|
+
this.config.defaultModel = AI_PROVIDERS_CONFIG.openai.model;
|
|
1056
1274
|
}
|
|
1057
|
-
this.availableModels =
|
|
1058
|
-
|
|
1059
|
-
"gpt-4-turbo-preview",
|
|
1060
|
-
"gpt-3.5-turbo",
|
|
1061
|
-
"gpt-3.5-turbo-16k"
|
|
1062
|
-
];
|
|
1063
|
-
if (!this.config.apiKey) {
|
|
1275
|
+
this.availableModels = AI_PROVIDERS_CONFIG.openai.availableModels;
|
|
1276
|
+
if (!this.config.apiKey && !AI_PROVIDERS_CONFIG.openai.apiKey) {
|
|
1064
1277
|
console.warn("[OpenAI Provider] No API key provided - will use template responses");
|
|
1065
1278
|
}
|
|
1066
1279
|
}
|
|
@@ -1205,21 +1418,16 @@ var AnthropicProvider;
|
|
|
1205
1418
|
var init_anthropic_provider = __esm({
|
|
1206
1419
|
"src/services/ai-response/providers/anthropic-provider.ts"() {
|
|
1207
1420
|
init_base_provider();
|
|
1421
|
+
init_defaults();
|
|
1208
1422
|
AnthropicProvider = class extends BaseAIProvider {
|
|
1209
1423
|
name = "Anthropic";
|
|
1210
|
-
apiEndpoint =
|
|
1424
|
+
apiEndpoint = AI_PROVIDERS_CONFIG.anthropic.endpoint;
|
|
1211
1425
|
async performInitialization() {
|
|
1212
1426
|
if (!this.config.defaultModel) {
|
|
1213
|
-
this.config.defaultModel =
|
|
1427
|
+
this.config.defaultModel = AI_PROVIDERS_CONFIG.anthropic.model;
|
|
1214
1428
|
}
|
|
1215
|
-
this.availableModels =
|
|
1216
|
-
|
|
1217
|
-
"claude-3-sonnet-20240229",
|
|
1218
|
-
"claude-3-haiku-20240307",
|
|
1219
|
-
"claude-2.1",
|
|
1220
|
-
"claude-instant-1.2"
|
|
1221
|
-
];
|
|
1222
|
-
if (!this.config.apiKey) {
|
|
1429
|
+
this.availableModels = AI_PROVIDERS_CONFIG.anthropic.availableModels;
|
|
1430
|
+
if (!this.config.apiKey && !AI_PROVIDERS_CONFIG.anthropic.apiKey) {
|
|
1223
1431
|
console.warn("[Anthropic Provider] No API key provided - will use template responses");
|
|
1224
1432
|
}
|
|
1225
1433
|
}
|
|
@@ -11682,157 +11890,6 @@ var init_startup_display = __esm({
|
|
|
11682
11890
|
}
|
|
11683
11891
|
});
|
|
11684
11892
|
|
|
11685
|
-
// src/config/defaults.ts
|
|
11686
|
-
var DEFAULT_MEMORY_CONFIG, DEFAULT_ROUTER_CONFIG, DEFAULT_UI_CONFIG, DEFAULT_PATHS, DEFAULT_ENV_VARS, DEFAULT_DUMMY_VALUES, DEFAULT_HELP_TEXTS, DEFAULT_COMMAND_OUTPUTS, DEFAULT_FEATURE_FLAGS, DEFAULT_PROVIDER_PREFS, DEFAULT_STARTUP, DEFAULT_HSR_CONFIG, DEFAULT_RL_CONFIG, DEFAULT_CONFIG;
|
|
11687
|
-
var init_defaults = __esm({
|
|
11688
|
-
"src/config/defaults.ts"() {
|
|
11689
|
-
DEFAULT_MEMORY_CONFIG = {
|
|
11690
|
-
system1: {
|
|
11691
|
-
maxKnowledgeNodes: 1e3,
|
|
11692
|
-
embeddingDimension: 1536,
|
|
11693
|
-
cacheSize: 100,
|
|
11694
|
-
compressionThreshold: 0.75,
|
|
11695
|
-
accessDecayRate: 0.03
|
|
11696
|
-
},
|
|
11697
|
-
system2: {
|
|
11698
|
-
maxReasoningTraces: 100,
|
|
11699
|
-
qualityThreshold: 0.75,
|
|
11700
|
-
reflectionFrequency: 12,
|
|
11701
|
-
enhancementEvaluationInterval: 6
|
|
11702
|
-
},
|
|
11703
|
-
coordinator: {
|
|
11704
|
-
syncInterval: 5e3,
|
|
11705
|
-
conflictResolutionStrategy: "balanced",
|
|
11706
|
-
learningRate: 0.15,
|
|
11707
|
-
adaptationThreshold: 0.7
|
|
11708
|
-
},
|
|
11709
|
-
performance: {
|
|
11710
|
-
targetLatency: 50,
|
|
11711
|
-
maxMemoryUsage: 256,
|
|
11712
|
-
cacheStrategy: "lru",
|
|
11713
|
-
preloadPriority: "medium",
|
|
11714
|
-
backgroundOptimization: true,
|
|
11715
|
-
batchSize: 10
|
|
11716
|
-
}
|
|
11717
|
-
};
|
|
11718
|
-
DEFAULT_ROUTER_CONFIG = {
|
|
11719
|
-
confidenceThreshold: 0.7,
|
|
11720
|
-
enableLearning: true,
|
|
11721
|
-
supportedLanguages: ["en", "ja", "zh", "ko", "vi"],
|
|
11722
|
-
// Fixed language codes
|
|
11723
|
-
enableConfirmation: false,
|
|
11724
|
-
maxSuggestions: 5,
|
|
11725
|
-
enableFuzzyMatch: true
|
|
11726
|
-
};
|
|
11727
|
-
DEFAULT_UI_CONFIG = {
|
|
11728
|
-
width: 120,
|
|
11729
|
-
borderColor: "white",
|
|
11730
|
-
promptColor: "cyan",
|
|
11731
|
-
textColor: "white",
|
|
11732
|
-
promptSymbol: ">",
|
|
11733
|
-
placeholder: "Type your command or question here...",
|
|
11734
|
-
enablePasteDetection: true,
|
|
11735
|
-
enableFileDrop: true,
|
|
11736
|
-
showModeIndicator: true,
|
|
11737
|
-
animationSpeed: 100
|
|
11738
|
-
};
|
|
11739
|
-
DEFAULT_PATHS = {
|
|
11740
|
-
avatar: "face_only_96x96_ramp.txt"
|
|
11741
|
-
// 相対パスに変更
|
|
11742
|
-
};
|
|
11743
|
-
DEFAULT_ENV_VARS = {
|
|
11744
|
-
OPENAIAPI_KEY: "OPENAI_API_KEY",
|
|
11745
|
-
ANTHROPICAPI_KEY: "ANTHROPIC_API_KEY",
|
|
11746
|
-
GOOGLEAI_API_KEY: "GOOGLE_AI_API_KEY"
|
|
11747
|
-
};
|
|
11748
|
-
DEFAULT_DUMMY_VALUES = {
|
|
11749
|
-
lintScore: 94,
|
|
11750
|
-
typeCoverage: 87,
|
|
11751
|
-
securityScore: 89,
|
|
11752
|
-
defaultModel: "gpt-4 (openai)",
|
|
11753
|
-
cognitiveModes: 58
|
|
11754
|
-
// Updated to actual count
|
|
11755
|
-
};
|
|
11756
|
-
DEFAULT_HELP_TEXTS = {
|
|
11757
|
-
priorityUsage: "Usage: /priority <privacy-first|performance|cost-effective|auto>",
|
|
11758
|
-
githubAppUrl: "https://github.com/apps/maria-ai-assistant",
|
|
11759
|
-
cognitiveModesCount: 58,
|
|
11760
|
-
// Updated to actual count
|
|
11761
|
-
priorityModes: ["privacy-first", "performance", "cost-effective", "auto"]
|
|
11762
|
-
};
|
|
11763
|
-
DEFAULT_COMMAND_OUTPUTS = {
|
|
11764
|
-
setupEnvVars: [
|
|
11765
|
-
"export OPENAI_API_KEY=",
|
|
11766
|
-
"export ANTHROPIC_API_KEY=",
|
|
11767
|
-
"export GOOGLE_AI_API_KEY="
|
|
11768
|
-
],
|
|
11769
|
-
modelSelector: {
|
|
11770
|
-
currentModel: "gpt-4 (openai) - Default",
|
|
11771
|
-
provider: "openai"
|
|
11772
|
-
}
|
|
11773
|
-
};
|
|
11774
|
-
DEFAULT_FEATURE_FLAGS = {
|
|
11775
|
-
enableEnhancedInterface: true,
|
|
11776
|
-
enableStreaming: true,
|
|
11777
|
-
enableAutoRouting: true,
|
|
11778
|
-
enableBgOptimization: true,
|
|
11779
|
-
enableMemorySystem: true,
|
|
11780
|
-
enableRLEvolution: true,
|
|
11781
|
-
enableHSRSystem: true,
|
|
11782
|
-
enableVisionAnalysis: true,
|
|
11783
|
-
enableProactiveReporting: true
|
|
11784
|
-
};
|
|
11785
|
-
DEFAULT_PROVIDER_PREFS = {
|
|
11786
|
-
provider: "openai",
|
|
11787
|
-
model: "gpt-4",
|
|
11788
|
-
offline: false,
|
|
11789
|
-
debug: false,
|
|
11790
|
-
priority: "auto",
|
|
11791
|
-
maxTokens: 4096,
|
|
11792
|
-
temperature: 0.7
|
|
11793
|
-
};
|
|
11794
|
-
DEFAULT_STARTUP = {
|
|
11795
|
-
showLogo: true,
|
|
11796
|
-
totalProvidersHint: 8,
|
|
11797
|
-
startupTimeout: 1e4,
|
|
11798
|
-
showLoadingAnimation: true,
|
|
11799
|
-
checkForUpdates: false
|
|
11800
|
-
};
|
|
11801
|
-
DEFAULT_HSR_CONFIG = {
|
|
11802
|
-
enableHumanInterruption: true,
|
|
11803
|
-
interruptionCheckInterval: 100,
|
|
11804
|
-
autoSaveInterval: 3e4,
|
|
11805
|
-
maxOperationTime: 6e4,
|
|
11806
|
-
safetyLevel: "medium"
|
|
11807
|
-
};
|
|
11808
|
-
DEFAULT_RL_CONFIG = {
|
|
11809
|
-
learningRate: 1e-3,
|
|
11810
|
-
discountFactor: 0.99,
|
|
11811
|
-
epsilon: 0.1,
|
|
11812
|
-
replayBufferSize: 1e4,
|
|
11813
|
-
batchSize: 32,
|
|
11814
|
-
updateFrequency: 100,
|
|
11815
|
-
enablePPO: true,
|
|
11816
|
-
enableDPO: false
|
|
11817
|
-
};
|
|
11818
|
-
DEFAULT_CONFIG = {
|
|
11819
|
-
memory: DEFAULT_MEMORY_CONFIG,
|
|
11820
|
-
router: DEFAULT_ROUTER_CONFIG,
|
|
11821
|
-
ui: DEFAULT_UI_CONFIG,
|
|
11822
|
-
paths: DEFAULT_PATHS,
|
|
11823
|
-
env: DEFAULT_ENV_VARS,
|
|
11824
|
-
values: DEFAULT_DUMMY_VALUES,
|
|
11825
|
-
help: DEFAULT_HELP_TEXTS,
|
|
11826
|
-
command: DEFAULT_COMMAND_OUTPUTS,
|
|
11827
|
-
flags: DEFAULT_FEATURE_FLAGS,
|
|
11828
|
-
provider: DEFAULT_PROVIDER_PREFS,
|
|
11829
|
-
startup: DEFAULT_STARTUP,
|
|
11830
|
-
hsr: DEFAULT_HSR_CONFIG,
|
|
11831
|
-
rl: DEFAULT_RL_CONFIG
|
|
11832
|
-
};
|
|
11833
|
-
}
|
|
11834
|
-
});
|
|
11835
|
-
|
|
11836
11893
|
// src/config/loader.ts
|
|
11837
11894
|
function loadCompleteConfig(options2 = {}) {
|
|
11838
11895
|
const _config = { ...DEFAULT_CONFIG };
|
|
@@ -22207,16 +22264,8 @@ var init_cli = __esm({
|
|
|
22207
22264
|
init_chat_context_fixed_service();
|
|
22208
22265
|
init_ai_response_service();
|
|
22209
22266
|
init_slash_command_handler();
|
|
22210
|
-
|
|
22211
|
-
|
|
22212
|
-
const packagePath = path17.join(process.cwd(), "package.json");
|
|
22213
|
-
const pkgContent = fs11.readFileSync(packagePath, "utf8");
|
|
22214
|
-
const pkg2 = JSON.parse(pkgContent);
|
|
22215
|
-
if (pkg2.version) {
|
|
22216
|
-
packageJson2.version = pkg2.version;
|
|
22217
|
-
}
|
|
22218
|
-
} catch {
|
|
22219
|
-
}
|
|
22267
|
+
init_defaults();
|
|
22268
|
+
packageJson2 = { version: APP_VERSION };
|
|
22220
22269
|
sessionMemory = [];
|
|
22221
22270
|
program = createCLI();
|
|
22222
22271
|
program.parse(process.argv);
|