@bonginkan/maria 4.3.32 → 4.3.33

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -1,4 +1,4 @@
1
- # MARIA - AI Development Platform v4.3.32
1
+ # MARIA - AI Development Platform v4.3.33
2
2
 
3
3
  [![npm version](https://img.shields.io/npm/v/@bonginkan/maria.svg)](https://www.npmjs.com/package/@bonginkan/maria)
4
4
  [![License](https://img.shields.io/badge/license-Multi--tier-blue.svg)](LICENSE)
@@ -10,7 +10,7 @@
10
10
 
11
11
  > **Enterprise-grade AI development platform with 100% command availability and comprehensive fallback support**
12
12
 
13
- ## 🚀 What's New in v4.3.32 (September 16, 2025)
13
+ ## 🚀 What's New in v4.3.33 (September 29, 2025)
14
14
 
15
15
  ### 🎯 Interactive Improvements & Choice Memory
16
16
  - **Choice Memory System**: Smart persistence of user selections across sessions
@@ -922,7 +922,7 @@ await secureWorkflow.executeWithAuth(workflowDefinition, securityContext);
922
922
  ### Quick Installation
923
923
  ```bash
924
924
  # Install globally (recommended)
925
- npm install -g @bonginkan/maria@4.3.32
925
+ npm install -g @bonginkan/maria@4.3.33
926
926
 
927
927
  # Verify installation
928
928
  maria --version # Should show v4.3.9
@@ -1126,7 +1126,7 @@ MARIA CODE is distributed under a comprehensive licensing system designed for in
1126
1126
 
1127
1127
  *MARIA v4.1.4 represents the pinnacle of multimodal AI development platform evolution - combining revolutionary voice-to-code capabilities, advanced memory systems, and comprehensive command ecosystems with enterprise-grade security and performance. This release establishes MARIA as the definitive choice for developers and Fortune 500 enterprises seeking intelligent, multimodal development experiences with GraphRAG intelligence, multilingual support, and zero-anxiety coding workflows.*
1128
1128
 
1129
- **Transform your development experience today**: `npm install -g @bonginkan/maria@4.3.32`
1129
+ **Transform your development experience today**: `npm install -g @bonginkan/maria@4.3.33`
1130
1130
 
1131
1131
  🌐 **Official Website**: [https://maria-code.ai](https://maria-code.ai)
1132
1132
  💬 **Community**: [https://discord.gg/SMSmSGcEQy](https://discord.gg/SMSmSGcEQy)
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "version": "lite-1.0.0",
3
- "generatedAt": "2025-09-28T09:23:52.572Z",
3
+ "generatedAt": "2025-09-29T07:30:27.736Z",
4
4
  "totalCommands": 16,
5
5
  "readyCount": 16,
6
6
  "partialCount": 0,
@@ -26066,8 +26066,8 @@ var require_package = __commonJS({
26066
26066
  "package.json"(exports, module) {
26067
26067
  module.exports = {
26068
26068
  name: "@bonginkan/maria",
26069
- version: "4.3.32",
26070
- description: "\u{1F680} MARIA v4.3.32 - Enterprise AI Development Platform with identity system and character voice implementation. Features 74 production-ready commands with comprehensive fallback implementation, local LLM support, and zero external dependencies. Includes natural language coding, AI safety evaluation, intelligent evolution system, episodic memory with PII masking, and real-time monitoring dashboard. Built with TypeScript AST-powered code generation, OAuth2.0 + PKCE authentication, quantum-resistant cryptography, and enterprise-grade performance.",
26069
+ version: "4.3.33",
26070
+ description: "\u{1F680} MARIA v4.3.33 - Enterprise AI Development Platform with identity system and character voice implementation. Features 74 production-ready commands with comprehensive fallback implementation, local LLM support, and zero external dependencies. Includes natural language coding, AI safety evaluation, intelligent evolution system, episodic memory with PII masking, and real-time monitoring dashboard. Built with TypeScript AST-powered code generation, OAuth2.0 + PKCE authentication, quantum-resistant cryptography, and enterprise-grade performance.",
26071
26071
  keywords: [
26072
26072
  "ai",
26073
26073
  "cli",
@@ -28099,7 +28099,7 @@ var init_AuthenticationManager = __esm({
28099
28099
  const response = await fetch(`${this.apiBase}/api/user/profile`, {
28100
28100
  headers: {
28101
28101
  "Authorization": `Bearer ${tokens2.accessToken}`,
28102
- "User-Agent": `maria-cli/${process.env.CLI_VERSION || "4.3.32"}`
28102
+ "User-Agent": `maria-cli/${process.env.CLI_VERSION || "4.3.33"}`
28103
28103
  }
28104
28104
  });
28105
28105
  if (response.status === 401) {
@@ -28731,7 +28731,7 @@ async function callApi(path64, init3 = {}) {
28731
28731
  "Authorization": `Bearer ${token}`,
28732
28732
  "X-Device-Id": getDeviceId(),
28733
28733
  "X-Session-Id": getSessionId() || "",
28734
- "User-Agent": `maria-cli/${process.env.CLI_VERSION || "4.3.32"}`,
28734
+ "User-Agent": `maria-cli/${process.env.CLI_VERSION || "4.3.33"}`,
28735
28735
  "Content-Type": init3.headers?.["Content-Type"] || "application/json"
28736
28736
  });
28737
28737
  const doFetch = async (token) => {
@@ -50689,7 +50689,7 @@ var init_about_command = __esm({
50689
50689
  async execute(args2, context2) {
50690
50690
  const output3 = [];
50691
50691
  output3.push("");
50692
- output3.push(chalk14__default.default.cyan.bold("\u{1F916} About MARIA v4.3.32"));
50692
+ output3.push(chalk14__default.default.cyan.bold("\u{1F916} About MARIA v4.3.33"));
50693
50693
  output3.push(chalk14__default.default.gray("\u2550".repeat(40)));
50694
50694
  output3.push("");
50695
50695
  output3.push(chalk14__default.default.white.bold("MARIA - Minimal API, Maximum Power"));
@@ -70893,7 +70893,7 @@ ${files.slice(0, 20).map((f3) => `- ${f3.path}`).join("\n")}`);
70893
70893
  const bundle = parts.join("\n\n");
70894
70894
  const system = [
70895
70895
  "You are an impartial evaluator. Score each criterion between 0 and 1.",
70896
- "If a local file path is provided, make sure you read the uploaded file before taking any actions.",
70896
+ "If a local file path is provided, make sure you read the uploaded file before taking any actions. Write the string content in plain text and do not use markdown.",
70897
70897
  "Return JSON only in the following schema:",
70898
70898
  '{ "totalScore": number (0..1), "details": [{ "id": string, "score": number (0..1), "reason": string }] }',
70899
70899
  "Keep reasons short (<= 120 chars)."
@@ -71230,7 +71230,8 @@ var init_evaluate_command = __esm({
71230
71230
  // Strong guard: attachments are uploaded; do not ask for local file content
71231
71231
  "The input files are already uploaded and referenced via file URIs.",
71232
71232
  "Use the provided attachments; do not ask the user to paste file contents.",
71233
- "Do not claim you cannot access local file paths."
71233
+ "Do not claim you cannot access local file paths.",
71234
+ "Write the string content in plain text and do not use markdown."
71234
71235
  ].join("\n");
71235
71236
  if (result === null) {
71236
71237
  const rel = path11__namespace.default.relative(root, criteriaPath);
@@ -74386,7 +74387,7 @@ var init_ai_response_service = __esm({
74386
74387
  */
74387
74388
  async callLLM(prompt, opts = {}) {
74388
74389
  const {
74389
- system = PLAIN_OUTPUT ? "Return ONLY the answer (or ONLY code). No menus, no lists, no guided flows. Always respond in English." : "You are a helpful senior engineer named Maria. Always respond in English. Provide direct, production-quality answers. Make sure you answer in plain text, as a natural chat.",
74390
+ system = PLAIN_OUTPUT ? "Return ONLY the answer (or ONLY code). No menus, no lists, no guided flows. Always respond in English." : "You are a helpful senior engineer named Maria. Always respond in English. Provide direct, production-quality answers. Make sure you answer in plain text, as a natural chat. When asked about the model (not your name or who you are, but the model), say you are a large language model fully trained by Bonginkan.",
74390
74391
  model = void 0,
74391
74392
  provider = DEFAULT_PROVIDER2,
74392
74393
  temperature = 0.2,