ruflo 3.6.30 → 3.7.0-alpha.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "ruflo",
3
- "version": "3.6.30",
3
+ "version": "3.7.0-alpha.2",
4
4
  "description": "Ruflo - Enterprise AI agent orchestration platform. Deploy 60+ specialized agents in coordinated swarms with self-learning, fault-tolerant consensus, vector memory, and MCP integration",
5
5
  "main": "bin/ruflo.js",
6
6
  "type": "module",
@@ -39,7 +39,7 @@
39
39
  "package:rvf": "bash src/scripts/package-rvf.sh"
40
40
  },
41
41
  "dependencies": {
42
- "@claude-flow/cli": ">=3.0.0-alpha.1"
42
+ "@claude-flow/cli": "^3.7.0-alpha.1"
43
43
  },
44
44
  "overrides": {
45
45
  "@ruvector/rvf-wasm": "0.1.5",
package/src/ruvocal/.env DELETED
@@ -1,194 +0,0 @@
1
- # Use .env.local to change these variables
2
- # DO NOT EDIT THIS FILE WITH SENSITIVE DATA
3
-
4
- ### Models ###
5
- # Models are sourced exclusively from an OpenAI-compatible base URL.
6
- # Example: https://router.huggingface.co/v1
7
- OPENAI_BASE_URL=https://router.huggingface.co/v1
8
-
9
- # Canonical auth token for any OpenAI-compatible provider
10
- OPENAI_API_KEY=#your provider API key (works for HF router, OpenAI, LM Studio, etc.).
11
- # When set to true, user token will be used for inference calls
12
- USE_USER_TOKEN=false
13
- # Automatically redirect to oauth login page if user is not logged in, when set to "true"
14
- AUTOMATIC_LOGIN=false
15
-
16
- ### PostgreSQL (RuVector) ###
17
- DATABASE_URL=#postgresql://ruvocal:password@localhost:5432/ruvocal
18
- # Legacy MongoDB vars (unused — kept for reference)
19
- # MONGODB_URL=
20
- # MONGODB_DB_NAME=chat-ui
21
- # MONGODB_DIRECT_CONNECTION=false
22
-
23
-
24
- ## Public app configuration ##
25
- PUBLIC_APP_NAME=ChatUI # name used as title throughout the app
26
- PUBLIC_APP_ASSETS=chatui # used to find logos & favicons in static/$PUBLIC_APP_ASSETS
27
- PUBLIC_APP_DESCRIPTION="Making the community's best AI chat models available to everyone."# description used throughout the app
28
- PUBLIC_ORIGIN=
29
- PUBLIC_SHARE_PREFIX=
30
- PUBLIC_GOOGLE_ANALYTICS_ID=
31
- PUBLIC_PLAUSIBLE_SCRIPT_URL=
32
- PUBLIC_APPLE_APP_ID=
33
-
34
- COUPLE_SESSION_WITH_COOKIE_NAME=
35
- # when OPEN_ID is configured, users are required to login after the welcome modal
36
- OPENID_CLIENT_ID="" # You can set to "__CIMD__" for automatic oauth app creation when deployed, see https://datatracker.ietf.org/doc/draft-ietf-oauth-client-id-metadata-document/
37
- OPENID_CLIENT_SECRET=
38
- OPENID_SCOPES="openid profile inference-api read-mcp read-billing"
39
- USE_USER_TOKEN=
40
- AUTOMATIC_LOGIN=# if true authentication is required on all routes
41
-
42
- ### Local Storage ###
43
- MONGO_STORAGE_PATH= # where is the db folder stored
44
-
45
- ## Models overrides
46
- MODELS=
47
-
48
- ## Task model
49
- # Optional: set to the model id/name from the `${OPENAI_BASE_URL}/models` list
50
- # to use for internal tasks (title summarization, etc). If not set, the current model will be used
51
- TASK_MODEL=
52
-
53
- # Arch router (OpenAI-compatible) endpoint base URL used for route selection
54
- # Example: https://api.openai.com/v1 or your hosted Arch endpoint
55
- LLM_ROUTER_ARCH_BASE_URL=
56
-
57
- ## LLM Router Configuration
58
- # Path to routes policy (JSON array). Required when the router is enabled; must point to a valid JSON file.
59
- LLM_ROUTER_ROUTES_PATH=
60
-
61
- # Model used at the Arch router endpoint for selection
62
- LLM_ROUTER_ARCH_MODEL=
63
-
64
- # Fallback behavior
65
- # Route to map "other" to (must exist in routes file)
66
- LLM_ROUTER_OTHER_ROUTE=casual_conversation
67
- # Model to call if the Arch selection fails entirely
68
- LLM_ROUTER_FALLBACK_MODEL=
69
- # Arch selection timeout in milliseconds (default 10000)
70
- LLM_ROUTER_ARCH_TIMEOUT_MS=10000
71
- # Maximum length (in characters) for assistant messages sent to router for route selection (default 500)
72
- LLM_ROUTER_MAX_ASSISTANT_LENGTH=500
73
- # Maximum length (in characters) for previous user messages sent to router (latest user message not trimmed, default 400)
74
- LLM_ROUTER_MAX_PREV_USER_LENGTH=400
75
-
76
- # Enable router multimodal handling (set to true to allow image inputs via router)
77
- LLM_ROUTER_ENABLE_MULTIMODAL=
78
- # Required when LLM_ROUTER_ENABLE_MULTIMODAL=true: id or name of the multimodal model to use for image requests
79
- LLM_ROUTER_MULTIMODAL_MODEL=
80
-
81
- # Enable router tool support (set to true to allow tool calling via router)
82
- LLM_ROUTER_ENABLE_TOOLS=
83
- # Required when tools are active: id or name of the model to use for MCP tool calls.
84
- LLM_ROUTER_TOOLS_MODEL=
85
-
86
- # Router UI overrides (client-visible)
87
- # Public display name for the router entry in the model list. Defaults to "Omni".
88
- PUBLIC_LLM_ROUTER_DISPLAY_NAME=Omni
89
- # Optional: public logo URL for the router entry. If unset, the UI shows a Carbon icon.
90
- PUBLIC_LLM_ROUTER_LOGO_URL=
91
- # Public alias id used for the virtual router model (Omni). Defaults to "omni".
92
- PUBLIC_LLM_ROUTER_ALIAS_ID=omni
93
-
94
- ### Transcription ###
95
- # Voice-to-text transcription using Whisper models
96
- # If set, enables the microphone button in the chat input
97
- # Example: openai/whisper-large-v3-turbo
98
- TRANSCRIPTION_MODEL=
99
- # Optional: Base URL for transcription API (defaults to HF inference)
100
- # Default: https://router.huggingface.co/hf-inference/models
101
- TRANSCRIPTION_BASE_URL=
102
-
103
- ### Authentication ###
104
- # Parameters to enable open id login
105
- OPENID_CONFIG=
106
- # if it's defined, only these emails will be allowed to use login
107
- ALLOWED_USER_EMAILS=[]
108
- # If it's defined, users with emails matching these domains will also be allowed to use login
109
- ALLOWED_USER_DOMAINS=[]
110
- # valid alternative redirect URLs for OAuth, used for HuggingChat apps
111
- ALTERNATIVE_REDIRECT_URLS=[]
112
- ### Cookies
113
- # name of the cookie used to store the session
114
- COOKIE_NAME=hf-chat
115
- # If the value of this cookie changes, the session is destroyed. Useful if chat-ui is deployed on a subpath
116
- # of your domain, and you want chat ui sessions to reset if the user's auth changes
117
- COUPLE_SESSION_WITH_COOKIE_NAME=
118
- # specify secure behaviour for cookies
119
- COOKIE_SAMESITE=# can be "lax", "strict", "none" or left empty
120
- COOKIE_SECURE=# set to true to only allow cookies over https
121
- TRUSTED_EMAIL_HEADER=# header to use to get the user email, only use if you know what you are doing
122
-
123
- ### Admin stuff ###
124
- ADMIN_CLI_LOGIN=true # set to false to disable the CLI login
125
- ADMIN_TOKEN=#We recommend leaving this empty, you can get the token from the terminal.
126
-
127
- ### Feature Flags ###
128
- LLM_SUMMARIZATION=true # generate conversation titles with LLMs
129
-
130
- ALLOW_IFRAME=true # Allow the app to be embedded in an iframe
131
-
132
- # Base servers list (JSON array). Example: MCP_SERVERS=[{"name": "Web Search (Exa)", "url": "https://mcp.exa.ai/mcp"}, {"name": "Hugging Face", "url": "https://hf.co/mcp"}]
133
- MCP_SERVERS=
134
- # When true, forward the logged-in user's Hugging Face access token
135
- MCP_FORWARD_HF_USER_TOKEN=
136
- # Exa API key (injected at runtime into mcp.exa.ai URLs as ?exaApiKey=)
137
- EXA_API_KEY=
138
- # Timeout in milliseconds for MCP tool calls (default: 120000 = 2 minutes)
139
- MCP_TOOL_TIMEOUT_MS=
140
- ENABLE_DATA_EXPORT=true
141
-
142
- ### Rate limits ###
143
- # See `src/lib/server/usageLimits.ts`
144
- # {
145
- # conversations: number, # how many conversations
146
- # messages: number, # how many messages in a conversation
147
- # assistants: number, # how many assistants
148
- # messageLength: number, # how long can a message be before we cut it off
149
- # messagesPerMinute: number, # how many messages per minute
150
- # tools: number # how many tools
151
- # }
152
- USAGE_LIMITS={}
153
-
154
- ### HuggingFace specific ###
155
- ## Feature flag & admin settings
156
- # Used for setting early access & admin flags to users
157
- HF_ORG_ADMIN=
158
- HF_ORG_EARLY_ACCESS=
159
- WEBHOOK_URL_REPORT_ASSISTANT=#provide slack webhook url to get notified for reports/feature requests
160
-
161
-
162
- ### Metrics ###
163
- METRICS_ENABLED=false
164
- METRICS_PORT=5565
165
- LOG_LEVEL=info
166
-
167
-
168
- ### Parquet export ###
169
- # Not in use anymore but useful to export conversations to a parquet file as a HuggingFace dataset
170
- PARQUET_EXPORT_DATASET=
171
- PARQUET_EXPORT_HF_TOKEN=
172
- ADMIN_API_SECRET=# secret to admin API calls, like computing usage stats or exporting parquet data
173
-
174
- ### Config ###
175
- ENABLE_CONFIG_MANAGER=true
176
-
177
- ### Docker build variables ###
178
- # These values cannot be updated at runtime
179
- # They need to be passed when building the docker image
180
- # See https://github.com/huggingface/chat-ui/main/.github/workflows/deploy-prod.yml#L44-L47
181
- APP_BASE="" # base path of the app, e.g. /chat, left blank as default
182
- ### Body size limit for SvelteKit https://svelte.dev/docs/kit/adapter-node#Environment-variables-BODY_SIZE_LIMIT
183
- BODY_SIZE_LIMIT=15728640
184
- PUBLIC_COMMIT_SHA=
185
-
186
- ### LEGACY parameters
187
- ALLOW_INSECURE_COOKIES=false # LEGACY! Use COOKIE_SECURE and COOKIE_SAMESITE instead
188
- PARQUET_EXPORT_SECRET=#DEPRECATED, use ADMIN_API_SECRET instead
189
- RATE_LIMIT= # /!\ DEPRECATED definition of messages per minute. Use USAGE_LIMITS.messagesPerMinute instead
190
- OPENID_NAME_CLAIM="name" # Change to "username" for some providers that do not provide name
191
- OPENID_PROVIDER_URL=https://huggingface.co # for Google, use https://accounts.google.com
192
- OPENID_TOLERANCE=
193
- OPENID_RESOURCE=
194
- EXPOSE_API=# deprecated, API is now always exposed
@@ -1 +0,0 @@
1
- MONGODB_URL=mongodb://localhost:27017/