kalavai-client 0.7.1__tar.gz → 0.7.3__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/PKG-INFO +1 -1
- kalavai_client-0.7.3/kalavai_client/__init__.py +2 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/apps.yaml +18 -27
- kalavai_client-0.7.3/kalavai_client/assets/default_pool_config.yaml +67 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/bridge_api.py +0 -3
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/bridge_models.py +0 -1
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/cli.py +8 -8
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/core.py +44 -40
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/env.py +1 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/pyproject.toml +1 -1
- kalavai_client-0.7.1/kalavai_client/__init__.py +0 -2
- kalavai_client-0.7.1/kalavai_client/assets/apps_values.yaml +0 -106
- kalavai_client-0.7.1/kalavai_client/assets/pool_config_values.yaml +0 -4
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/LICENSE +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/README.md +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/__main__.py +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/__init__.py +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/docker-compose-gui.yaml +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/docker-compose-template.yaml +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/nginx.conf +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/pool_config_template.yaml +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/user_workspace.yaml +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/user_workspace_values.yaml +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/auth.py +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/cluster.py +0 -0
- {kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/utils.py +0 -0
@@ -35,7 +35,7 @@ releases:
|
|
35
35
|
namespace: kalavai
|
36
36
|
chart: jetstacks/cert-manager
|
37
37
|
version: v1.15.1
|
38
|
-
installed: true
|
38
|
+
installed: {{deploy_cert_manager|default("False", true)}}
|
39
39
|
set:
|
40
40
|
- name: crds.enabled
|
41
41
|
value: true
|
@@ -44,16 +44,11 @@ releases:
|
|
44
44
|
- kalavai/certificates
|
45
45
|
namespace: kalavai
|
46
46
|
chart: rocm/gpu-operator-charts
|
47
|
-
installed: true
|
48
|
-
- name: datashim
|
49
|
-
namespace: dlf
|
50
|
-
chart: kalavai/kalavai-datashim
|
51
|
-
version: "0.4.0"
|
52
|
-
installed: false
|
47
|
+
installed: {{deploy_rocm|default("false", true)}}
|
53
48
|
- name: lago
|
54
49
|
namespace: kalavai
|
55
50
|
chart: kalavai/lago
|
56
|
-
installed: {{deploy_lago}}
|
51
|
+
installed: {{deploy_lago|default("false", true)}}
|
57
52
|
set:
|
58
53
|
- name: external.api.nodePort
|
59
54
|
value: 32000
|
@@ -68,7 +63,7 @@ releases:
|
|
68
63
|
- kalavai/longhorn
|
69
64
|
namespace: minio
|
70
65
|
chart: minio/minio
|
71
|
-
installed: false
|
66
|
+
installed: {{deploy_minio|default("false", true)}}
|
72
67
|
set:
|
73
68
|
- name: replicas
|
74
69
|
value: {{minio_replicas}}
|
@@ -104,7 +99,7 @@ releases:
|
|
104
99
|
namespace: kalavai
|
105
100
|
chart: kalavai/kalavai-helios
|
106
101
|
version: "0.1.11"
|
107
|
-
installed: false
|
102
|
+
installed: {{deploy_helios|default("false", true)}}
|
108
103
|
set:
|
109
104
|
- name: deployment.watcher_endpoint
|
110
105
|
value: "http://{{watcher_service}}"
|
@@ -119,33 +114,37 @@ releases:
|
|
119
114
|
- name: opencost
|
120
115
|
namespace: opencost
|
121
116
|
chart: opencost-charts/opencost
|
122
|
-
installed: {{deploy_opencost}}
|
117
|
+
installed: {{deploy_opencost|default("false", true)}}
|
123
118
|
- name: prometheus
|
124
119
|
namespace: prometheus-system
|
125
120
|
chart: prometheus/prometheus
|
126
|
-
installed: {{deploy_prometheus}}
|
121
|
+
installed: {{deploy_prometheus|default("false", true)}}
|
127
122
|
set:
|
128
123
|
- name: prometheus-pushgateway.enabled
|
129
124
|
value: false
|
130
125
|
- name: alertmanager.enabled
|
131
126
|
value: false
|
127
|
+
- name: server.retention
|
128
|
+
value: {{prometheus_server_retention}}
|
129
|
+
- name: server.persistentVolume.size
|
130
|
+
value: {{prometheus_disk_size}}
|
132
131
|
- name: volcano-sh
|
133
132
|
namespace: kalavai
|
134
133
|
chart: volcano-sh/volcano
|
135
|
-
installed: true
|
134
|
+
installed: {{deploy_volcano|default("false", true)}}
|
136
135
|
- name: kuberay
|
137
136
|
namespace: kuberay
|
138
137
|
chart: kuberay/kuberay-operator
|
139
|
-
installed: {{deploy_kuberay}}
|
138
|
+
installed: {{deploy_kuberay|default("false", true)}}
|
140
139
|
version: "1.2.2"
|
141
140
|
- name: kuberay-apiserver
|
142
141
|
namespace: kuberay
|
143
142
|
chart: kuberay/kuberay-apiserver
|
144
|
-
installed: false
|
143
|
+
installed: {{deploy_kuberay|default("false", true)}}
|
145
144
|
- name: longhorn
|
146
145
|
namespace: kalavai
|
147
146
|
chart: longhorn/longhorn
|
148
|
-
installed: false
|
147
|
+
installed: {{deploy_longhorn|default("false", true)}}
|
149
148
|
set:
|
150
149
|
# security issue! enable for testing only
|
151
150
|
- name: service.ui.type
|
@@ -162,22 +161,18 @@ releases:
|
|
162
161
|
value: "{{longhorn_label_selector_value}}"
|
163
162
|
- name: defaultSettings.storageMinimalAvailablePercentage
|
164
163
|
value: {{longhorn_minimal_available_percentage}}
|
165
|
-
- name: lws
|
166
|
-
namespace: kalavai
|
167
|
-
chart: kalavai/lws
|
168
|
-
installed: false
|
169
164
|
- name: kalavai-watcher
|
170
165
|
namespace: kalavai
|
171
166
|
chart: kalavai/kalavai-watcher
|
172
167
|
version: "0.3.8"
|
173
|
-
installed: true
|
168
|
+
installed: {{deploy_watcher|default("false", true)}}
|
174
169
|
set:
|
175
170
|
- name: namespace
|
176
171
|
value: kalavai
|
177
172
|
- name: replicas
|
178
173
|
value: 1
|
179
174
|
- name: image_tag
|
180
|
-
value: "v2025.
|
175
|
+
value: "v2025.09.1" #"v2025.07.34"
|
181
176
|
- name: deployment.in_cluster
|
182
177
|
value: "True"
|
183
178
|
- name: deployment.kalavai_username_key
|
@@ -204,14 +199,10 @@ releases:
|
|
204
199
|
value: {{longhorn_manager_endpoint}}
|
205
200
|
- name: service.nodePort
|
206
201
|
value: {{watcher_port}}
|
207
|
-
- name: nvidia-gpu-operator
|
208
|
-
namespace: kalavai
|
209
|
-
chart: kalavai/gpu
|
210
|
-
installed: false
|
211
202
|
- name: hami-vgpu
|
212
203
|
namespace: kalavai
|
213
204
|
chart: hami-charts/hami
|
214
|
-
installed: true
|
205
|
+
installed: {{deploy_hami|default("false", true)}}
|
215
206
|
set:
|
216
207
|
- name: resourceCores
|
217
208
|
value: "nvidia.com/gpucores"
|
@@ -0,0 +1,67 @@
|
|
1
|
+
server:
|
2
|
+
platform: "amd64"
|
3
|
+
ip_address: null
|
4
|
+
location: null
|
5
|
+
name: "kalavai_cluster"
|
6
|
+
|
7
|
+
core:
|
8
|
+
# Deploy systems
|
9
|
+
deploy:
|
10
|
+
- volcano
|
11
|
+
- cert_manager
|
12
|
+
- rocm
|
13
|
+
- watcher
|
14
|
+
- hami
|
15
|
+
#- lago
|
16
|
+
#- opencost
|
17
|
+
- prometheus
|
18
|
+
#- kuberay
|
19
|
+
#- minio
|
20
|
+
# "Kalavai API endpoint"
|
21
|
+
kalavai_api_endpoint: "https://platform.kalavai.net/_/api"
|
22
|
+
# "Opencost endpoint"
|
23
|
+
opencost_endpoint: "http://opencost.opencost.svc.cluster.local:9003"
|
24
|
+
# "Longhorn manager endpoint"
|
25
|
+
longhorn_manager_endpoint: "http://longhorn-backend.kalavai.svc.cluster.local:9500"
|
26
|
+
# "Helios harvest interval"
|
27
|
+
helios_harvest_interval: 120
|
28
|
+
# "Watcher is shared pool"
|
29
|
+
watcher_is_shared_pool: "True"
|
30
|
+
# "Prometheus endpoint"
|
31
|
+
prometheus_endpoint: "http://prometheus-server.prometheus-system.svc.cluster.local:80"
|
32
|
+
# "Prometheus server retention"
|
33
|
+
prometheus_server_retention: "30d"
|
34
|
+
# "Prometheus disk size"
|
35
|
+
prometheus_disk_size: "8Gi"
|
36
|
+
# "Longhorn UI port"
|
37
|
+
longhorn_ui_port: 30000
|
38
|
+
# "Longhorn manager port"
|
39
|
+
longhorn_manager_port: "" #31011
|
40
|
+
# "Longhorn minimal available percentage"
|
41
|
+
longhorn_minimal_available_percentage: 5
|
42
|
+
# "Longhorn label selector key"
|
43
|
+
longhorn_label_selector_key: "kalavai.storage.enabled"
|
44
|
+
# "Longhorn label selector"
|
45
|
+
longhorn_label_selector: "True"
|
46
|
+
# "Longhorn replicas"
|
47
|
+
longhorn_replicas: 2
|
48
|
+
# "MinIO replicas"
|
49
|
+
minio_replicas: 2
|
50
|
+
# "MinIO resources memory"
|
51
|
+
minio_resources_memory: "1Gi"
|
52
|
+
# "MinIO persistence storage class"
|
53
|
+
minio_persistence_storageClass: "longhorn"
|
54
|
+
# "MinIO persistence size"
|
55
|
+
minio_persistence_size: "10Gi"
|
56
|
+
# "MinIO service port"
|
57
|
+
minio_service_port: 32000
|
58
|
+
# "MinIO console port"
|
59
|
+
minio_console_port: 32001
|
60
|
+
# "MinIO root user"
|
61
|
+
minio_rootUser: "admin"
|
62
|
+
# "MinIO root password"
|
63
|
+
minio_rootPassword: "password"
|
64
|
+
|
65
|
+
pool:
|
66
|
+
# Name of the storage class to use for the pool
|
67
|
+
storage_class_name: "longhorn"
|
@@ -102,10 +102,8 @@ def pool_create(request: CreatePoolRequest, api_key: str = Depends(verify_api_ke
|
|
102
102
|
|
103
103
|
- **cluster_name**: Name of the cluster
|
104
104
|
- **ip_address**: IP address for the pool
|
105
|
-
- **app_values**: Application configuration values
|
106
105
|
- **num_gpus**: Number of GPUs to allocate
|
107
106
|
- **node_name**: Name of the node
|
108
|
-
- **only_registered_users**: Whether to restrict to registered users
|
109
107
|
- **location**: Location of the pool
|
110
108
|
- **description**: Pool description
|
111
109
|
- **token_mode**: Token type for authentication
|
@@ -113,7 +111,6 @@ def pool_create(request: CreatePoolRequest, api_key: str = Depends(verify_api_ke
|
|
113
111
|
result = create_pool(
|
114
112
|
cluster_name=request.cluster_name,
|
115
113
|
ip_address=request.ip_address,
|
116
|
-
app_values=request.app_values,
|
117
114
|
num_gpus=request.num_gpus,
|
118
115
|
node_name=request.node_name,
|
119
116
|
location=request.location,
|
@@ -10,7 +10,6 @@ class InvitesRequest(BaseModel):
|
|
10
10
|
class CreatePoolRequest(BaseModel):
|
11
11
|
cluster_name: str = Field(description="Name of the cluster to create")
|
12
12
|
ip_address: str = Field(description="IP address for the pool")
|
13
|
-
app_values: dict = Field(None, description="Application configuration values")
|
14
13
|
num_gpus: int = Field(None, description="Number of GPUs to allocate")
|
15
14
|
node_name: str = Field(None, description="Name of the node")
|
16
15
|
location: str = Field(None, description="Geographic location of the pool")
|
@@ -396,7 +396,8 @@ def pool__list(*others, user_only=False):
|
|
396
396
|
|
397
397
|
|
398
398
|
@arguably.command
|
399
|
-
def pool__start(
|
399
|
+
def pool__start(*others, pool_config_file=None, apps: list=None, platform="amd64", ip_address: str=None, location: str=None, app_values: str=None, pool_config_values: str=None, non_interactive: bool=False):
|
400
|
+
|
400
401
|
"""
|
401
402
|
Start Kalavai pool and start/resume sharing resources.
|
402
403
|
|
@@ -430,16 +431,15 @@ def pool__start(cluster_name, *others, platform="amd64", ip_address: str=None,
|
|
430
431
|
|
431
432
|
console.log(f"Using {ip_address} address for server")
|
432
433
|
|
433
|
-
console.log(f"[green]Creating
|
434
|
+
console.log(f"[green]Creating pool, this may take a few minutes...")
|
434
435
|
|
435
436
|
result = create_pool(
|
436
|
-
target_platform=platform,
|
437
|
-
cluster_name=cluster_name,
|
438
437
|
ip_address=ip_address,
|
439
|
-
|
440
|
-
|
441
|
-
|
442
|
-
|
438
|
+
location=location,
|
439
|
+
target_platform=platform,
|
440
|
+
pool_config_file=pool_config_file,
|
441
|
+
apps=apps,
|
442
|
+
num_gpus=input_gpus(non_interactive=non_interactive)
|
443
443
|
)
|
444
444
|
|
445
445
|
if "warning" in result:
|
@@ -1,4 +1,5 @@
|
|
1
1
|
import os
|
2
|
+
import yaml
|
2
3
|
import time
|
3
4
|
from collections import defaultdict
|
4
5
|
import math
|
@@ -71,7 +72,8 @@ from kalavai_client.env import (
|
|
71
72
|
HELM_APPS_VALUES,
|
72
73
|
POOL_CONFIG_DEFAULT_VALUES,
|
73
74
|
POOL_CONFIG_TEMPLATE,
|
74
|
-
FORBIDEDEN_IPS
|
75
|
+
FORBIDEDEN_IPS,
|
76
|
+
DEFAULT_POOL_CONFIG_TEMPLATE
|
75
77
|
)
|
76
78
|
|
77
79
|
class Job(BaseModel):
|
@@ -505,7 +507,7 @@ def fetch_gpus(available=False):
|
|
505
507
|
status = gpu["ready"] if "ready" in gpu else True
|
506
508
|
if available and not status:
|
507
509
|
continue
|
508
|
-
row_gpus.append( (f"{gpu['model']} ({gpu['memory']}
|
510
|
+
row_gpus.append( (f"{gpu['model']} ({gpu['memory']} vRAM)", str(status)))
|
509
511
|
if len(row_gpus) > 0:
|
510
512
|
models, statuses = zip(*row_gpus)
|
511
513
|
#rows.append([node, "\n".join(statuses), "\n".join(models), str(gpus["available"]), str(gpus["capacity"])])
|
@@ -752,47 +754,56 @@ def join_pool(
|
|
752
754
|
return cluster_name
|
753
755
|
|
754
756
|
def create_pool(
|
755
|
-
cluster_name: str,
|
756
|
-
ip_address: str,
|
757
|
+
cluster_name: str=None,
|
758
|
+
ip_address: str=None,
|
759
|
+
location: str=None,
|
760
|
+
target_platform: str="amd64",
|
761
|
+
pool_config_file: str=None,
|
757
762
|
description: str="",
|
758
763
|
token_mode: TokenType=TokenType.USER,
|
759
|
-
|
760
|
-
pool_config_values: str=None,
|
761
|
-
num_gpus: int=0,
|
764
|
+
num_gpus: int=-1,
|
762
765
|
node_name: str=None,
|
763
|
-
|
764
|
-
target_platform: str="amd64"
|
766
|
+
apps: list=[]
|
765
767
|
):
|
766
768
|
|
767
769
|
if not check_seed_compatibility():
|
768
770
|
return {"error": "Requirements failed"}
|
769
771
|
|
770
|
-
if
|
771
|
-
|
772
|
-
|
773
|
-
if
|
774
|
-
|
772
|
+
if pool_config_file is None:
|
773
|
+
pool_config_file = DEFAULT_POOL_CONFIG_TEMPLATE
|
774
|
+
|
775
|
+
if node_name is None:
|
776
|
+
node_name = socket.gethostname()
|
775
777
|
|
776
|
-
node_name = socket.gethostname()
|
777
778
|
user_id = load_user_id()
|
778
779
|
|
779
780
|
node_labels = {
|
780
781
|
STORAGE_CLASS_LABEL: is_storage_compatible(),
|
781
782
|
NODE_ROLE_LABEL: "server"
|
782
783
|
}
|
783
|
-
|
784
|
-
# if location is not None:
|
785
|
-
# try:
|
786
|
-
# vpn = get_vpn_details(
|
787
|
-
# location=location,
|
788
|
-
# user_cookie=USER_COOKIE)
|
789
|
-
# node_labels[USER_NODE_LABEL] = user["username"]
|
790
|
-
# except Exception as e:
|
791
|
-
# return {"error": f"[red]Error when joining network: {str(e)}"}
|
792
784
|
|
793
|
-
if num_gpus
|
785
|
+
if num_gpus < 0:
|
794
786
|
num_gpus = get_max_gpus()
|
795
|
-
|
787
|
+
|
788
|
+
# load values from pool config
|
789
|
+
with open(pool_config_file, "r") as f:
|
790
|
+
config_values = yaml.safe_load(f)
|
791
|
+
# use default values if not provided
|
792
|
+
try:
|
793
|
+
cluster_name = config_values["server"]["name"] if cluster_name is None else cluster_name
|
794
|
+
ip_address = config_values["server"]["ip_address"] if ip_address is None else ip_address
|
795
|
+
location = config_values["server"]["location"] if location is None else location
|
796
|
+
target_platform = config_values["server"]["platform"] if target_platform is None else target_platform
|
797
|
+
app_values = config_values["core"]
|
798
|
+
post_config_values = config_values["pool"]
|
799
|
+
deploy_apps = {
|
800
|
+
f"deploy_{app}": True for app in config_values["core"]["deploy"]
|
801
|
+
}
|
802
|
+
for app in apps:
|
803
|
+
deploy_apps[f"deploy_{app}"] = True
|
804
|
+
except Exception as e:
|
805
|
+
return {"error": f"Error when loading pool config. Missing format? {str(e)}"}
|
806
|
+
|
796
807
|
# Generate docker compose recipe
|
797
808
|
generate_compose_config(
|
798
809
|
target_platform=target_platform,
|
@@ -818,13 +829,12 @@ def create_pool(
|
|
818
829
|
time.sleep(10)
|
819
830
|
|
820
831
|
# populate local cred files
|
821
|
-
|
822
832
|
auth_key = user_id if user_id is not None else str(uuid.uuid4())
|
823
833
|
write_auth_key = str(uuid.uuid4())
|
824
834
|
readonly_auth_key = str(uuid.uuid4())
|
825
835
|
watcher_service = f"{ip_address}:{DEFAULT_WATCHER_PORT}"
|
826
836
|
values = {
|
827
|
-
CLUSTER_NAME_KEY: cluster_name,
|
837
|
+
#CLUSTER_NAME_KEY: cluster_name,
|
828
838
|
CLUSTER_IP_KEY: ip_address,
|
829
839
|
USER_ID_KEY: user_id if user_id is not None else "",
|
830
840
|
AUTH_KEY: auth_key,
|
@@ -851,8 +861,8 @@ def create_pool(
|
|
851
861
|
# Generate helmfile recipe
|
852
862
|
helm_yaml = load_template(
|
853
863
|
template_path=HELM_APPS_FILE,
|
854
|
-
values=values,
|
855
|
-
default_values_path=app_values,
|
864
|
+
values={**values, **deploy_apps, **app_values},
|
865
|
+
#default_values_path=app_values,
|
856
866
|
force_defaults=True)
|
857
867
|
with open(USER_HELM_APPS_FILE, "w") as f:
|
858
868
|
f.write(helm_yaml)
|
@@ -864,11 +874,6 @@ def create_pool(
|
|
864
874
|
)
|
865
875
|
except Exception as e:
|
866
876
|
return {"error": f"Error when updating dependencies: {str(e)}"}
|
867
|
-
|
868
|
-
if location is not None:
|
869
|
-
# TODO: register with kalavai if it's a public cluster
|
870
|
-
pass
|
871
|
-
#pool__publish()
|
872
877
|
|
873
878
|
# wait until the server is ready to create objects
|
874
879
|
while True:
|
@@ -876,7 +881,7 @@ def create_pool(
|
|
876
881
|
if is_watcher_alive(server_creds=USER_LOCAL_SERVER_FILE, user_cookie=USER_COOKIE):
|
877
882
|
break
|
878
883
|
|
879
|
-
result = pool_init(
|
884
|
+
result = pool_init(config_values=post_config_values)
|
880
885
|
if "error" in result or ("failed" in result and len(result['failed']) > 0):
|
881
886
|
return {"error": f"Error when initialising pool: {result}"}
|
882
887
|
# init default namespace
|
@@ -953,16 +958,15 @@ def get_pool_token(mode: TokenType):
|
|
953
958
|
except Exception as e:
|
954
959
|
return {"error": f"Error when generating token: {str(e)}"}
|
955
960
|
|
956
|
-
def pool_init(
|
961
|
+
def pool_init(config_values=None):
|
957
962
|
"""Deploy configured objects to initialise pool"""
|
958
|
-
if
|
963
|
+
if config_values is None:
|
959
964
|
return
|
960
965
|
|
961
966
|
# load template config and populate with values
|
962
967
|
sidecar_template_yaml = load_template(
|
963
968
|
template_path=POOL_CONFIG_TEMPLATE,
|
964
|
-
values=
|
965
|
-
default_values_path=pool_config_values_path)
|
969
|
+
values=config_values)
|
966
970
|
|
967
971
|
try:
|
968
972
|
result = request_to_server(
|
@@ -51,6 +51,7 @@ HELM_APPS_FILE = resource_path("kalavai_client/assets/apps.yaml")
|
|
51
51
|
HELM_APPS_VALUES = resource_path("kalavai_client/assets/apps_values.yaml")
|
52
52
|
DOCKER_COMPOSE_TEMPLATE = resource_path("kalavai_client/assets/docker-compose-template.yaml")
|
53
53
|
DOCKER_COMPOSE_GUI = resource_path("kalavai_client/assets/docker-compose-gui.yaml")
|
54
|
+
DEFAULT_POOL_CONFIG_TEMPLATE = resource_path("kalavai_client/assets/default_pool_config.yaml")
|
54
55
|
USER_WORKSPACE_TEMPLATE = resource_path("kalavai_client/assets/user_workspace.yaml")
|
55
56
|
DEFAULT_USER_WORKSPACE_VALUES = resource_path("kalavai_client/assets/user_workspace_values.yaml")
|
56
57
|
POOL_CONFIG_TEMPLATE = resource_path("kalavai_client/assets/pool_config_template.yaml")
|
@@ -1,106 +0,0 @@
|
|
1
|
-
### APS ###
|
2
|
-
- name: deploy_lago
|
3
|
-
default: "False"
|
4
|
-
description: "Deploy Lago payment system"
|
5
|
-
|
6
|
-
- name: deploy_opencost
|
7
|
-
default: "False"
|
8
|
-
description: "Deploy Opencost cost monitoring system"
|
9
|
-
|
10
|
-
- name: deploy_prometheus
|
11
|
-
default: "False"
|
12
|
-
description: "Deploy Prometheus system monitoring system"
|
13
|
-
|
14
|
-
- name: deploy_langfuse
|
15
|
-
default: "False"
|
16
|
-
description: "Deploy Langfuse LLM tracing system"
|
17
|
-
|
18
|
-
- name: deploy_kuberay
|
19
|
-
default: "False"
|
20
|
-
description: "Deploy Langfuse LLM tracing system"
|
21
|
-
|
22
|
-
######
|
23
|
-
|
24
|
-
### VARIABLES ###
|
25
|
-
|
26
|
-
- name: kalavai_api_endpoint
|
27
|
-
default: https://platform.kalavai.net/_/api
|
28
|
-
description: ""
|
29
|
-
|
30
|
-
- name: prometheus_endpoint
|
31
|
-
default: http://prometheus-server.prometheus-system.svc.cluster.local:80
|
32
|
-
description: ""
|
33
|
-
|
34
|
-
- name: opencost_endpoint
|
35
|
-
default: http://opencost.opencost.svc.cluster.local:9003
|
36
|
-
description: ""
|
37
|
-
|
38
|
-
- name: longhorn_manager_endpoint
|
39
|
-
default: http://longhorn-backend.kalavai.svc.cluster.local:9500
|
40
|
-
description: ""
|
41
|
-
|
42
|
-
- name: helios_harvest_interval
|
43
|
-
default: 120
|
44
|
-
description: "Interval (minutes) at which to report resource usage in public pools"
|
45
|
-
|
46
|
-
- name: watcher_is_shared_pool
|
47
|
-
default: "True"
|
48
|
-
description: "All users see each other's work"
|
49
|
-
|
50
|
-
## Longhorn
|
51
|
-
- name: longhorn_ui_port
|
52
|
-
default: 30000
|
53
|
-
description: ""
|
54
|
-
|
55
|
-
- name: longhorn_manager_port
|
56
|
-
default: "" #31011
|
57
|
-
description: ""
|
58
|
-
|
59
|
-
- name: longhorn_minimal_available_percentage
|
60
|
-
default: 5
|
61
|
-
description: ""
|
62
|
-
|
63
|
-
- name: longhorn_label_selector_key
|
64
|
-
default: "kalavai.storage.enabled"
|
65
|
-
description: ""
|
66
|
-
|
67
|
-
- name: longhorn_label_selector
|
68
|
-
default: "True"
|
69
|
-
description: ""
|
70
|
-
|
71
|
-
- name: longhorn_replicas
|
72
|
-
default: 2
|
73
|
-
description: ""
|
74
|
-
|
75
|
-
## MinIO
|
76
|
-
- name: minio_replicas
|
77
|
-
default: 2
|
78
|
-
description: ""
|
79
|
-
|
80
|
-
- name: minio_resources_memory
|
81
|
-
default: "1Gi"
|
82
|
-
description: ""
|
83
|
-
|
84
|
-
- name: minio_persistence_storageClass
|
85
|
-
default: "longhorn"
|
86
|
-
description: ""
|
87
|
-
|
88
|
-
- name: minio_persistence_size
|
89
|
-
default: "10Gi"
|
90
|
-
description: ""
|
91
|
-
|
92
|
-
- name: minio_service_port
|
93
|
-
default: 32000
|
94
|
-
description: ""
|
95
|
-
|
96
|
-
- name: minio_console_port
|
97
|
-
default: 32001
|
98
|
-
description: ""
|
99
|
-
|
100
|
-
- name: minio_rootUser
|
101
|
-
default: "admin"
|
102
|
-
description: ""
|
103
|
-
|
104
|
-
- name: minio_rootPassword
|
105
|
-
default: "password"
|
106
|
-
description: ""
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
{kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/docker-compose-template.yaml
RENAMED
File without changes
|
File without changes
|
{kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/pool_config_template.yaml
RENAMED
File without changes
|
File without changes
|
{kalavai_client-0.7.1 → kalavai_client-0.7.3}/kalavai_client/assets/user_workspace_values.yaml
RENAMED
File without changes
|
File without changes
|
File without changes
|
File without changes
|