skypilot-nightly 1.0.0.dev20250723__py3-none-any.whl → 1.0.0.dev20250725__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of skypilot-nightly might be problematic. Click here for more details.
- sky/__init__.py +2 -2
- sky/backends/backend_utils.py +27 -1
- sky/client/cli/command.py +61 -21
- sky/client/sdk.pyi +296 -0
- sky/clouds/utils/oci_utils.py +16 -40
- sky/dashboard/out/404.html +1 -1
- sky/dashboard/out/clusters/[cluster]/[job].html +1 -1
- sky/dashboard/out/clusters/[cluster].html +1 -1
- sky/dashboard/out/clusters.html +1 -1
- sky/dashboard/out/config.html +1 -1
- sky/dashboard/out/index.html +1 -1
- sky/dashboard/out/infra/[context].html +1 -1
- sky/dashboard/out/infra.html +1 -1
- sky/dashboard/out/jobs/[job].html +1 -1
- sky/dashboard/out/jobs.html +1 -1
- sky/dashboard/out/users.html +1 -1
- sky/dashboard/out/volumes.html +1 -1
- sky/dashboard/out/workspace/new.html +1 -1
- sky/dashboard/out/workspaces/[name].html +1 -1
- sky/dashboard/out/workspaces.html +1 -1
- sky/exceptions.py +8 -0
- sky/global_user_state.py +12 -23
- sky/jobs/state.py +12 -24
- sky/logs/__init__.py +4 -0
- sky/logs/agent.py +14 -0
- sky/logs/aws.py +276 -0
- sky/server/common.py +14 -1
- sky/server/requests/payloads.py +20 -4
- sky/server/rest.py +6 -0
- sky/server/server.py +5 -1
- sky/templates/aws-ray.yml.j2 +7 -1
- sky/templates/azure-ray.yml.j2 +1 -1
- sky/templates/do-ray.yml.j2 +1 -1
- sky/templates/lambda-ray.yml.j2 +1 -1
- sky/templates/nebius-ray.yml.j2 +1 -1
- sky/templates/paperspace-ray.yml.j2 +1 -1
- sky/templates/runpod-ray.yml.j2 +1 -1
- sky/utils/config_utils.py +6 -4
- sky/utils/db/migration_utils.py +60 -19
- sky/utils/rich_utils.py +2 -3
- sky/utils/schemas.py +67 -22
- {skypilot_nightly-1.0.0.dev20250723.dist-info → skypilot_nightly-1.0.0.dev20250725.dist-info}/METADATA +1 -1
- {skypilot_nightly-1.0.0.dev20250723.dist-info → skypilot_nightly-1.0.0.dev20250725.dist-info}/RECORD +49 -47
- /sky/dashboard/out/_next/static/{mym3Ciwp-zqU7ZpOLGnrW → SiA7c33x_DqO42M373Okd}/_buildManifest.js +0 -0
- /sky/dashboard/out/_next/static/{mym3Ciwp-zqU7ZpOLGnrW → SiA7c33x_DqO42M373Okd}/_ssgManifest.js +0 -0
- {skypilot_nightly-1.0.0.dev20250723.dist-info → skypilot_nightly-1.0.0.dev20250725.dist-info}/WHEEL +0 -0
- {skypilot_nightly-1.0.0.dev20250723.dist-info → skypilot_nightly-1.0.0.dev20250725.dist-info}/entry_points.txt +0 -0
- {skypilot_nightly-1.0.0.dev20250723.dist-info → skypilot_nightly-1.0.0.dev20250725.dist-info}/licenses/LICENSE +0 -0
- {skypilot_nightly-1.0.0.dev20250723.dist-info → skypilot_nightly-1.0.0.dev20250725.dist-info}/top_level.txt +0 -0
sky/__init__.py
CHANGED
|
@@ -5,7 +5,7 @@ from typing import Optional
|
|
|
5
5
|
import urllib.request
|
|
6
6
|
|
|
7
7
|
# Replaced with the current commit when building the wheels.
|
|
8
|
-
_SKYPILOT_COMMIT_SHA = '
|
|
8
|
+
_SKYPILOT_COMMIT_SHA = 'c8383b047969b0b18ca50a7ad02ea9d8cf77d236'
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
def _get_git_commit():
|
|
@@ -35,7 +35,7 @@ def _get_git_commit():
|
|
|
35
35
|
|
|
36
36
|
|
|
37
37
|
__commit__ = _get_git_commit()
|
|
38
|
-
__version__ = '1.0.0.
|
|
38
|
+
__version__ = '1.0.0.dev20250725'
|
|
39
39
|
__root_dir__ = os.path.dirname(os.path.abspath(__file__))
|
|
40
40
|
|
|
41
41
|
|
sky/backends/backend_utils.py
CHANGED
|
@@ -35,6 +35,7 @@ from sky.adaptors import common as adaptors_common
|
|
|
35
35
|
from sky.jobs import utils as managed_job_utils
|
|
36
36
|
from sky.provision import instance_setup
|
|
37
37
|
from sky.provision.kubernetes import utils as kubernetes_utils
|
|
38
|
+
from sky.server.requests import requests as requests_lib
|
|
38
39
|
from sky.skylet import constants
|
|
39
40
|
from sky.usage import usage_lib
|
|
40
41
|
from sky.utils import cluster_utils
|
|
@@ -798,6 +799,12 @@ def write_cluster_config(
|
|
|
798
799
|
'volume_name_on_cloud': vol.volume_config.name_on_cloud,
|
|
799
800
|
})
|
|
800
801
|
|
|
802
|
+
runcmd = skypilot_config.get_effective_region_config(
|
|
803
|
+
cloud=str(to_provision.cloud).lower(),
|
|
804
|
+
region=to_provision.region,
|
|
805
|
+
keys=('post_provision_runcmd',),
|
|
806
|
+
default_value=None)
|
|
807
|
+
|
|
801
808
|
# Use a tmp file path to avoid incomplete YAML file being re-used in the
|
|
802
809
|
# future.
|
|
803
810
|
tmp_yaml_path = yaml_path + '.tmp'
|
|
@@ -833,7 +840,7 @@ def write_cluster_config(
|
|
|
833
840
|
# User-supplied remote_identity
|
|
834
841
|
'remote_identity': remote_identity,
|
|
835
842
|
# The reservation pools that specified by the user. This is
|
|
836
|
-
# currently only used by GCP.
|
|
843
|
+
# currently only used by AWS and GCP.
|
|
837
844
|
'specific_reservations': specific_reservations,
|
|
838
845
|
|
|
839
846
|
# Conda setup
|
|
@@ -896,6 +903,10 @@ def write_cluster_config(
|
|
|
896
903
|
|
|
897
904
|
# Volume mounts
|
|
898
905
|
'volume_mounts': volume_mount_vars,
|
|
906
|
+
|
|
907
|
+
# runcmd to append to the cloud-init cloud config passed to the
|
|
908
|
+
# machine's UserData. This is currently only used by AWS.
|
|
909
|
+
'runcmd': runcmd,
|
|
899
910
|
}),
|
|
900
911
|
output_path=tmp_yaml_path)
|
|
901
912
|
config_dict['cluster_name'] = cluster_name
|
|
@@ -2831,6 +2842,21 @@ def get_clusters(
|
|
|
2831
2842
|
force_refresh_statuses = None
|
|
2832
2843
|
|
|
2833
2844
|
def _refresh_cluster(cluster_name):
|
|
2845
|
+
# TODO(syang): we should try not to leak
|
|
2846
|
+
# request info in backend_utils.py.
|
|
2847
|
+
# Refactor this to use some other info to
|
|
2848
|
+
# determine if a launch is in progress.
|
|
2849
|
+
request = requests_lib.get_request_tasks(
|
|
2850
|
+
status=[requests_lib.RequestStatus.RUNNING],
|
|
2851
|
+
cluster_names=[cluster_name],
|
|
2852
|
+
include_request_names=['sky.launch'])
|
|
2853
|
+
if len(request) > 0:
|
|
2854
|
+
# There is an active launch request on the cluster,
|
|
2855
|
+
# so we don't want to update the cluster status until
|
|
2856
|
+
# the request is completed.
|
|
2857
|
+
logger.debug(f'skipping refresh for cluster {cluster_name} '
|
|
2858
|
+
'as there is an active launch request')
|
|
2859
|
+
return global_user_state.get_cluster_from_name(cluster_name)
|
|
2834
2860
|
try:
|
|
2835
2861
|
record = refresh_cluster_record(
|
|
2836
2862
|
cluster_name,
|
sky/client/cli/command.py
CHANGED
|
@@ -24,6 +24,7 @@ listed in "sky --help". Take care to put logically connected commands close to
|
|
|
24
24
|
each other.
|
|
25
25
|
"""
|
|
26
26
|
import collections
|
|
27
|
+
import concurrent.futures
|
|
27
28
|
import fnmatch
|
|
28
29
|
import os
|
|
29
30
|
import pathlib
|
|
@@ -358,6 +359,9 @@ def _install_shell_completion(ctx: click.Context, param: click.Parameter,
|
|
|
358
359
|
bashrc_diff = ('\n# For SkyPilot shell completion'
|
|
359
360
|
'\n. ~/.sky/.sky-complete.bash')
|
|
360
361
|
|
|
362
|
+
cmd: Optional[str] = None
|
|
363
|
+
reload_cmd: Optional[str] = None
|
|
364
|
+
|
|
361
365
|
if value == 'bash':
|
|
362
366
|
install_cmd = f'_SKY_COMPLETE=bash_source sky > \
|
|
363
367
|
~/.sky/.sky-complete.bash && \
|
|
@@ -389,6 +393,7 @@ def _install_shell_completion(ctx: click.Context, param: click.Parameter,
|
|
|
389
393
|
click.secho(f'Unsupported shell: {value}', fg='red')
|
|
390
394
|
ctx.exit()
|
|
391
395
|
|
|
396
|
+
assert cmd is not None # This should never be None due to ctx.exit() above
|
|
392
397
|
try:
|
|
393
398
|
subprocess.run(cmd,
|
|
394
399
|
shell=True,
|
|
@@ -420,6 +425,9 @@ def _uninstall_shell_completion(ctx: click.Context, param: click.Parameter,
|
|
|
420
425
|
else:
|
|
421
426
|
value = os.path.basename(os.environ['SHELL'])
|
|
422
427
|
|
|
428
|
+
cmd: Optional[str] = None
|
|
429
|
+
reload_cmd: Optional[str] = None
|
|
430
|
+
|
|
423
431
|
if value == 'bash':
|
|
424
432
|
cmd = 'sed -i"" -e "/# For SkyPilot shell completion/d" ~/.bashrc && \
|
|
425
433
|
sed -i"" -e "/sky-complete.bash/d" ~/.bashrc && \
|
|
@@ -444,6 +452,7 @@ def _uninstall_shell_completion(ctx: click.Context, param: click.Parameter,
|
|
|
444
452
|
click.secho(f'Unsupported shell: {value}', fg='red')
|
|
445
453
|
ctx.exit()
|
|
446
454
|
|
|
455
|
+
assert cmd is not None # This should never be None due to ctx.exit() above
|
|
447
456
|
try:
|
|
448
457
|
subprocess.run(cmd, shell=True, check=True)
|
|
449
458
|
click.secho(f'Shell completion uninstalled for {value}', fg='green')
|
|
@@ -1745,19 +1754,16 @@ def status(verbose: bool, refresh: bool, ip: bool, endpoints: bool,
|
|
|
1745
1754
|
# Do not show job queue if user specifies clusters, and if user
|
|
1746
1755
|
# specifies --ip or --endpoint(s).
|
|
1747
1756
|
show_managed_jobs = show_managed_jobs and not any([clusters, ip, endpoints])
|
|
1748
|
-
if show_managed_jobs:
|
|
1749
|
-
managed_jobs_queue_request_id = managed_jobs.queue(refresh=False,
|
|
1750
|
-
skip_finished=True,
|
|
1751
|
-
all_users=all_users)
|
|
1752
1757
|
show_endpoints = endpoints or endpoint is not None
|
|
1753
1758
|
show_single_endpoint = endpoint is not None
|
|
1754
1759
|
show_services = show_services and not any([clusters, ip, endpoints])
|
|
1755
|
-
if show_services:
|
|
1756
|
-
# Run the sky serve service query in parallel to speed up the
|
|
1757
|
-
# status query.
|
|
1758
|
-
service_status_request_id = serve_lib.status(service_names=None)
|
|
1759
1760
|
|
|
1760
|
-
|
|
1761
|
+
query_clusters: Optional[List[str]] = None if not clusters else clusters
|
|
1762
|
+
refresh_mode = common.StatusRefreshMode.NONE
|
|
1763
|
+
if refresh:
|
|
1764
|
+
refresh_mode = common.StatusRefreshMode.FORCE
|
|
1765
|
+
|
|
1766
|
+
# Phase 1: Validate arguments for IP/endpoint queries
|
|
1761
1767
|
if ip or show_endpoints:
|
|
1762
1768
|
if refresh:
|
|
1763
1769
|
raise click.UsageError(
|
|
@@ -1791,9 +1797,19 @@ def status(verbose: bool, refresh: bool, ip: bool, endpoints: bool,
|
|
|
1791
1797
|
flag='ip' if ip else
|
|
1792
1798
|
('endpoint port'
|
|
1793
1799
|
if show_single_endpoint else 'endpoints')))
|
|
1794
|
-
|
|
1800
|
+
|
|
1801
|
+
# Phase 2: Parallel submission of all API requests
|
|
1802
|
+
def submit_managed_jobs():
|
|
1803
|
+
return managed_jobs.queue(refresh=False,
|
|
1804
|
+
skip_finished=True,
|
|
1805
|
+
all_users=all_users)
|
|
1806
|
+
|
|
1807
|
+
def submit_services() -> Optional[str]:
|
|
1808
|
+
return serve_lib.status(service_names=None)
|
|
1809
|
+
|
|
1810
|
+
def submit_workspace() -> Optional[str]:
|
|
1795
1811
|
try:
|
|
1796
|
-
|
|
1812
|
+
return sdk.workspaces()
|
|
1797
1813
|
except RuntimeError:
|
|
1798
1814
|
# Backward compatibility for API server before #5660.
|
|
1799
1815
|
# TODO(zhwu): remove this after 0.10.0.
|
|
@@ -1802,12 +1818,35 @@ def status(verbose: bool, refresh: bool, ip: bool, endpoints: bool,
|
|
|
1802
1818
|
'workspaces. Update with: sky api stop; '
|
|
1803
1819
|
'sky api start'
|
|
1804
1820
|
f'{colorama.Style.RESET_ALL}')
|
|
1805
|
-
|
|
1821
|
+
return None
|
|
1806
1822
|
|
|
1807
|
-
|
|
1808
|
-
|
|
1809
|
-
|
|
1810
|
-
|
|
1823
|
+
managed_jobs_queue_request_id = None
|
|
1824
|
+
service_status_request_id = None
|
|
1825
|
+
workspace_request_id = None
|
|
1826
|
+
|
|
1827
|
+
# Submit all requests in parallel
|
|
1828
|
+
with concurrent.futures.ThreadPoolExecutor(max_workers=1) as executor:
|
|
1829
|
+
if show_managed_jobs:
|
|
1830
|
+
managed_jobs_request_future = executor.submit(submit_managed_jobs)
|
|
1831
|
+
if show_services:
|
|
1832
|
+
services_request_future = executor.submit(submit_services)
|
|
1833
|
+
if not (ip or show_endpoints):
|
|
1834
|
+
workspace_request_future = executor.submit(submit_workspace)
|
|
1835
|
+
|
|
1836
|
+
# Get the request IDs
|
|
1837
|
+
if show_managed_jobs:
|
|
1838
|
+
managed_jobs_queue_request_id = managed_jobs_request_future.result()
|
|
1839
|
+
if show_services:
|
|
1840
|
+
service_status_request_id = services_request_future.result()
|
|
1841
|
+
if not (ip or show_endpoints):
|
|
1842
|
+
workspace_request_id = workspace_request_future.result()
|
|
1843
|
+
|
|
1844
|
+
managed_jobs_queue_request_id = '' if not managed_jobs_queue_request_id \
|
|
1845
|
+
else managed_jobs_queue_request_id
|
|
1846
|
+
service_status_request_id = '' if not service_status_request_id \
|
|
1847
|
+
else service_status_request_id
|
|
1848
|
+
|
|
1849
|
+
# Phase 3: Get cluster records and handle special cases
|
|
1811
1850
|
cluster_records = _get_cluster_records_and_set_ssh_config(
|
|
1812
1851
|
query_clusters, refresh_mode, all_users)
|
|
1813
1852
|
|
|
@@ -2143,7 +2182,9 @@ def logs(
|
|
|
2143
2182
|
if sync_down:
|
|
2144
2183
|
with rich_utils.client_status(
|
|
2145
2184
|
ux_utils.spinner_message('Downloading logs')):
|
|
2146
|
-
log_local_path_dict = sdk.download_logs(
|
|
2185
|
+
log_local_path_dict = sdk.download_logs(
|
|
2186
|
+
cluster,
|
|
2187
|
+
list(job_ids) if job_ids else None)
|
|
2147
2188
|
style = colorama.Style
|
|
2148
2189
|
fore = colorama.Fore
|
|
2149
2190
|
for job, log_local_path in log_local_path_dict.items():
|
|
@@ -2195,8 +2236,7 @@ def logs(
|
|
|
2195
2236
|
f'{colorama.Style.RESET_ALL}')
|
|
2196
2237
|
|
|
2197
2238
|
# Stream logs from the server.
|
|
2198
|
-
|
|
2199
|
-
sys.exit(returncode)
|
|
2239
|
+
sys.exit(sdk.tail_logs(cluster, job_id, follow, tail=tail))
|
|
2200
2240
|
|
|
2201
2241
|
|
|
2202
2242
|
@cli.command()
|
|
@@ -3244,7 +3284,7 @@ def show_gpus(
|
|
|
3244
3284
|
infra: Optional[str],
|
|
3245
3285
|
cloud: Optional[str],
|
|
3246
3286
|
region: Optional[str],
|
|
3247
|
-
all_regions:
|
|
3287
|
+
all_regions: bool):
|
|
3248
3288
|
"""Show supported GPU/TPU/accelerators and their prices.
|
|
3249
3289
|
|
|
3250
3290
|
The names and counts shown can be set in the ``accelerators`` field in task
|
|
@@ -5398,7 +5438,7 @@ def api():
|
|
|
5398
5438
|
required=False,
|
|
5399
5439
|
help='Enable basic authentication in the SkyPilot API server.')
|
|
5400
5440
|
@usage_lib.entrypoint
|
|
5401
|
-
def api_start(deploy: bool, host:
|
|
5441
|
+
def api_start(deploy: bool, host: str, foreground: bool,
|
|
5402
5442
|
enable_basic_auth: bool):
|
|
5403
5443
|
"""Starts the SkyPilot API server locally."""
|
|
5404
5444
|
sdk.api_start(deploy=deploy,
|
sky/client/sdk.pyi
ADDED
|
@@ -0,0 +1,296 @@
|
|
|
1
|
+
"""Stub for sky.client.sdk, generated by stubgen."""
|
|
2
|
+
import io
|
|
3
|
+
from typing import Any, Dict, List, Optional, Tuple, Union
|
|
4
|
+
|
|
5
|
+
from _typeshed import Incomplete
|
|
6
|
+
import requests
|
|
7
|
+
|
|
8
|
+
import sky
|
|
9
|
+
from sky import admin_policy as admin_policy
|
|
10
|
+
from sky import backends as backends
|
|
11
|
+
from sky import exceptions as exceptions
|
|
12
|
+
from sky import sky_logging as sky_logging
|
|
13
|
+
from sky import skypilot_config as skypilot_config
|
|
14
|
+
from sky.server import common as server_common
|
|
15
|
+
from sky.server import rest as rest
|
|
16
|
+
from sky.server.requests import payloads as payloads
|
|
17
|
+
from sky.skylet import constants as constants
|
|
18
|
+
from sky.usage import usage_lib as usage_lib
|
|
19
|
+
from sky.utils import admin_policy_utils as admin_policy_utils
|
|
20
|
+
from sky.utils import annotations as annotations
|
|
21
|
+
from sky.utils import cluster_utils as cluster_utils
|
|
22
|
+
from sky.utils import common as common
|
|
23
|
+
from sky.utils import common_utils as common_utils
|
|
24
|
+
from sky.utils import dag_utils as dag_utils
|
|
25
|
+
from sky.utils import env_options as env_options
|
|
26
|
+
from sky.utils import infra_utils as infra_utils
|
|
27
|
+
from sky.utils import rich_utils as rich_utils
|
|
28
|
+
from sky.utils import status_lib as status_lib
|
|
29
|
+
from sky.utils import subprocess_utils as subprocess_utils
|
|
30
|
+
from sky.utils import ux_utils as ux_utils
|
|
31
|
+
from sky.utils.kubernetes import ssh_utils as ssh_utils
|
|
32
|
+
|
|
33
|
+
logger: Incomplete
|
|
34
|
+
|
|
35
|
+
|
|
36
|
+
def reload_config() -> None:
|
|
37
|
+
...
|
|
38
|
+
|
|
39
|
+
|
|
40
|
+
def stream_response(request_id: Optional[str],
|
|
41
|
+
response: requests.Response,
|
|
42
|
+
output_stream: Optional['io.TextIOBase'] = ...,
|
|
43
|
+
resumable: bool = ...) -> Any:
|
|
44
|
+
...
|
|
45
|
+
|
|
46
|
+
|
|
47
|
+
def check(infra_list: Optional[Tuple[str, ...]],
|
|
48
|
+
verbose: bool,
|
|
49
|
+
workspace: Optional[str] = ...) -> server_common.RequestId:
|
|
50
|
+
...
|
|
51
|
+
|
|
52
|
+
|
|
53
|
+
def enabled_clouds(workspace: Optional[str] = ...,
|
|
54
|
+
expand: bool = ...) -> server_common.RequestId:
|
|
55
|
+
...
|
|
56
|
+
|
|
57
|
+
|
|
58
|
+
def list_accelerators(gpus_only: bool = ...,
|
|
59
|
+
name_filter: Optional[str] = ...,
|
|
60
|
+
region_filter: Optional[str] = ...,
|
|
61
|
+
quantity_filter: Optional[int] = ...,
|
|
62
|
+
clouds: Optional[Union[List[str], str]] = ...,
|
|
63
|
+
all_regions: bool = ...,
|
|
64
|
+
require_price: bool = ...,
|
|
65
|
+
case_sensitive: bool = ...) -> server_common.RequestId:
|
|
66
|
+
...
|
|
67
|
+
|
|
68
|
+
|
|
69
|
+
def list_accelerator_counts(
|
|
70
|
+
gpus_only: bool = ...,
|
|
71
|
+
name_filter: Optional[str] = ...,
|
|
72
|
+
region_filter: Optional[str] = ...,
|
|
73
|
+
quantity_filter: Optional[int] = ...,
|
|
74
|
+
clouds: Optional[Union[List[str],
|
|
75
|
+
str]] = ...) -> server_common.RequestId:
|
|
76
|
+
...
|
|
77
|
+
|
|
78
|
+
|
|
79
|
+
def optimize(
|
|
80
|
+
dag: sky.Dag,
|
|
81
|
+
minimize: common.OptimizeTarget = ...,
|
|
82
|
+
admin_policy_request_options: Optional[admin_policy.RequestOptions] = ...
|
|
83
|
+
) -> server_common.RequestId:
|
|
84
|
+
...
|
|
85
|
+
|
|
86
|
+
|
|
87
|
+
def workspaces() -> server_common.RequestId:
|
|
88
|
+
...
|
|
89
|
+
|
|
90
|
+
|
|
91
|
+
def validate(
|
|
92
|
+
dag: sky.Dag,
|
|
93
|
+
workdir_only: bool = ...,
|
|
94
|
+
admin_policy_request_options: Optional[admin_policy.RequestOptions] = ...
|
|
95
|
+
) -> None:
|
|
96
|
+
...
|
|
97
|
+
|
|
98
|
+
|
|
99
|
+
def dashboard(starting_page: Optional[str] = ...) -> None:
|
|
100
|
+
...
|
|
101
|
+
|
|
102
|
+
|
|
103
|
+
def launch(task: Union['sky.Task', 'sky.Dag'],
|
|
104
|
+
cluster_name: Optional[str] = ...,
|
|
105
|
+
retry_until_up: bool = ...,
|
|
106
|
+
idle_minutes_to_autostop: Optional[int] = ...,
|
|
107
|
+
dryrun: bool = ...,
|
|
108
|
+
down: bool = ...,
|
|
109
|
+
backend: Optional['backends.Backend'] = ...,
|
|
110
|
+
optimize_target: common.OptimizeTarget = ...,
|
|
111
|
+
no_setup: bool = ...,
|
|
112
|
+
clone_disk_from: Optional[str] = ...,
|
|
113
|
+
fast: bool = ...,
|
|
114
|
+
_need_confirmation: bool = ...,
|
|
115
|
+
_is_launched_by_jobs_controller: bool = ...,
|
|
116
|
+
_is_launched_by_sky_serve_controller: bool = ...,
|
|
117
|
+
_disable_controller_check: bool = ...) -> server_common.RequestId:
|
|
118
|
+
...
|
|
119
|
+
|
|
120
|
+
|
|
121
|
+
def exec(
|
|
122
|
+
task: Union['sky.Task', 'sky.Dag'],
|
|
123
|
+
cluster_name: Optional[str] = ...,
|
|
124
|
+
dryrun: bool = ...,
|
|
125
|
+
down: bool = ...,
|
|
126
|
+
backend: Optional['backends.Backend'] = ...) -> server_common.RequestId:
|
|
127
|
+
...
|
|
128
|
+
|
|
129
|
+
|
|
130
|
+
def tail_logs(cluster_name: str,
|
|
131
|
+
job_id: Optional[int],
|
|
132
|
+
follow: bool,
|
|
133
|
+
tail: int = ...,
|
|
134
|
+
output_stream: Optional['io.TextIOBase'] = ...) -> int:
|
|
135
|
+
...
|
|
136
|
+
|
|
137
|
+
|
|
138
|
+
def download_logs(cluster_name: str,
|
|
139
|
+
job_ids: Optional[List[str]]) -> Dict[str, str]:
|
|
140
|
+
...
|
|
141
|
+
|
|
142
|
+
|
|
143
|
+
def start(cluster_name: str,
|
|
144
|
+
idle_minutes_to_autostop: Optional[int] = ...,
|
|
145
|
+
retry_until_up: bool = ...,
|
|
146
|
+
down: bool = ...,
|
|
147
|
+
force: bool = ...) -> server_common.RequestId:
|
|
148
|
+
...
|
|
149
|
+
|
|
150
|
+
|
|
151
|
+
def down(cluster_name: str, purge: bool = ...) -> server_common.RequestId:
|
|
152
|
+
...
|
|
153
|
+
|
|
154
|
+
|
|
155
|
+
def stop(cluster_name: str, purge: bool = ...) -> server_common.RequestId:
|
|
156
|
+
...
|
|
157
|
+
|
|
158
|
+
|
|
159
|
+
def autostop(cluster_name: str,
|
|
160
|
+
idle_minutes: int,
|
|
161
|
+
down: bool = ...) -> server_common.RequestId:
|
|
162
|
+
...
|
|
163
|
+
|
|
164
|
+
|
|
165
|
+
def queue(cluster_name: str,
|
|
166
|
+
skip_finished: bool = ...,
|
|
167
|
+
all_users: bool = ...) -> server_common.RequestId:
|
|
168
|
+
...
|
|
169
|
+
|
|
170
|
+
|
|
171
|
+
def job_status(cluster_name: str,
|
|
172
|
+
job_ids: Optional[List[int]] = ...) -> server_common.RequestId:
|
|
173
|
+
...
|
|
174
|
+
|
|
175
|
+
|
|
176
|
+
def cancel(
|
|
177
|
+
cluster_name: str,
|
|
178
|
+
all: bool = ...,
|
|
179
|
+
all_users: bool = ...,
|
|
180
|
+
job_ids: Optional[List[int]] = ...,
|
|
181
|
+
_try_cancel_if_cluster_is_init: bool = ...) -> server_common.RequestId:
|
|
182
|
+
...
|
|
183
|
+
|
|
184
|
+
|
|
185
|
+
def status(cluster_names: Optional[List[str]] = ...,
|
|
186
|
+
refresh: common.StatusRefreshMode = ...,
|
|
187
|
+
all_users: bool = ...) -> server_common.RequestId:
|
|
188
|
+
...
|
|
189
|
+
|
|
190
|
+
|
|
191
|
+
def endpoints(cluster: str,
|
|
192
|
+
port: Optional[Union[int, str]] = ...) -> server_common.RequestId:
|
|
193
|
+
...
|
|
194
|
+
|
|
195
|
+
|
|
196
|
+
def cost_report(days: Optional[int] = ...) -> server_common.RequestId:
|
|
197
|
+
...
|
|
198
|
+
|
|
199
|
+
|
|
200
|
+
def storage_ls() -> server_common.RequestId:
|
|
201
|
+
...
|
|
202
|
+
|
|
203
|
+
|
|
204
|
+
def storage_delete(name: str) -> server_common.RequestId:
|
|
205
|
+
...
|
|
206
|
+
|
|
207
|
+
|
|
208
|
+
def local_up(gpus: bool,
|
|
209
|
+
ips: Optional[List[str]],
|
|
210
|
+
ssh_user: Optional[str],
|
|
211
|
+
ssh_key: Optional[str],
|
|
212
|
+
cleanup: bool,
|
|
213
|
+
context_name: Optional[str] = ...,
|
|
214
|
+
password: Optional[str] = ...) -> server_common.RequestId:
|
|
215
|
+
...
|
|
216
|
+
|
|
217
|
+
|
|
218
|
+
def local_down() -> server_common.RequestId:
|
|
219
|
+
...
|
|
220
|
+
|
|
221
|
+
|
|
222
|
+
def ssh_up(infra: Optional[str] = ...,
|
|
223
|
+
file: Optional[str] = ...) -> server_common.RequestId:
|
|
224
|
+
...
|
|
225
|
+
|
|
226
|
+
|
|
227
|
+
def ssh_down(infra: Optional[str] = ...) -> server_common.RequestId:
|
|
228
|
+
...
|
|
229
|
+
|
|
230
|
+
|
|
231
|
+
def realtime_kubernetes_gpu_availability(
|
|
232
|
+
context: Optional[str] = ...,
|
|
233
|
+
name_filter: Optional[str] = ...,
|
|
234
|
+
quantity_filter: Optional[int] = ...,
|
|
235
|
+
is_ssh: Optional[bool] = ...) -> server_common.RequestId:
|
|
236
|
+
...
|
|
237
|
+
|
|
238
|
+
|
|
239
|
+
def kubernetes_node_info(
|
|
240
|
+
context: Optional[str] = ...) -> server_common.RequestId:
|
|
241
|
+
...
|
|
242
|
+
|
|
243
|
+
|
|
244
|
+
def status_kubernetes() -> server_common.RequestId:
|
|
245
|
+
...
|
|
246
|
+
|
|
247
|
+
|
|
248
|
+
def get(request_id: str) -> Any:
|
|
249
|
+
...
|
|
250
|
+
|
|
251
|
+
|
|
252
|
+
def stream_and_get(request_id: Optional[str] = ...,
|
|
253
|
+
log_path: Optional[str] = ...,
|
|
254
|
+
tail: Optional[int] = ...,
|
|
255
|
+
follow: bool = ...,
|
|
256
|
+
output_stream: Optional['io.TextIOBase'] = ...) -> Any:
|
|
257
|
+
...
|
|
258
|
+
|
|
259
|
+
|
|
260
|
+
def api_cancel(request_ids: Optional[Union[str, List[str]]] = ...,
|
|
261
|
+
all_users: bool = ...,
|
|
262
|
+
silent: bool = ...) -> server_common.RequestId:
|
|
263
|
+
...
|
|
264
|
+
|
|
265
|
+
|
|
266
|
+
def api_status(request_ids: Optional[List[str]] = ...,
|
|
267
|
+
all_status: bool = ...) -> List[payloads.RequestPayload]:
|
|
268
|
+
...
|
|
269
|
+
|
|
270
|
+
|
|
271
|
+
def api_info() -> Dict[str, Any]:
|
|
272
|
+
...
|
|
273
|
+
|
|
274
|
+
|
|
275
|
+
def api_start(*,
|
|
276
|
+
deploy: bool = ...,
|
|
277
|
+
host: str = ...,
|
|
278
|
+
foreground: bool = ...,
|
|
279
|
+
metrics: bool = ...,
|
|
280
|
+
metrics_port: Optional[int] = ...,
|
|
281
|
+
enable_basic_auth: bool = ...) -> None:
|
|
282
|
+
...
|
|
283
|
+
|
|
284
|
+
|
|
285
|
+
def api_stop() -> None:
|
|
286
|
+
...
|
|
287
|
+
|
|
288
|
+
|
|
289
|
+
def api_server_logs(follow: bool = ..., tail: Optional[int] = ...) -> None:
|
|
290
|
+
...
|
|
291
|
+
|
|
292
|
+
|
|
293
|
+
def api_login(endpoint: Optional[str] = ...,
|
|
294
|
+
relogin: bool = ...,
|
|
295
|
+
service_account_token: Optional[str] = ...) -> None:
|
|
296
|
+
...
|
sky/clouds/utils/oci_utils.py
CHANGED
|
@@ -107,19 +107,13 @@ class OCIConfig:
|
|
|
107
107
|
# Allow task(cluster)-specific compartment/VCN parameters.
|
|
108
108
|
default_compartment_ocid = skypilot_config.get_effective_region_config(
|
|
109
109
|
cloud='oci',
|
|
110
|
-
region=
|
|
111
|
-
keys=(
|
|
112
|
-
'default',
|
|
113
|
-
'compartment_ocid',
|
|
114
|
-
),
|
|
110
|
+
region='default',
|
|
111
|
+
keys=('compartment_ocid',),
|
|
115
112
|
default_value=None)
|
|
116
113
|
compartment = skypilot_config.get_effective_region_config(
|
|
117
114
|
cloud='oci',
|
|
118
|
-
region=
|
|
119
|
-
keys=(
|
|
120
|
-
region,
|
|
121
|
-
'compartment_ocid',
|
|
122
|
-
),
|
|
115
|
+
region=region,
|
|
116
|
+
keys=('compartment_ocid',),
|
|
123
117
|
default_value=default_compartment_ocid)
|
|
124
118
|
return compartment
|
|
125
119
|
|
|
@@ -127,11 +121,8 @@ class OCIConfig:
|
|
|
127
121
|
def get_vcn_ocid(cls, region):
|
|
128
122
|
# Will reuse the regional VCN if specified.
|
|
129
123
|
vcn = skypilot_config.get_effective_region_config(cloud='oci',
|
|
130
|
-
region=
|
|
131
|
-
keys=(
|
|
132
|
-
region,
|
|
133
|
-
'vcn_ocid',
|
|
134
|
-
),
|
|
124
|
+
region=region,
|
|
125
|
+
keys=('vcn_ocid',),
|
|
135
126
|
default_value=None)
|
|
136
127
|
return vcn
|
|
137
128
|
|
|
@@ -139,11 +130,8 @@ class OCIConfig:
|
|
|
139
130
|
def get_vcn_subnet(cls, region):
|
|
140
131
|
# Will reuse the subnet if specified.
|
|
141
132
|
vcn = skypilot_config.get_effective_region_config(cloud='oci',
|
|
142
|
-
region=
|
|
143
|
-
keys=(
|
|
144
|
-
region,
|
|
145
|
-
'vcn_subnet',
|
|
146
|
-
),
|
|
133
|
+
region=region,
|
|
134
|
+
keys=('vcn_subnet',),
|
|
147
135
|
default_value=None)
|
|
148
136
|
return vcn
|
|
149
137
|
|
|
@@ -155,11 +143,8 @@ class OCIConfig:
|
|
|
155
143
|
# last)
|
|
156
144
|
return skypilot_config.get_effective_region_config(
|
|
157
145
|
cloud='oci',
|
|
158
|
-
region=
|
|
159
|
-
keys=(
|
|
160
|
-
'default',
|
|
161
|
-
'image_tag_gpu',
|
|
162
|
-
),
|
|
146
|
+
region='default',
|
|
147
|
+
keys=('image_tag_gpu',),
|
|
163
148
|
default_value='skypilot:gpu-ubuntu-2204')
|
|
164
149
|
|
|
165
150
|
@classmethod
|
|
@@ -169,11 +154,8 @@ class OCIConfig:
|
|
|
169
154
|
# specified, use the hardcode one at last)
|
|
170
155
|
return skypilot_config.get_effective_region_config(
|
|
171
156
|
cloud='oci',
|
|
172
|
-
region=
|
|
173
|
-
keys=(
|
|
174
|
-
'default',
|
|
175
|
-
'image_tag_general',
|
|
176
|
-
),
|
|
157
|
+
region='default',
|
|
158
|
+
keys=('image_tag_general',),
|
|
177
159
|
default_value='skypilot:cpu-ubuntu-2204')
|
|
178
160
|
|
|
179
161
|
@classmethod
|
|
@@ -190,11 +172,8 @@ class OCIConfig:
|
|
|
190
172
|
def get_profile(cls) -> str:
|
|
191
173
|
return skypilot_config.get_effective_region_config(
|
|
192
174
|
cloud='oci',
|
|
193
|
-
region=
|
|
194
|
-
keys=(
|
|
195
|
-
'default',
|
|
196
|
-
'oci_config_profile',
|
|
197
|
-
),
|
|
175
|
+
region='default',
|
|
176
|
+
keys=('oci_config_profile',),
|
|
198
177
|
default_value='DEFAULT')
|
|
199
178
|
|
|
200
179
|
@classmethod
|
|
@@ -204,11 +183,8 @@ class OCIConfig:
|
|
|
204
183
|
# specified, use the hardcode one at last)
|
|
205
184
|
return skypilot_config.get_effective_region_config(
|
|
206
185
|
cloud='oci',
|
|
207
|
-
region=
|
|
208
|
-
keys=(
|
|
209
|
-
'default',
|
|
210
|
-
'image_os_type',
|
|
211
|
-
),
|
|
186
|
+
region='default',
|
|
187
|
+
keys=('image_os_type',),
|
|
212
188
|
default_value='ubuntu')
|
|
213
189
|
|
|
214
190
|
|
sky/dashboard/out/404.html
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
<!DOCTYPE html><html><head><meta charSet="utf-8"/><meta name="viewport" content="width=device-width"/><meta name="next-head-count" content="2"/><link rel="preload" href="/dashboard/_next/static/css/b3227360726f12eb.css" as="style"/><link rel="stylesheet" href="/dashboard/_next/static/css/b3227360726f12eb.css" data-n-g=""/><noscript data-n-css=""></noscript><script defer="" nomodule="" src="/dashboard/_next/static/chunks/polyfills-78c92fac7aa8fdd8.js"></script><script src="/dashboard/_next/static/chunks/webpack-a305898dc479711e.js" defer=""></script><script src="/dashboard/_next/static/chunks/framework-efc06c2733009cd3.js" defer=""></script><script src="/dashboard/_next/static/chunks/main-c0a4f1ea606d48d2.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/_app-da491665d4289aae.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/_error-c72a1f77a3c0be1b.js" defer=""></script><script src="/dashboard/_next/static/
|
|
1
|
+
<!DOCTYPE html><html><head><meta charSet="utf-8"/><meta name="viewport" content="width=device-width"/><meta name="next-head-count" content="2"/><link rel="preload" href="/dashboard/_next/static/css/b3227360726f12eb.css" as="style"/><link rel="stylesheet" href="/dashboard/_next/static/css/b3227360726f12eb.css" data-n-g=""/><noscript data-n-css=""></noscript><script defer="" nomodule="" src="/dashboard/_next/static/chunks/polyfills-78c92fac7aa8fdd8.js"></script><script src="/dashboard/_next/static/chunks/webpack-a305898dc479711e.js" defer=""></script><script src="/dashboard/_next/static/chunks/framework-efc06c2733009cd3.js" defer=""></script><script src="/dashboard/_next/static/chunks/main-c0a4f1ea606d48d2.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/_app-da491665d4289aae.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/_error-c72a1f77a3c0be1b.js" defer=""></script><script src="/dashboard/_next/static/SiA7c33x_DqO42M373Okd/_buildManifest.js" defer=""></script><script src="/dashboard/_next/static/SiA7c33x_DqO42M373Okd/_ssgManifest.js" defer=""></script></head><body><div id="__next"></div><script id="__NEXT_DATA__" type="application/json">{"props":{"pageProps":{"statusCode":404}},"page":"/_error","query":{},"buildId":"SiA7c33x_DqO42M373Okd","assetPrefix":"/dashboard","nextExport":true,"isFallback":false,"gip":true,"scriptLoader":[]}</script></body></html>
|
|
@@ -1 +1 @@
|
|
|
1
|
-
<!DOCTYPE html><html><head><meta charSet="utf-8"/><meta name="viewport" content="width=device-width"/><meta name="next-head-count" content="2"/><link rel="preload" href="/dashboard/_next/static/css/b3227360726f12eb.css" as="style"/><link rel="stylesheet" href="/dashboard/_next/static/css/b3227360726f12eb.css" data-n-g=""/><noscript data-n-css=""></noscript><script defer="" nomodule="" src="/dashboard/_next/static/chunks/polyfills-78c92fac7aa8fdd8.js"></script><script src="/dashboard/_next/static/chunks/webpack-a305898dc479711e.js" defer=""></script><script src="/dashboard/_next/static/chunks/framework-efc06c2733009cd3.js" defer=""></script><script src="/dashboard/_next/static/chunks/main-c0a4f1ea606d48d2.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/_app-da491665d4289aae.js" defer=""></script><script src="/dashboard/_next/static/chunks/616-162f3033ffcd3d31.js" defer=""></script><script src="/dashboard/_next/static/chunks/5230-df791914b54d91d9.js" defer=""></script><script src="/dashboard/_next/static/chunks/5739-5ea3ffa10fc884f2.js" defer=""></script><script src="/dashboard/_next/static/chunks/1664-d65361e92b85e786.js" defer=""></script><script src="/dashboard/_next/static/chunks/7411-2cc31dc0fdf2a9ad.js" defer=""></script><script src="/dashboard/_next/static/chunks/6989-eab0e9c16b64fd9f.js" defer=""></script><script src="/dashboard/_next/static/chunks/3698-9fa11dafb5cad4a6.js" defer=""></script><script src="/dashboard/_next/static/chunks/6135-2abbd0352f8ee061.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/clusters/%5Bcluster%5D/%5Bjob%5D-2186770cc2de1623.js" defer=""></script><script src="/dashboard/_next/static/
|
|
1
|
+
<!DOCTYPE html><html><head><meta charSet="utf-8"/><meta name="viewport" content="width=device-width"/><meta name="next-head-count" content="2"/><link rel="preload" href="/dashboard/_next/static/css/b3227360726f12eb.css" as="style"/><link rel="stylesheet" href="/dashboard/_next/static/css/b3227360726f12eb.css" data-n-g=""/><noscript data-n-css=""></noscript><script defer="" nomodule="" src="/dashboard/_next/static/chunks/polyfills-78c92fac7aa8fdd8.js"></script><script src="/dashboard/_next/static/chunks/webpack-a305898dc479711e.js" defer=""></script><script src="/dashboard/_next/static/chunks/framework-efc06c2733009cd3.js" defer=""></script><script src="/dashboard/_next/static/chunks/main-c0a4f1ea606d48d2.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/_app-da491665d4289aae.js" defer=""></script><script src="/dashboard/_next/static/chunks/616-162f3033ffcd3d31.js" defer=""></script><script src="/dashboard/_next/static/chunks/5230-df791914b54d91d9.js" defer=""></script><script src="/dashboard/_next/static/chunks/5739-5ea3ffa10fc884f2.js" defer=""></script><script src="/dashboard/_next/static/chunks/1664-d65361e92b85e786.js" defer=""></script><script src="/dashboard/_next/static/chunks/7411-2cc31dc0fdf2a9ad.js" defer=""></script><script src="/dashboard/_next/static/chunks/6989-eab0e9c16b64fd9f.js" defer=""></script><script src="/dashboard/_next/static/chunks/3698-9fa11dafb5cad4a6.js" defer=""></script><script src="/dashboard/_next/static/chunks/6135-2abbd0352f8ee061.js" defer=""></script><script src="/dashboard/_next/static/chunks/pages/clusters/%5Bcluster%5D/%5Bjob%5D-2186770cc2de1623.js" defer=""></script><script src="/dashboard/_next/static/SiA7c33x_DqO42M373Okd/_buildManifest.js" defer=""></script><script src="/dashboard/_next/static/SiA7c33x_DqO42M373Okd/_ssgManifest.js" defer=""></script></head><body><div id="__next"></div><script id="__NEXT_DATA__" type="application/json">{"props":{"pageProps":{}},"page":"/clusters/[cluster]/[job]","query":{},"buildId":"SiA7c33x_DqO42M373Okd","assetPrefix":"/dashboard","nextExport":true,"autoExport":true,"isFallback":false,"scriptLoader":[]}</script></body></html>
|