skypilot-nightly 1.0.0.dev20250905__py3-none-any.whl → 1.0.0.dev20251210__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- sky/__init__.py +12 -2
- sky/adaptors/aws.py +27 -22
- sky/adaptors/common.py +25 -2
- sky/adaptors/coreweave.py +278 -0
- sky/adaptors/do.py +8 -2
- sky/adaptors/gcp.py +11 -0
- sky/adaptors/ibm.py +5 -2
- sky/adaptors/kubernetes.py +64 -0
- sky/adaptors/nebius.py +3 -1
- sky/adaptors/primeintellect.py +1 -0
- sky/adaptors/seeweb.py +183 -0
- sky/adaptors/shadeform.py +89 -0
- sky/adaptors/slurm.py +478 -0
- sky/admin_policy.py +20 -0
- sky/authentication.py +157 -263
- sky/backends/__init__.py +3 -2
- sky/backends/backend.py +11 -3
- sky/backends/backend_utils.py +630 -185
- sky/backends/cloud_vm_ray_backend.py +1111 -928
- sky/backends/local_docker_backend.py +9 -5
- sky/backends/task_codegen.py +971 -0
- sky/backends/wheel_utils.py +18 -0
- sky/catalog/__init__.py +8 -3
- sky/catalog/aws_catalog.py +4 -0
- sky/catalog/common.py +19 -1
- sky/catalog/data_fetchers/fetch_aws.py +102 -80
- sky/catalog/data_fetchers/fetch_gcp.py +30 -3
- sky/catalog/data_fetchers/fetch_nebius.py +9 -6
- sky/catalog/data_fetchers/fetch_runpod.py +698 -0
- sky/catalog/data_fetchers/fetch_seeweb.py +329 -0
- sky/catalog/data_fetchers/fetch_shadeform.py +142 -0
- sky/catalog/kubernetes_catalog.py +36 -32
- sky/catalog/primeintellect_catalog.py +95 -0
- sky/catalog/runpod_catalog.py +5 -1
- sky/catalog/seeweb_catalog.py +184 -0
- sky/catalog/shadeform_catalog.py +165 -0
- sky/catalog/slurm_catalog.py +243 -0
- sky/check.py +87 -46
- sky/client/cli/command.py +1004 -434
- sky/client/cli/flags.py +4 -2
- sky/{volumes/utils.py → client/cli/table_utils.py} +111 -13
- sky/client/cli/utils.py +79 -0
- sky/client/common.py +12 -2
- sky/client/sdk.py +188 -65
- sky/client/sdk_async.py +34 -33
- sky/cloud_stores.py +82 -3
- sky/clouds/__init__.py +8 -0
- sky/clouds/aws.py +337 -129
- sky/clouds/azure.py +24 -18
- sky/clouds/cloud.py +47 -13
- sky/clouds/cudo.py +16 -13
- sky/clouds/do.py +9 -7
- sky/clouds/fluidstack.py +12 -5
- sky/clouds/gcp.py +14 -7
- sky/clouds/hyperbolic.py +12 -5
- sky/clouds/ibm.py +12 -5
- sky/clouds/kubernetes.py +80 -45
- sky/clouds/lambda_cloud.py +12 -5
- sky/clouds/nebius.py +23 -9
- sky/clouds/oci.py +19 -12
- sky/clouds/paperspace.py +4 -1
- sky/clouds/primeintellect.py +317 -0
- sky/clouds/runpod.py +85 -24
- sky/clouds/scp.py +12 -8
- sky/clouds/seeweb.py +477 -0
- sky/clouds/shadeform.py +400 -0
- sky/clouds/slurm.py +578 -0
- sky/clouds/ssh.py +6 -3
- sky/clouds/utils/scp_utils.py +61 -50
- sky/clouds/vast.py +43 -27
- sky/clouds/vsphere.py +14 -16
- sky/core.py +296 -195
- sky/dashboard/out/404.html +1 -1
- sky/dashboard/out/_next/static/KYAhEFa3FTfq4JyKVgo-s/_buildManifest.js +1 -0
- sky/dashboard/out/_next/static/chunks/1141-9c810f01ff4f398a.js +11 -0
- sky/dashboard/out/_next/static/chunks/1871-7e202677c42f43fe.js +6 -0
- sky/dashboard/out/_next/static/chunks/2260-7703229c33c5ebd5.js +1 -0
- sky/dashboard/out/_next/static/chunks/2369.fc20f0c2c8ed9fe7.js +15 -0
- sky/dashboard/out/_next/static/chunks/2755.edd818326d489a1d.js +26 -0
- sky/dashboard/out/_next/static/chunks/3294.ddda8c6c6f9f24dc.js +1 -0
- sky/dashboard/out/_next/static/chunks/3785.7e245f318f9d1121.js +1 -0
- sky/dashboard/out/_next/static/chunks/{6601-06114c982db410b6.js → 3800-b589397dc09c5b4e.js} +1 -1
- sky/dashboard/out/_next/static/chunks/3850-fd5696f3bbbaddae.js +1 -0
- sky/dashboard/out/_next/static/chunks/4725.172ede95d1b21022.js +1 -0
- sky/dashboard/out/_next/static/chunks/4937.a2baa2df5572a276.js +15 -0
- sky/dashboard/out/_next/static/chunks/6212-7bd06f60ba693125.js +13 -0
- sky/dashboard/out/_next/static/chunks/6856-da20c5fd999f319c.js +1 -0
- sky/dashboard/out/_next/static/chunks/6990-09cbf02d3cd518c3.js +1 -0
- sky/dashboard/out/_next/static/chunks/7359-c8d04e06886000b3.js +30 -0
- sky/dashboard/out/_next/static/chunks/7615-019513abc55b3b47.js +1 -0
- sky/dashboard/out/_next/static/chunks/8640.5b9475a2d18c5416.js +16 -0
- sky/dashboard/out/_next/static/chunks/8969-452f9d5cbdd2dc73.js +1 -0
- sky/dashboard/out/_next/static/chunks/9025.fa408f3242e9028d.js +6 -0
- sky/dashboard/out/_next/static/chunks/9353-8369df1cf105221c.js +1 -0
- sky/dashboard/out/_next/static/chunks/9360.a536cf6b1fa42355.js +31 -0
- sky/dashboard/out/_next/static/chunks/9847.3aaca6bb33455140.js +30 -0
- sky/dashboard/out/_next/static/chunks/pages/_app-68b647e26f9d2793.js +34 -0
- sky/dashboard/out/_next/static/chunks/pages/clusters/[cluster]/[job]-33f525539665fdfd.js +16 -0
- sky/dashboard/out/_next/static/chunks/pages/clusters/[cluster]-a7565f586ef86467.js +1 -0
- sky/dashboard/out/_next/static/chunks/pages/clusters-9e5d47818b9bdadd.js +1 -0
- sky/dashboard/out/_next/static/chunks/pages/{config-dfb9bf07b13045f4.js → config-718cdc365de82689.js} +1 -1
- sky/dashboard/out/_next/static/chunks/pages/infra/{[context]-6563820e094f68ca.js → [context]-12c559ec4d81fdbd.js} +1 -1
- sky/dashboard/out/_next/static/chunks/pages/{infra-aabba60d57826e0f.js → infra-d187cd0413d72475.js} +1 -1
- sky/dashboard/out/_next/static/chunks/pages/jobs/[job]-895847b6cf200b04.js +16 -0
- sky/dashboard/out/_next/static/chunks/pages/jobs/pools/[pool]-8d0f4655400b4eb9.js +21 -0
- sky/dashboard/out/_next/static/chunks/pages/jobs-e5a98f17f8513a96.js +1 -0
- sky/dashboard/out/_next/static/chunks/pages/plugins/[...slug]-4f46050ca065d8f8.js +1 -0
- sky/dashboard/out/_next/static/chunks/pages/users-2f7646eb77785a2c.js +1 -0
- sky/dashboard/out/_next/static/chunks/pages/volumes-ef19d49c6d0e8500.js +1 -0
- sky/dashboard/out/_next/static/chunks/pages/workspaces/{[name]-af76bb06dbb3954f.js → [name]-96e0f298308da7e2.js} +1 -1
- sky/dashboard/out/_next/static/chunks/pages/{workspaces-7598c33a746cdc91.js → workspaces-cb4da3abe08ebf19.js} +1 -1
- sky/dashboard/out/_next/static/chunks/webpack-fba3de387ff6bb08.js +1 -0
- sky/dashboard/out/_next/static/css/c5a4cfd2600fc715.css +3 -0
- sky/dashboard/out/clusters/[cluster]/[job].html +1 -1
- sky/dashboard/out/clusters/[cluster].html +1 -1
- sky/dashboard/out/clusters.html +1 -1
- sky/dashboard/out/config.html +1 -1
- sky/dashboard/out/index.html +1 -1
- sky/dashboard/out/infra/[context].html +1 -1
- sky/dashboard/out/infra.html +1 -1
- sky/dashboard/out/jobs/[job].html +1 -1
- sky/dashboard/out/jobs/pools/[pool].html +1 -1
- sky/dashboard/out/jobs.html +1 -1
- sky/dashboard/out/plugins/[...slug].html +1 -0
- sky/dashboard/out/users.html +1 -1
- sky/dashboard/out/volumes.html +1 -1
- sky/dashboard/out/workspace/new.html +1 -1
- sky/dashboard/out/workspaces/[name].html +1 -1
- sky/dashboard/out/workspaces.html +1 -1
- sky/data/data_utils.py +92 -1
- sky/data/mounting_utils.py +177 -30
- sky/data/storage.py +200 -19
- sky/data/storage_utils.py +10 -45
- sky/exceptions.py +18 -7
- sky/execution.py +74 -31
- sky/global_user_state.py +605 -191
- sky/jobs/__init__.py +2 -0
- sky/jobs/client/sdk.py +101 -4
- sky/jobs/client/sdk_async.py +31 -5
- sky/jobs/constants.py +15 -8
- sky/jobs/controller.py +726 -284
- sky/jobs/file_content_utils.py +128 -0
- sky/jobs/log_gc.py +193 -0
- sky/jobs/recovery_strategy.py +250 -100
- sky/jobs/scheduler.py +271 -173
- sky/jobs/server/core.py +367 -114
- sky/jobs/server/server.py +81 -35
- sky/jobs/server/utils.py +89 -35
- sky/jobs/state.py +1498 -620
- sky/jobs/utils.py +771 -306
- sky/logs/agent.py +40 -5
- sky/logs/aws.py +9 -19
- sky/metrics/utils.py +282 -39
- sky/models.py +2 -0
- sky/optimizer.py +7 -6
- sky/provision/__init__.py +38 -1
- sky/provision/aws/config.py +34 -13
- sky/provision/aws/instance.py +5 -2
- sky/provision/azure/instance.py +5 -3
- sky/provision/common.py +22 -0
- sky/provision/cudo/instance.py +4 -3
- sky/provision/do/instance.py +4 -3
- sky/provision/docker_utils.py +112 -28
- sky/provision/fluidstack/instance.py +6 -5
- sky/provision/gcp/config.py +6 -1
- sky/provision/gcp/instance.py +4 -2
- sky/provision/hyperbolic/instance.py +4 -2
- sky/provision/instance_setup.py +66 -20
- sky/provision/kubernetes/__init__.py +2 -0
- sky/provision/kubernetes/config.py +7 -44
- sky/provision/kubernetes/constants.py +0 -1
- sky/provision/kubernetes/instance.py +609 -213
- sky/provision/kubernetes/manifests/fusermount-server-daemonset.yaml +1 -2
- sky/provision/kubernetes/network.py +12 -8
- sky/provision/kubernetes/network_utils.py +8 -25
- sky/provision/kubernetes/utils.py +422 -422
- sky/provision/kubernetes/volume.py +150 -18
- sky/provision/lambda_cloud/instance.py +16 -13
- sky/provision/nebius/instance.py +6 -2
- sky/provision/nebius/utils.py +103 -86
- sky/provision/oci/instance.py +4 -2
- sky/provision/paperspace/instance.py +4 -3
- sky/provision/primeintellect/__init__.py +10 -0
- sky/provision/primeintellect/config.py +11 -0
- sky/provision/primeintellect/instance.py +454 -0
- sky/provision/primeintellect/utils.py +398 -0
- sky/provision/provisioner.py +45 -15
- sky/provision/runpod/__init__.py +2 -0
- sky/provision/runpod/instance.py +4 -3
- sky/provision/runpod/volume.py +69 -13
- sky/provision/scp/instance.py +307 -130
- sky/provision/seeweb/__init__.py +11 -0
- sky/provision/seeweb/config.py +13 -0
- sky/provision/seeweb/instance.py +812 -0
- sky/provision/shadeform/__init__.py +11 -0
- sky/provision/shadeform/config.py +12 -0
- sky/provision/shadeform/instance.py +351 -0
- sky/provision/shadeform/shadeform_utils.py +83 -0
- sky/provision/slurm/__init__.py +12 -0
- sky/provision/slurm/config.py +13 -0
- sky/provision/slurm/instance.py +572 -0
- sky/provision/slurm/utils.py +583 -0
- sky/provision/vast/instance.py +9 -4
- sky/provision/vast/utils.py +10 -6
- sky/provision/volume.py +164 -0
- sky/provision/vsphere/common/ssl_helper.py +1 -1
- sky/provision/vsphere/common/vapiconnect.py +2 -1
- sky/provision/vsphere/common/vim_utils.py +3 -2
- sky/provision/vsphere/instance.py +8 -6
- sky/provision/vsphere/vsphere_utils.py +8 -1
- sky/resources.py +11 -3
- sky/schemas/api/responses.py +107 -6
- sky/schemas/db/global_user_state/008_skylet_ssh_tunnel_metadata.py +34 -0
- sky/schemas/db/global_user_state/009_last_activity_and_launched_at.py +89 -0
- sky/schemas/db/global_user_state/010_save_ssh_key.py +66 -0
- sky/schemas/db/global_user_state/011_is_ephemeral.py +34 -0
- sky/schemas/db/kv_cache/001_initial_schema.py +29 -0
- sky/schemas/db/serve_state/002_yaml_content.py +34 -0
- sky/schemas/db/skypilot_config/001_initial_schema.py +30 -0
- sky/schemas/db/spot_jobs/002_cluster_pool.py +3 -3
- sky/schemas/db/spot_jobs/004_job_file_contents.py +42 -0
- sky/schemas/db/spot_jobs/005_logs_gc.py +38 -0
- sky/schemas/db/spot_jobs/006_controller_pid_started_at.py +34 -0
- sky/schemas/db/spot_jobs/007_config_file_content.py +34 -0
- sky/schemas/generated/jobsv1_pb2.py +86 -0
- sky/schemas/generated/jobsv1_pb2.pyi +254 -0
- sky/schemas/generated/jobsv1_pb2_grpc.py +542 -0
- sky/schemas/generated/managed_jobsv1_pb2.py +76 -0
- sky/schemas/generated/managed_jobsv1_pb2.pyi +278 -0
- sky/schemas/generated/managed_jobsv1_pb2_grpc.py +278 -0
- sky/schemas/generated/servev1_pb2.py +58 -0
- sky/schemas/generated/servev1_pb2.pyi +115 -0
- sky/schemas/generated/servev1_pb2_grpc.py +322 -0
- sky/serve/autoscalers.py +2 -0
- sky/serve/client/impl.py +55 -21
- sky/serve/constants.py +4 -3
- sky/serve/controller.py +17 -11
- sky/serve/load_balancing_policies.py +1 -1
- sky/serve/replica_managers.py +219 -142
- sky/serve/serve_rpc_utils.py +179 -0
- sky/serve/serve_state.py +63 -54
- sky/serve/serve_utils.py +145 -109
- sky/serve/server/core.py +46 -25
- sky/serve/server/impl.py +311 -162
- sky/serve/server/server.py +21 -19
- sky/serve/service.py +84 -68
- sky/serve/service_spec.py +45 -7
- sky/server/auth/loopback.py +38 -0
- sky/server/auth/oauth2_proxy.py +12 -7
- sky/server/common.py +47 -24
- sky/server/config.py +62 -28
- sky/server/constants.py +9 -1
- sky/server/daemons.py +109 -38
- sky/server/metrics.py +76 -96
- sky/server/middleware_utils.py +166 -0
- sky/server/plugins.py +222 -0
- sky/server/requests/executor.py +384 -145
- sky/server/requests/payloads.py +83 -19
- sky/server/requests/preconditions.py +15 -13
- sky/server/requests/request_names.py +123 -0
- sky/server/requests/requests.py +511 -157
- sky/server/requests/serializers/decoders.py +48 -17
- sky/server/requests/serializers/encoders.py +102 -20
- sky/server/requests/serializers/return_value_serializers.py +60 -0
- sky/server/requests/threads.py +117 -0
- sky/server/rest.py +116 -24
- sky/server/server.py +497 -179
- sky/server/server_utils.py +30 -0
- sky/server/stream_utils.py +219 -45
- sky/server/uvicorn.py +30 -19
- sky/setup_files/MANIFEST.in +6 -1
- sky/setup_files/alembic.ini +8 -0
- sky/setup_files/dependencies.py +64 -19
- sky/setup_files/setup.py +44 -44
- sky/sky_logging.py +13 -5
- sky/skylet/attempt_skylet.py +116 -24
- sky/skylet/configs.py +3 -1
- sky/skylet/constants.py +139 -29
- sky/skylet/events.py +74 -14
- sky/skylet/executor/__init__.py +1 -0
- sky/skylet/executor/slurm.py +189 -0
- sky/skylet/job_lib.py +143 -105
- sky/skylet/log_lib.py +252 -8
- sky/skylet/log_lib.pyi +47 -7
- sky/skylet/providers/ibm/node_provider.py +12 -8
- sky/skylet/providers/ibm/vpc_provider.py +13 -12
- sky/skylet/runtime_utils.py +21 -0
- sky/skylet/services.py +524 -0
- sky/skylet/skylet.py +27 -2
- sky/skylet/subprocess_daemon.py +104 -28
- sky/skypilot_config.py +99 -79
- sky/ssh_node_pools/constants.py +12 -0
- sky/ssh_node_pools/core.py +40 -3
- sky/ssh_node_pools/deploy/__init__.py +4 -0
- sky/ssh_node_pools/deploy/deploy.py +952 -0
- sky/ssh_node_pools/deploy/tunnel_utils.py +199 -0
- sky/ssh_node_pools/deploy/utils.py +173 -0
- sky/ssh_node_pools/server.py +20 -21
- sky/{utils/kubernetes/ssh_utils.py → ssh_node_pools/utils.py} +9 -6
- sky/task.py +221 -104
- sky/templates/aws-ray.yml.j2 +1 -0
- sky/templates/azure-ray.yml.j2 +1 -0
- sky/templates/cudo-ray.yml.j2 +1 -0
- sky/templates/do-ray.yml.j2 +1 -0
- sky/templates/fluidstack-ray.yml.j2 +1 -0
- sky/templates/gcp-ray.yml.j2 +1 -0
- sky/templates/hyperbolic-ray.yml.j2 +1 -0
- sky/templates/ibm-ray.yml.j2 +2 -1
- sky/templates/jobs-controller.yaml.j2 +3 -0
- sky/templates/kubernetes-ray.yml.j2 +204 -55
- sky/templates/lambda-ray.yml.j2 +1 -0
- sky/templates/nebius-ray.yml.j2 +3 -0
- sky/templates/oci-ray.yml.j2 +1 -0
- sky/templates/paperspace-ray.yml.j2 +1 -0
- sky/templates/primeintellect-ray.yml.j2 +72 -0
- sky/templates/runpod-ray.yml.j2 +1 -0
- sky/templates/scp-ray.yml.j2 +1 -0
- sky/templates/seeweb-ray.yml.j2 +171 -0
- sky/templates/shadeform-ray.yml.j2 +73 -0
- sky/templates/slurm-ray.yml.j2 +85 -0
- sky/templates/vast-ray.yml.j2 +2 -0
- sky/templates/vsphere-ray.yml.j2 +1 -0
- sky/templates/websocket_proxy.py +188 -43
- sky/usage/usage_lib.py +16 -4
- sky/users/model.conf +1 -1
- sky/users/permission.py +84 -44
- sky/users/rbac.py +31 -3
- sky/utils/accelerator_registry.py +6 -3
- sky/utils/admin_policy_utils.py +18 -5
- sky/utils/annotations.py +128 -6
- sky/utils/asyncio_utils.py +78 -0
- sky/utils/atomic.py +1 -1
- sky/utils/auth_utils.py +153 -0
- sky/utils/cli_utils/status_utils.py +12 -7
- sky/utils/cluster_utils.py +28 -6
- sky/utils/command_runner.py +283 -30
- sky/utils/command_runner.pyi +63 -7
- sky/utils/common.py +3 -1
- sky/utils/common_utils.py +55 -7
- sky/utils/config_utils.py +1 -14
- sky/utils/context.py +127 -40
- sky/utils/context_utils.py +73 -18
- sky/utils/controller_utils.py +229 -70
- sky/utils/db/db_utils.py +95 -18
- sky/utils/db/kv_cache.py +149 -0
- sky/utils/db/migration_utils.py +24 -7
- sky/utils/env_options.py +4 -0
- sky/utils/git.py +559 -1
- sky/utils/kubernetes/create_cluster.sh +15 -30
- sky/utils/kubernetes/delete_cluster.sh +10 -7
- sky/utils/kubernetes/generate_kind_config.py +6 -66
- sky/utils/kubernetes/gpu_labeler.py +13 -3
- sky/utils/kubernetes/k8s_gpu_labeler_job.yaml +2 -1
- sky/utils/kubernetes/k8s_gpu_labeler_setup.yaml +16 -16
- sky/utils/kubernetes/kubernetes_deploy_utils.py +187 -260
- sky/utils/kubernetes/rsync_helper.sh +11 -3
- sky/utils/kubernetes/ssh-tunnel.sh +7 -376
- sky/utils/kubernetes_enums.py +7 -15
- sky/utils/lock_events.py +4 -4
- sky/utils/locks.py +128 -31
- sky/utils/log_utils.py +0 -319
- sky/utils/resource_checker.py +13 -10
- sky/utils/resources_utils.py +53 -29
- sky/utils/rich_utils.py +8 -4
- sky/utils/schemas.py +138 -52
- sky/utils/subprocess_utils.py +17 -4
- sky/utils/thread_utils.py +91 -0
- sky/utils/timeline.py +2 -1
- sky/utils/ux_utils.py +35 -1
- sky/utils/volume.py +88 -4
- sky/utils/yaml_utils.py +9 -0
- sky/volumes/client/sdk.py +48 -10
- sky/volumes/server/core.py +59 -22
- sky/volumes/server/server.py +46 -17
- sky/volumes/volume.py +54 -42
- sky/workspaces/core.py +57 -21
- sky/workspaces/server.py +13 -12
- sky_templates/README.md +3 -0
- sky_templates/__init__.py +3 -0
- sky_templates/ray/__init__.py +0 -0
- sky_templates/ray/start_cluster +183 -0
- sky_templates/ray/stop_cluster +75 -0
- {skypilot_nightly-1.0.0.dev20250905.dist-info → skypilot_nightly-1.0.0.dev20251210.dist-info}/METADATA +343 -65
- skypilot_nightly-1.0.0.dev20251210.dist-info/RECORD +629 -0
- skypilot_nightly-1.0.0.dev20251210.dist-info/top_level.txt +2 -0
- sky/client/cli/git.py +0 -549
- sky/dashboard/out/_next/static/chunks/1121-408ed10b2f9fce17.js +0 -1
- sky/dashboard/out/_next/static/chunks/1141-943efc7aff0f0c06.js +0 -1
- sky/dashboard/out/_next/static/chunks/1836-37fede578e2da5f8.js +0 -40
- sky/dashboard/out/_next/static/chunks/3015-86cabed5d4669ad0.js +0 -1
- sky/dashboard/out/_next/static/chunks/3294.c80326aec9bfed40.js +0 -6
- sky/dashboard/out/_next/static/chunks/3785.4872a2f3aa489880.js +0 -1
- sky/dashboard/out/_next/static/chunks/3850-ff4a9a69d978632b.js +0 -1
- sky/dashboard/out/_next/static/chunks/4045.b30465273dc5e468.js +0 -21
- sky/dashboard/out/_next/static/chunks/4676-9da7fdbde90b5549.js +0 -10
- sky/dashboard/out/_next/static/chunks/4725.10f7a9a5d3ea8208.js +0 -1
- sky/dashboard/out/_next/static/chunks/5339.3fda4a4010ff4e06.js +0 -51
- sky/dashboard/out/_next/static/chunks/6135-4b4d5e824b7f9d3c.js +0 -1
- sky/dashboard/out/_next/static/chunks/649.b9d7f7d10c1b8c53.js +0 -45
- sky/dashboard/out/_next/static/chunks/6856-dca7962af4814e1b.js +0 -1
- sky/dashboard/out/_next/static/chunks/6990-08b2a1cae076a943.js +0 -1
- sky/dashboard/out/_next/static/chunks/7325.b4bc99ce0892dcd5.js +0 -6
- sky/dashboard/out/_next/static/chunks/754-d0da8ab45f9509e9.js +0 -18
- sky/dashboard/out/_next/static/chunks/7669.1f5d9a402bf5cc42.js +0 -36
- sky/dashboard/out/_next/static/chunks/8969-0be3036bf86f8256.js +0 -1
- sky/dashboard/out/_next/static/chunks/9025.c12318fb6a1a9093.js +0 -6
- sky/dashboard/out/_next/static/chunks/9037-fa1737818d0a0969.js +0 -6
- sky/dashboard/out/_next/static/chunks/pages/_app-ce361c6959bc2001.js +0 -34
- sky/dashboard/out/_next/static/chunks/pages/clusters/[cluster]/[job]-1cbba24bd1bd35f8.js +0 -16
- sky/dashboard/out/_next/static/chunks/pages/clusters/[cluster]-0b4b35dc1dfe046c.js +0 -16
- sky/dashboard/out/_next/static/chunks/pages/clusters-469814d711d63b1b.js +0 -1
- sky/dashboard/out/_next/static/chunks/pages/jobs/[job]-dd64309c3fe67ed2.js +0 -11
- sky/dashboard/out/_next/static/chunks/pages/jobs/pools/[pool]-07349868f7905d37.js +0 -16
- sky/dashboard/out/_next/static/chunks/pages/jobs-1f70d9faa564804f.js +0 -1
- sky/dashboard/out/_next/static/chunks/pages/users-018bf31cda52e11b.js +0 -1
- sky/dashboard/out/_next/static/chunks/pages/volumes-739726d6b823f532.js +0 -1
- sky/dashboard/out/_next/static/chunks/webpack-4fe903277b57b523.js +0 -1
- sky/dashboard/out/_next/static/css/4614e06482d7309e.css +0 -3
- sky/dashboard/out/_next/static/mS-4qZPSkRuA1u-g2wQhg/_buildManifest.js +0 -1
- sky/templates/kubernetes-ssh-jump.yml.j2 +0 -94
- sky/utils/kubernetes/cleanup-tunnel.sh +0 -62
- sky/utils/kubernetes/deploy_remote_cluster.py +0 -1299
- sky/utils/kubernetes/ssh_jump_lifecycle_manager.py +0 -191
- skypilot_nightly-1.0.0.dev20250905.dist-info/RECORD +0 -547
- skypilot_nightly-1.0.0.dev20250905.dist-info/top_level.txt +0 -1
- /sky/dashboard/out/_next/static/{mS-4qZPSkRuA1u-g2wQhg → KYAhEFa3FTfq4JyKVgo-s}/_ssgManifest.js +0 -0
- {skypilot_nightly-1.0.0.dev20250905.dist-info → skypilot_nightly-1.0.0.dev20251210.dist-info}/WHEEL +0 -0
- {skypilot_nightly-1.0.0.dev20250905.dist-info → skypilot_nightly-1.0.0.dev20251210.dist-info}/entry_points.txt +0 -0
- {skypilot_nightly-1.0.0.dev20250905.dist-info → skypilot_nightly-1.0.0.dev20251210.dist-info}/licenses/LICENSE +0 -0
|
@@ -1,379 +1,10 @@
|
|
|
1
1
|
#!/bin/bash
|
|
2
|
-
# ssh-tunnel.sh - SSH tunnel script for Kubernetes API access
|
|
3
|
-
# Used as kubectl exec credential plugin to establish SSH tunnel on demand.
|
|
4
|
-
# Returns a valid credential format for kubectl with expiration. The expiration
|
|
5
|
-
# is calculated based on the TTL argument and is required to force kubectl to
|
|
6
|
-
# check the tunnel status frequently.
|
|
7
2
|
|
|
8
|
-
#
|
|
3
|
+
# This redirect stub is needed because we use this script in the
|
|
4
|
+
# exec auth section when creating our kubeconfig. Therefore, node pools
|
|
5
|
+
# launched in older versions of SkyPilot will have kubeconfigs pointing
|
|
6
|
+
# to this path.
|
|
9
7
|
|
|
10
|
-
#
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
# Parse arguments
|
|
15
|
-
USE_SSH_CONFIG=0
|
|
16
|
-
SSH_KEY=""
|
|
17
|
-
CONTEXT=""
|
|
18
|
-
HOST=""
|
|
19
|
-
USER=""
|
|
20
|
-
PORT=6443 # Default port if not specified
|
|
21
|
-
|
|
22
|
-
# Debug log to ~/.sky/ssh_node_pools_info/$CONTEXT-tunnel.log
|
|
23
|
-
debug_log() {
|
|
24
|
-
local message="$(date): $1"
|
|
25
|
-
echo "$message" >> "$LOG_FILE"
|
|
26
|
-
}
|
|
27
|
-
|
|
28
|
-
# Generate expiration timestamp for credential
|
|
29
|
-
generate_expiration_timestamp() {
|
|
30
|
-
# Try macOS date format first, fallback to Linux format
|
|
31
|
-
date -u -v+${TTL_SECONDS}S +"%Y-%m-%dT%H:%M:%SZ" 2>/dev/null || date -u -d "+${TTL_SECONDS} seconds" +"%Y-%m-%dT%H:%M:%SZ"
|
|
32
|
-
}
|
|
33
|
-
|
|
34
|
-
# Acquire the lock, return 0 if successful, 1 if another process is already holding the lock
|
|
35
|
-
acquire_lock() {
|
|
36
|
-
# Check for flock command
|
|
37
|
-
if ! command -v flock >/dev/null 2>&1; then
|
|
38
|
-
debug_log "flock command not available, using alternative lock mechanism"
|
|
39
|
-
# Simple file-based locking
|
|
40
|
-
if [ -f "$LOCK_FILE" ]; then
|
|
41
|
-
lock_pid=$(cat "$LOCK_FILE" 2>/dev/null)
|
|
42
|
-
if [ -n "$lock_pid" ] && kill -0 "$lock_pid" 2>/dev/null; then
|
|
43
|
-
debug_log "Another process ($lock_pid) is starting the tunnel, waiting briefly"
|
|
44
|
-
return 1
|
|
45
|
-
else
|
|
46
|
-
# Stale lock file
|
|
47
|
-
debug_log "Removing stale lock file"
|
|
48
|
-
rm -f "$LOCK_FILE"
|
|
49
|
-
fi
|
|
50
|
-
fi
|
|
51
|
-
# Create our lock
|
|
52
|
-
echo $$ > "$LOCK_FILE"
|
|
53
|
-
return 0
|
|
54
|
-
else
|
|
55
|
-
# Use flock for better locking
|
|
56
|
-
exec 9>"$LOCK_FILE"
|
|
57
|
-
if ! flock -n 9; then
|
|
58
|
-
debug_log "Another process is starting the tunnel, waiting briefly"
|
|
59
|
-
return 1
|
|
60
|
-
fi
|
|
61
|
-
return 0
|
|
62
|
-
fi
|
|
63
|
-
}
|
|
64
|
-
|
|
65
|
-
# Release the lock
|
|
66
|
-
release_lock() {
|
|
67
|
-
if command -v flock >/dev/null 2>&1; then
|
|
68
|
-
# Using flock
|
|
69
|
-
exec 9>&- # Close file descriptor to release lock
|
|
70
|
-
else
|
|
71
|
-
# Using simple lock
|
|
72
|
-
rm -f "$LOCK_FILE"
|
|
73
|
-
fi
|
|
74
|
-
debug_log "Lock released"
|
|
75
|
-
}
|
|
76
|
-
|
|
77
|
-
# Generate SSH command based on available tools and parameters
|
|
78
|
-
generate_ssh_command() {
|
|
79
|
-
# Check for autossh
|
|
80
|
-
if ! command -v autossh >/dev/null 2>&1; then
|
|
81
|
-
debug_log "WARNING: autossh is not installed but recommended for reliable SSH tunnels"
|
|
82
|
-
debug_log "Install autossh: brew install autossh (macOS), apt-get install autossh (Ubuntu/Debian)"
|
|
83
|
-
|
|
84
|
-
# Fall back to regular ssh
|
|
85
|
-
if [[ $USE_SSH_CONFIG -eq 1 ]]; then
|
|
86
|
-
SSH_CMD=("ssh" "-o" "ServerAliveInterval=30" "-o" "ServerAliveCountMax=3" "-o" "ExitOnForwardFailure=yes" "-L" "$PORT:127.0.0.1:6443" "-N" "$HOST")
|
|
87
|
-
else
|
|
88
|
-
SSH_CMD=("ssh" "-o" "StrictHostKeyChecking=no" "-o" "IdentitiesOnly=yes" "-o" "ServerAliveInterval=30" "-o" "ServerAliveCountMax=3" "-o" "ExitOnForwardFailure=yes" "-L" "$PORT:127.0.0.1:6443" "-N")
|
|
89
|
-
|
|
90
|
-
# Add SSH key if provided
|
|
91
|
-
if [[ -n "$SSH_KEY" ]]; then
|
|
92
|
-
SSH_CMD+=("-i" "$SSH_KEY")
|
|
93
|
-
fi
|
|
94
|
-
|
|
95
|
-
# Add user@host
|
|
96
|
-
SSH_CMD+=("$USER@$HOST")
|
|
97
|
-
fi
|
|
98
|
-
else
|
|
99
|
-
# Configure autossh
|
|
100
|
-
if [[ $USE_SSH_CONFIG -eq 1 ]]; then
|
|
101
|
-
SSH_CMD=("autossh" "-M" "0" "-o" "ServerAliveInterval=30" "-o" "ServerAliveCountMax=3" "-o" "ExitOnForwardFailure=yes" "-L" "$PORT:127.0.0.1:6443" "-N" "$HOST")
|
|
102
|
-
else
|
|
103
|
-
SSH_CMD=("autossh" "-M" "0" "-o" "StrictHostKeyChecking=no" "-o" "IdentitiesOnly=yes" "-o" "ServerAliveInterval=30" "-o" "ServerAliveCountMax=3" "-o" "ExitOnForwardFailure=yes" "-L" "$PORT:127.0.0.1:6443" "-N")
|
|
104
|
-
|
|
105
|
-
# Add SSH key if provided
|
|
106
|
-
if [[ -n "$SSH_KEY" ]]; then
|
|
107
|
-
SSH_CMD+=("-i" "$SSH_KEY")
|
|
108
|
-
fi
|
|
109
|
-
|
|
110
|
-
# Add user@host
|
|
111
|
-
SSH_CMD+=("$USER@$HOST")
|
|
112
|
-
fi
|
|
113
|
-
fi
|
|
114
|
-
}
|
|
115
|
-
|
|
116
|
-
# Function to read certificate files if they exist
|
|
117
|
-
read_certificate_data() {
|
|
118
|
-
local client_cert_file="$TUNNEL_DIR/$CONTEXT-cert.pem"
|
|
119
|
-
local client_key_file="$TUNNEL_DIR/$CONTEXT-key.pem"
|
|
120
|
-
local cert_data=""
|
|
121
|
-
local key_data=""
|
|
122
|
-
|
|
123
|
-
if [[ -f "$client_cert_file" ]]; then
|
|
124
|
-
# Read the certificate file as is - it's already in PEM format
|
|
125
|
-
cert_data=$(cat "$client_cert_file")
|
|
126
|
-
debug_log "Found client certificate data for context $CONTEXT"
|
|
127
|
-
|
|
128
|
-
# Log the first and last few characters to verify PEM format
|
|
129
|
-
local cert_start=$(head -1 "$client_cert_file")
|
|
130
|
-
local cert_end=$(tail -1 "$client_cert_file")
|
|
131
|
-
debug_log "Certificate starts with: $cert_start"
|
|
132
|
-
debug_log "Certificate ends with: $cert_end"
|
|
133
|
-
|
|
134
|
-
# Check if it has proper PEM format
|
|
135
|
-
if ! grep -q "BEGIN CERTIFICATE" "$client_cert_file" || ! grep -q "END CERTIFICATE" "$client_cert_file"; then
|
|
136
|
-
debug_log "WARNING: Certificate file may not be in proper PEM format"
|
|
137
|
-
# Try to fix it if needed
|
|
138
|
-
if ! grep -q "BEGIN CERTIFICATE" "$client_cert_file"; then
|
|
139
|
-
echo "-----BEGIN CERTIFICATE-----" > "$client_cert_file.fixed"
|
|
140
|
-
cat "$client_cert_file" >> "$client_cert_file.fixed"
|
|
141
|
-
echo "-----END CERTIFICATE-----" >> "$client_cert_file.fixed"
|
|
142
|
-
mv "$client_cert_file.fixed" "$client_cert_file"
|
|
143
|
-
cert_data=$(cat "$client_cert_file")
|
|
144
|
-
debug_log "Fixed certificate format by adding BEGIN/END markers"
|
|
145
|
-
fi
|
|
146
|
-
fi
|
|
147
|
-
fi
|
|
148
|
-
|
|
149
|
-
if [[ -f "$client_key_file" ]]; then
|
|
150
|
-
# Read the key file as is - it's already in PEM format
|
|
151
|
-
key_data=$(cat "$client_key_file")
|
|
152
|
-
debug_log "Found client key data for context $CONTEXT"
|
|
153
|
-
|
|
154
|
-
# Log the first and last few characters to verify PEM format
|
|
155
|
-
local key_start=$(head -1 "$client_key_file")
|
|
156
|
-
local key_end=$(tail -1 "$client_key_file")
|
|
157
|
-
debug_log "Key starts with: $key_start"
|
|
158
|
-
debug_log "Key ends with: $key_end"
|
|
159
|
-
|
|
160
|
-
# Check if it has proper PEM format
|
|
161
|
-
if ! grep -q "BEGIN" "$client_key_file" || ! grep -q "END" "$client_key_file"; then
|
|
162
|
-
debug_log "WARNING: Key file may not be in proper PEM format"
|
|
163
|
-
# Try to fix it if needed
|
|
164
|
-
if ! grep -q "BEGIN" "$client_key_file"; then
|
|
165
|
-
echo "-----BEGIN PRIVATE KEY-----" > "$client_key_file.fixed"
|
|
166
|
-
cat "$client_key_file" >> "$client_key_file.fixed"
|
|
167
|
-
echo "-----END PRIVATE KEY-----" >> "$client_key_file.fixed"
|
|
168
|
-
mv "$client_key_file.fixed" "$client_key_file"
|
|
169
|
-
key_data=$(cat "$client_key_file")
|
|
170
|
-
debug_log "Fixed key format by adding BEGIN/END markers"
|
|
171
|
-
fi
|
|
172
|
-
fi
|
|
173
|
-
fi
|
|
174
|
-
|
|
175
|
-
echo "$cert_data:$key_data"
|
|
176
|
-
}
|
|
177
|
-
|
|
178
|
-
# Function to generate credentials JSON
|
|
179
|
-
generate_credentials_json() {
|
|
180
|
-
local expiration_time=$(generate_expiration_timestamp)
|
|
181
|
-
local cert_bundle=$(read_certificate_data)
|
|
182
|
-
local client_cert_data=${cert_bundle%:*}
|
|
183
|
-
local client_key_data=${cert_bundle#*:}
|
|
184
|
-
|
|
185
|
-
if [[ -n "$client_cert_data" && -n "$client_key_data" ]]; then
|
|
186
|
-
# Debug the certificate data
|
|
187
|
-
debug_log "Certificate data length: $(echo -n "$client_cert_data" | wc -c) bytes"
|
|
188
|
-
debug_log "Key data length: $(echo -n "$client_key_data" | wc -c) bytes"
|
|
189
|
-
|
|
190
|
-
# Check if we can create proper JSON with `jq`
|
|
191
|
-
if ! command -v jq &>/dev/null; then
|
|
192
|
-
echo "jq is not installed. Please install jq to use this script." >&2
|
|
193
|
-
exit 1
|
|
194
|
-
fi
|
|
195
|
-
debug_log "Using jq for JSON formatting"
|
|
196
|
-
|
|
197
|
-
# Create a temporary file for the JSON output to avoid shell escaping issues
|
|
198
|
-
local TEMP_JSON_FILE=$(mktemp)
|
|
199
|
-
|
|
200
|
-
# Write the JSON to the temporary file using jq for proper JSON formatting
|
|
201
|
-
cat > "$TEMP_JSON_FILE" << EOL
|
|
202
|
-
{
|
|
203
|
-
"apiVersion": "client.authentication.k8s.io/v1beta1",
|
|
204
|
-
"kind": "ExecCredential",
|
|
205
|
-
"status": {
|
|
206
|
-
"clientCertificateData": $(printf '%s' "$client_cert_data" | jq -R -s .),
|
|
207
|
-
"clientKeyData": $(printf '%s' "$client_key_data" | jq -R -s .),
|
|
208
|
-
"expirationTimestamp": "$expiration_time"
|
|
209
|
-
}
|
|
210
|
-
}
|
|
211
|
-
EOL
|
|
212
|
-
|
|
213
|
-
# Read the JSON from the file
|
|
214
|
-
local json_response=$(cat "$TEMP_JSON_FILE")
|
|
215
|
-
|
|
216
|
-
# Clean up
|
|
217
|
-
rm -f "$TEMP_JSON_FILE"
|
|
218
|
-
|
|
219
|
-
# Output the JSON
|
|
220
|
-
echo "$json_response"
|
|
221
|
-
else
|
|
222
|
-
# Fallback to token-based credential for tunnel-only authentication
|
|
223
|
-
echo "{\"apiVersion\":\"client.authentication.k8s.io/v1beta1\",\"kind\":\"ExecCredential\",\"status\":{\"token\":\"k8s-ssh-tunnel-token\",\"expirationTimestamp\":\"$expiration_time\"}}"
|
|
224
|
-
fi
|
|
225
|
-
}
|
|
226
|
-
|
|
227
|
-
while [[ $# -gt 0 ]]; do
|
|
228
|
-
case $1 in
|
|
229
|
-
--use-ssh-config)
|
|
230
|
-
USE_SSH_CONFIG=1
|
|
231
|
-
shift
|
|
232
|
-
;;
|
|
233
|
-
--ssh-key)
|
|
234
|
-
SSH_KEY="$2"
|
|
235
|
-
shift 2
|
|
236
|
-
;;
|
|
237
|
-
--context)
|
|
238
|
-
CONTEXT="$2"
|
|
239
|
-
shift 2
|
|
240
|
-
;;
|
|
241
|
-
--port)
|
|
242
|
-
PORT="$2"
|
|
243
|
-
shift 2
|
|
244
|
-
;;
|
|
245
|
-
--host)
|
|
246
|
-
HOST="$2"
|
|
247
|
-
shift 2
|
|
248
|
-
;;
|
|
249
|
-
--user)
|
|
250
|
-
USER="$2"
|
|
251
|
-
shift 2
|
|
252
|
-
;;
|
|
253
|
-
--ttl)
|
|
254
|
-
TTL_SECONDS="$2"
|
|
255
|
-
shift 2
|
|
256
|
-
;;
|
|
257
|
-
*)
|
|
258
|
-
echo "Unknown parameter: $1" >&2
|
|
259
|
-
exit 1
|
|
260
|
-
;;
|
|
261
|
-
esac
|
|
262
|
-
done
|
|
263
|
-
|
|
264
|
-
# Validate required parameters
|
|
265
|
-
if [[ -z "$HOST" ]]; then
|
|
266
|
-
echo "Error: --host parameter is required" >&2
|
|
267
|
-
exit 1
|
|
268
|
-
fi
|
|
269
|
-
|
|
270
|
-
# Setup directories
|
|
271
|
-
TUNNEL_DIR="$HOME/.sky/ssh_node_pools_info"
|
|
272
|
-
mkdir -p "$TUNNEL_DIR"
|
|
273
|
-
|
|
274
|
-
# Get context name for PID file
|
|
275
|
-
if [[ -z "$CONTEXT" ]]; then
|
|
276
|
-
CONTEXT="default"
|
|
277
|
-
fi
|
|
278
|
-
|
|
279
|
-
PID_FILE="$TUNNEL_DIR/$CONTEXT-tunnel.pid"
|
|
280
|
-
LOG_FILE="$TUNNEL_DIR/$CONTEXT-tunnel.log"
|
|
281
|
-
LOCK_FILE="$TUNNEL_DIR/$CONTEXT-tunnel.lock"
|
|
282
|
-
|
|
283
|
-
debug_log "Starting ssh-tunnel.sh for context $CONTEXT, host $HOST, port $PORT"
|
|
284
|
-
debug_log "SSH Config: $USE_SSH_CONFIG, User: $USER, TTL: ${TTL_SECONDS}s"
|
|
285
|
-
|
|
286
|
-
# Check if specified port is already in use (tunnel may be running)
|
|
287
|
-
if nc -z 127.0.0.1 "$PORT" 2>/dev/null; then
|
|
288
|
-
debug_log "Port $PORT already in use, checking if it's our tunnel"
|
|
289
|
-
|
|
290
|
-
# Check if there's a PID file and if that process is running
|
|
291
|
-
if [[ -f "$PID_FILE" ]]; then
|
|
292
|
-
OLD_PID=$(cat "$PID_FILE")
|
|
293
|
-
if kill -0 "$OLD_PID" 2>/dev/null; then
|
|
294
|
-
debug_log "Tunnel appears to be running with PID $OLD_PID"
|
|
295
|
-
else
|
|
296
|
-
debug_log "PID file exists but process $OLD_PID is not running"
|
|
297
|
-
fi
|
|
298
|
-
else
|
|
299
|
-
debug_log "Port $PORT is in use but no PID file exists"
|
|
300
|
-
fi
|
|
301
|
-
|
|
302
|
-
# Return valid credential format for kubectl with expiration
|
|
303
|
-
generate_credentials_json
|
|
304
|
-
exit 0
|
|
305
|
-
fi
|
|
306
|
-
|
|
307
|
-
# Try to acquire the lock
|
|
308
|
-
if ! acquire_lock; then
|
|
309
|
-
# Wait briefly for the tunnel to be established
|
|
310
|
-
for i in {1..10}; do
|
|
311
|
-
if nc -z 127.0.0.1 "$PORT" 2>/dev/null; then
|
|
312
|
-
debug_log "Tunnel is now active"
|
|
313
|
-
|
|
314
|
-
# Return valid credential format for kubectl with expiration
|
|
315
|
-
generate_credentials_json
|
|
316
|
-
exit 0
|
|
317
|
-
fi
|
|
318
|
-
sleep 0.2
|
|
319
|
-
done
|
|
320
|
-
debug_log "Waited for tunnel but port $PORT still not available"
|
|
321
|
-
fi
|
|
322
|
-
|
|
323
|
-
# Check if we have a PID file with running process
|
|
324
|
-
if [[ -f "$PID_FILE" ]]; then
|
|
325
|
-
OLD_PID=$(cat "$PID_FILE")
|
|
326
|
-
if kill -0 "$OLD_PID" 2>/dev/null; then
|
|
327
|
-
# Process exists but port isn't open - something's wrong, kill it
|
|
328
|
-
kill "$OLD_PID" 2>/dev/null
|
|
329
|
-
debug_log "Killed stale tunnel process $OLD_PID"
|
|
330
|
-
else
|
|
331
|
-
debug_log "PID file exists but process $OLD_PID is not running anymore"
|
|
332
|
-
fi
|
|
333
|
-
# Remove the stale PID file
|
|
334
|
-
rm -f "$PID_FILE"
|
|
335
|
-
fi
|
|
336
|
-
|
|
337
|
-
# Generate the SSH command
|
|
338
|
-
generate_ssh_command
|
|
339
|
-
|
|
340
|
-
debug_log "Starting SSH tunnel: ${SSH_CMD[*]}"
|
|
341
|
-
|
|
342
|
-
# Start the tunnel in foreground and wait for it to establish
|
|
343
|
-
"${SSH_CMD[@]}" >> "$LOG_FILE" 2>&1 &
|
|
344
|
-
TUNNEL_PID=$!
|
|
345
|
-
|
|
346
|
-
# Save PID
|
|
347
|
-
echo $TUNNEL_PID > "$PID_FILE"
|
|
348
|
-
debug_log "Tunnel started with PID $TUNNEL_PID"
|
|
349
|
-
|
|
350
|
-
# Wait for tunnel to establish
|
|
351
|
-
tunnel_up=0
|
|
352
|
-
for i in {1..20}; do
|
|
353
|
-
if nc -z 127.0.0.1 "$PORT" 2>/dev/null; then
|
|
354
|
-
debug_log "Tunnel established successfully on port $PORT"
|
|
355
|
-
tunnel_up=1
|
|
356
|
-
break
|
|
357
|
-
fi
|
|
358
|
-
sleep 0.2
|
|
359
|
-
done
|
|
360
|
-
|
|
361
|
-
# Clean up lock file
|
|
362
|
-
release_lock
|
|
363
|
-
|
|
364
|
-
# Check if the tunnel process is still running
|
|
365
|
-
if ! kill -0 $TUNNEL_PID 2>/dev/null; then
|
|
366
|
-
debug_log "ERROR: Tunnel process exited unexpectedly! Check logs for details"
|
|
367
|
-
if [[ -f "$PID_FILE" ]]; then
|
|
368
|
-
rm -f "$PID_FILE"
|
|
369
|
-
fi
|
|
370
|
-
# Return error in case of tunnel failure
|
|
371
|
-
echo "Failed to establish SSH tunnel. See $TUNNEL_DIR/$CONTEXT-tunnel.log for details." >&2
|
|
372
|
-
exit 1
|
|
373
|
-
elif [[ $tunnel_up -eq 0 ]]; then
|
|
374
|
-
debug_log "WARNING: Tunnel process is running but port $PORT is not responding"
|
|
375
|
-
fi
|
|
376
|
-
|
|
377
|
-
# Return valid credential format with certificates if available
|
|
378
|
-
generate_credentials_json
|
|
379
|
-
exit 0
|
|
8
|
+
# TODO (kyuds): remove this script after v0.13.0. Kept here for backwards compat.
|
|
9
|
+
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
|
10
|
+
exec "$SCRIPT_DIR/../../ssh_node_pools/deploy/tunnel/ssh-tunnel.sh" "$@"
|
sky/utils/kubernetes_enums.py
CHANGED
|
@@ -2,26 +2,13 @@
|
|
|
2
2
|
import enum
|
|
3
3
|
|
|
4
4
|
|
|
5
|
+
# TODO(kevin): Remove this enum in v0.13.0.
|
|
5
6
|
class KubernetesNetworkingMode(enum.Enum):
|
|
6
|
-
"""Enum for the different types of networking modes for accessing
|
|
7
|
-
jump pods.
|
|
7
|
+
"""Enum for the different types of networking modes for accessing pods.
|
|
8
8
|
"""
|
|
9
9
|
NODEPORT = 'nodeport'
|
|
10
10
|
PORTFORWARD = 'portforward'
|
|
11
11
|
|
|
12
|
-
@classmethod
|
|
13
|
-
def from_str(cls, mode: str) -> 'KubernetesNetworkingMode':
|
|
14
|
-
"""Returns the enum value for the given string."""
|
|
15
|
-
if mode.lower() == cls.NODEPORT.value:
|
|
16
|
-
return cls.NODEPORT
|
|
17
|
-
elif mode.lower() == cls.PORTFORWARD.value:
|
|
18
|
-
return cls.PORTFORWARD
|
|
19
|
-
else:
|
|
20
|
-
raise ValueError(f'Unsupported kubernetes networking mode: '
|
|
21
|
-
f'{mode}. The mode must be either '
|
|
22
|
-
f'\'{cls.PORTFORWARD.value}\' or '
|
|
23
|
-
f'\'{cls.NODEPORT.value}\'. ')
|
|
24
|
-
|
|
25
12
|
|
|
26
13
|
class KubernetesServiceType(enum.Enum):
|
|
27
14
|
"""Enum for the different types of services."""
|
|
@@ -44,3 +31,8 @@ class KubernetesAutoscalerType(enum.Enum):
|
|
|
44
31
|
KARPENTER = 'karpenter'
|
|
45
32
|
COREWEAVE = 'coreweave'
|
|
46
33
|
GENERIC = 'generic'
|
|
34
|
+
|
|
35
|
+
def emits_autoscale_event(self) -> bool:
|
|
36
|
+
"""Returns whether specific autoscaler emits the event reason
|
|
37
|
+
TriggeredScaleUp."""
|
|
38
|
+
return self not in {self.KARPENTER}
|
sky/utils/lock_events.py
CHANGED
|
@@ -20,17 +20,17 @@ class DistributedLockEvent:
|
|
|
20
20
|
f'[DistributedLock.hold]:{lock_id}')
|
|
21
21
|
|
|
22
22
|
def acquire(self):
|
|
23
|
-
was_locked = self._lock.is_locked
|
|
23
|
+
was_locked = self._lock.is_locked # type: ignore[truthy-function]
|
|
24
24
|
with timeline.Event(f'[DistributedLock.acquire]:{self._lock_id}'):
|
|
25
25
|
self._lock.acquire()
|
|
26
|
-
if not was_locked and self._lock.is_locked:
|
|
26
|
+
if not was_locked and self._lock.is_locked: # type: ignore[truthy-function] # pylint: disable=line-too-long
|
|
27
27
|
# start holding the lock after initial acquiring
|
|
28
28
|
self._hold_lock_event.begin()
|
|
29
29
|
|
|
30
30
|
def release(self):
|
|
31
|
-
was_locked = self._lock.is_locked
|
|
31
|
+
was_locked = self._lock.is_locked # type: ignore[truthy-function]
|
|
32
32
|
self._lock.release()
|
|
33
|
-
if was_locked and not self._lock.is_locked:
|
|
33
|
+
if was_locked and not self._lock.is_locked: # type: ignore[truthy-function] # pylint: disable=line-too-long
|
|
34
34
|
# stop holding the lock after initial releasing
|
|
35
35
|
self._hold_lock_event.end()
|
|
36
36
|
|