pulumiverse-scaleway 1.27.0a1744869119__py3-none-any.whl → 1.39.0a1764841480__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- pulumiverse_scaleway/__init__.py +109 -2
- pulumiverse_scaleway/_inputs.py +3936 -2603
- pulumiverse_scaleway/_utilities.py +1 -1
- pulumiverse_scaleway/account/__init__.py +4 -1
- pulumiverse_scaleway/account/get_availability_zones.py +12 -12
- pulumiverse_scaleway/account/get_project.py +28 -28
- pulumiverse_scaleway/account/get_projects.py +241 -0
- pulumiverse_scaleway/account/get_ssh_key.py +34 -34
- pulumiverse_scaleway/account/outputs.py +95 -0
- pulumiverse_scaleway/account/project.py +72 -71
- pulumiverse_scaleway/account/ssh_key.py +107 -106
- pulumiverse_scaleway/account_project.py +72 -71
- pulumiverse_scaleway/account_ssh_key.py +107 -106
- pulumiverse_scaleway/apple_silicon_server.py +334 -201
- pulumiverse_scaleway/applesilicon/__init__.py +2 -1
- pulumiverse_scaleway/applesilicon/_inputs.py +105 -51
- pulumiverse_scaleway/applesilicon/outputs.py +65 -33
- pulumiverse_scaleway/applesilicon/server.py +334 -201
- pulumiverse_scaleway/audittrail/__init__.py +10 -0
- pulumiverse_scaleway/audittrail/get_event.py +410 -0
- pulumiverse_scaleway/audittrail/outputs.py +234 -0
- pulumiverse_scaleway/autoscaling/__init__.py +13 -0
- pulumiverse_scaleway/autoscaling/_inputs.py +558 -0
- pulumiverse_scaleway/autoscaling/instance_group.py +807 -0
- pulumiverse_scaleway/autoscaling/instance_policy.py +634 -0
- pulumiverse_scaleway/autoscaling/instance_template.py +878 -0
- pulumiverse_scaleway/autoscaling/outputs.py +438 -0
- pulumiverse_scaleway/baremetal_server.py +622 -524
- pulumiverse_scaleway/billing/__init__.py +2 -1
- pulumiverse_scaleway/billing/get_consumptions.py +19 -15
- pulumiverse_scaleway/billing/get_invoices.py +25 -25
- pulumiverse_scaleway/billing/outputs.py +96 -96
- pulumiverse_scaleway/block/__init__.py +4 -1
- pulumiverse_scaleway/block/_inputs.py +125 -0
- pulumiverse_scaleway/block/get_snapshot.py +60 -37
- pulumiverse_scaleway/block/get_volume.py +38 -38
- pulumiverse_scaleway/block/outputs.py +139 -0
- pulumiverse_scaleway/block/snapshot.py +272 -98
- pulumiverse_scaleway/block/volume.py +145 -144
- pulumiverse_scaleway/block_snapshot.py +272 -98
- pulumiverse_scaleway/block_volume.py +145 -144
- pulumiverse_scaleway/cockpit.py +71 -49
- pulumiverse_scaleway/cockpit_alert_manager.py +66 -65
- pulumiverse_scaleway/cockpit_grafana_user.py +72 -113
- pulumiverse_scaleway/cockpit_source.py +142 -141
- pulumiverse_scaleway/cockpit_token.py +84 -83
- pulumiverse_scaleway/config/__init__.py +2 -1
- pulumiverse_scaleway/config/__init__.pyi +2 -2
- pulumiverse_scaleway/config/vars.py +11 -11
- pulumiverse_scaleway/container.py +653 -424
- pulumiverse_scaleway/container_cron.py +97 -96
- pulumiverse_scaleway/container_domain.py +63 -62
- pulumiverse_scaleway/container_namespace.py +251 -172
- pulumiverse_scaleway/container_token.py +97 -96
- pulumiverse_scaleway/container_trigger.py +77 -76
- pulumiverse_scaleway/containers/__init__.py +2 -1
- pulumiverse_scaleway/containers/_inputs.py +87 -87
- pulumiverse_scaleway/containers/container.py +653 -424
- pulumiverse_scaleway/containers/cron.py +97 -96
- pulumiverse_scaleway/containers/domain.py +63 -62
- pulumiverse_scaleway/containers/get_container.py +117 -73
- pulumiverse_scaleway/containers/get_namespace.py +56 -45
- pulumiverse_scaleway/containers/namespace.py +251 -172
- pulumiverse_scaleway/containers/outputs.py +84 -84
- pulumiverse_scaleway/containers/token.py +97 -96
- pulumiverse_scaleway/containers/trigger.py +77 -76
- pulumiverse_scaleway/database.py +81 -80
- pulumiverse_scaleway/database_acl.py +148 -39
- pulumiverse_scaleway/database_backup.py +124 -123
- pulumiverse_scaleway/database_instance.py +531 -436
- pulumiverse_scaleway/database_privilege.py +88 -87
- pulumiverse_scaleway/database_read_replica.py +60 -59
- pulumiverse_scaleway/database_user.py +171 -98
- pulumiverse_scaleway/databases/__init__.py +2 -1
- pulumiverse_scaleway/databases/_inputs.py +364 -216
- pulumiverse_scaleway/databases/acl.py +148 -39
- pulumiverse_scaleway/databases/database.py +81 -80
- pulumiverse_scaleway/databases/database_backup.py +124 -123
- pulumiverse_scaleway/databases/get_acl.py +17 -17
- pulumiverse_scaleway/databases/get_database.py +26 -26
- pulumiverse_scaleway/databases/get_database_backup.py +46 -46
- pulumiverse_scaleway/databases/get_instance.py +95 -73
- pulumiverse_scaleway/databases/get_privilege.py +30 -30
- pulumiverse_scaleway/databases/instance.py +531 -436
- pulumiverse_scaleway/databases/outputs.py +419 -232
- pulumiverse_scaleway/databases/privilege.py +88 -87
- pulumiverse_scaleway/databases/read_replica.py +60 -59
- pulumiverse_scaleway/databases/serverless_database.py +97 -96
- pulumiverse_scaleway/databases/snapshot.py +127 -126
- pulumiverse_scaleway/databases/user.py +171 -98
- pulumiverse_scaleway/datawarehouse/__init__.py +13 -0
- pulumiverse_scaleway/datawarehouse/_inputs.py +149 -0
- pulumiverse_scaleway/datawarehouse/database.py +338 -0
- pulumiverse_scaleway/datawarehouse/deployment.py +821 -0
- pulumiverse_scaleway/datawarehouse/outputs.py +113 -0
- pulumiverse_scaleway/datawarehouse/user.py +451 -0
- pulumiverse_scaleway/domain/__init__.py +2 -1
- pulumiverse_scaleway/domain/_inputs.py +661 -661
- pulumiverse_scaleway/domain/get_record.py +53 -64
- pulumiverse_scaleway/domain/get_zone.py +36 -30
- pulumiverse_scaleway/domain/outputs.py +501 -501
- pulumiverse_scaleway/domain/record.py +156 -202
- pulumiverse_scaleway/domain/registration.py +123 -122
- pulumiverse_scaleway/domain/zone.py +108 -107
- pulumiverse_scaleway/domain_record.py +156 -202
- pulumiverse_scaleway/domain_zone.py +108 -107
- pulumiverse_scaleway/edge_services_backend_stage.py +61 -60
- pulumiverse_scaleway/edge_services_cache_stage.py +283 -111
- pulumiverse_scaleway/edge_services_dns_stage.py +169 -140
- pulumiverse_scaleway/edge_services_head_stage.py +37 -36
- pulumiverse_scaleway/edge_services_pipeline.py +125 -82
- pulumiverse_scaleway/edge_services_plan.py +37 -36
- pulumiverse_scaleway/edge_services_route_stage.py +75 -74
- pulumiverse_scaleway/edge_services_tls_stage.py +219 -124
- pulumiverse_scaleway/edge_services_waf_stage.py +106 -105
- pulumiverse_scaleway/elasticmetal/__init__.py +3 -1
- pulumiverse_scaleway/elasticmetal/_inputs.py +203 -129
- pulumiverse_scaleway/elasticmetal/get_ip.py +70 -54
- pulumiverse_scaleway/elasticmetal/get_ips.py +31 -31
- pulumiverse_scaleway/elasticmetal/get_offer.py +43 -41
- pulumiverse_scaleway/elasticmetal/get_option.py +24 -24
- pulumiverse_scaleway/elasticmetal/get_os.py +28 -28
- pulumiverse_scaleway/elasticmetal/get_partition_schema.py +215 -0
- pulumiverse_scaleway/elasticmetal/get_server.py +101 -68
- pulumiverse_scaleway/elasticmetal/ip.py +170 -169
- pulumiverse_scaleway/elasticmetal/ip_mac_address.py +107 -106
- pulumiverse_scaleway/elasticmetal/outputs.py +379 -287
- pulumiverse_scaleway/elasticmetal/server.py +622 -524
- pulumiverse_scaleway/file_filesystem.py +555 -0
- pulumiverse_scaleway/flexible_ip.py +170 -169
- pulumiverse_scaleway/flexible_ip_mac_address.py +107 -106
- pulumiverse_scaleway/function.py +457 -348
- pulumiverse_scaleway/function_cron.py +97 -96
- pulumiverse_scaleway/function_domain.py +63 -62
- pulumiverse_scaleway/function_namespace.py +234 -155
- pulumiverse_scaleway/function_token.py +97 -96
- pulumiverse_scaleway/function_trigger.py +77 -76
- pulumiverse_scaleway/functions/__init__.py +2 -1
- pulumiverse_scaleway/functions/_inputs.py +50 -50
- pulumiverse_scaleway/functions/cron.py +97 -96
- pulumiverse_scaleway/functions/domain.py +63 -62
- pulumiverse_scaleway/functions/function.py +457 -348
- pulumiverse_scaleway/functions/get_function.py +93 -71
- pulumiverse_scaleway/functions/get_namespace.py +54 -43
- pulumiverse_scaleway/functions/namespace.py +234 -155
- pulumiverse_scaleway/functions/outputs.py +34 -34
- pulumiverse_scaleway/functions/token.py +97 -96
- pulumiverse_scaleway/functions/trigger.py +77 -76
- pulumiverse_scaleway/get_account_project.py +28 -28
- pulumiverse_scaleway/get_account_ssh_key.py +34 -34
- pulumiverse_scaleway/get_availability_zones.py +12 -12
- pulumiverse_scaleway/get_baremetal_offer.py +43 -41
- pulumiverse_scaleway/get_baremetal_option.py +24 -24
- pulumiverse_scaleway/get_baremetal_os.py +28 -28
- pulumiverse_scaleway/get_baremetal_server.py +101 -68
- pulumiverse_scaleway/get_billing_consumptions.py +19 -15
- pulumiverse_scaleway/get_billing_invoices.py +25 -25
- pulumiverse_scaleway/get_block_snapshot.py +60 -37
- pulumiverse_scaleway/get_block_volume.py +38 -38
- pulumiverse_scaleway/get_cockpit.py +16 -16
- pulumiverse_scaleway/get_cockpit_source.py +88 -45
- pulumiverse_scaleway/get_config.py +24 -24
- pulumiverse_scaleway/get_container.py +117 -73
- pulumiverse_scaleway/get_container_namespace.py +56 -45
- pulumiverse_scaleway/get_database.py +26 -26
- pulumiverse_scaleway/get_database_acl.py +17 -17
- pulumiverse_scaleway/get_database_backup.py +46 -46
- pulumiverse_scaleway/get_database_instance.py +95 -73
- pulumiverse_scaleway/get_database_privilege.py +30 -30
- pulumiverse_scaleway/get_domain_record.py +53 -64
- pulumiverse_scaleway/get_domain_zone.py +36 -30
- pulumiverse_scaleway/get_flexible_ip.py +70 -54
- pulumiverse_scaleway/get_flexible_ips.py +31 -31
- pulumiverse_scaleway/get_function.py +93 -71
- pulumiverse_scaleway/get_function_namespace.py +54 -43
- pulumiverse_scaleway/get_iam_api_key.py +28 -28
- pulumiverse_scaleway/get_iam_application.py +32 -32
- pulumiverse_scaleway/get_iam_group.py +36 -36
- pulumiverse_scaleway/get_iam_ssh_key.py +34 -34
- pulumiverse_scaleway/get_iam_user.py +30 -30
- pulumiverse_scaleway/get_instance_image.py +58 -58
- pulumiverse_scaleway/get_instance_ip.py +42 -30
- pulumiverse_scaleway/get_instance_placement_group.py +38 -38
- pulumiverse_scaleway/get_instance_private_nic.py +53 -41
- pulumiverse_scaleway/get_instance_security_group.py +46 -46
- pulumiverse_scaleway/get_instance_server.py +108 -145
- pulumiverse_scaleway/get_instance_servers.py +31 -31
- pulumiverse_scaleway/get_instance_snapshot.py +41 -41
- pulumiverse_scaleway/get_instance_volume.py +42 -42
- pulumiverse_scaleway/get_iot_device.py +46 -46
- pulumiverse_scaleway/get_iot_hub.py +60 -60
- pulumiverse_scaleway/get_ipam_ip.py +65 -65
- pulumiverse_scaleway/get_ipam_ips.py +56 -56
- pulumiverse_scaleway/get_k8s_version.py +34 -23
- pulumiverse_scaleway/get_kubernetes_cluster.py +98 -65
- pulumiverse_scaleway/get_kubernetes_node_pool.py +86 -75
- pulumiverse_scaleway/get_lb_acls.py +29 -29
- pulumiverse_scaleway/get_lb_backend.py +75 -75
- pulumiverse_scaleway/get_lb_backends.py +29 -29
- pulumiverse_scaleway/get_lb_frontend.py +76 -43
- pulumiverse_scaleway/get_lb_frontends.py +29 -29
- pulumiverse_scaleway/get_lb_ips.py +37 -37
- pulumiverse_scaleway/get_lb_route.py +36 -25
- pulumiverse_scaleway/get_lb_routes.py +23 -23
- pulumiverse_scaleway/get_lbs.py +31 -31
- pulumiverse_scaleway/get_loadbalancer.py +78 -56
- pulumiverse_scaleway/get_loadbalancer_certificate.py +96 -36
- pulumiverse_scaleway/get_loadbalancer_ip.py +40 -40
- pulumiverse_scaleway/get_marketplace_image.py +28 -28
- pulumiverse_scaleway/get_mnq_sns.py +18 -18
- pulumiverse_scaleway/get_mnq_sqs.py +18 -18
- pulumiverse_scaleway/get_mongo_db_instance.py +110 -55
- pulumiverse_scaleway/get_object_bucket.py +35 -35
- pulumiverse_scaleway/get_object_bucket_policy.py +22 -22
- pulumiverse_scaleway/get_redis_cluster.py +65 -54
- pulumiverse_scaleway/get_registry_image.py +48 -48
- pulumiverse_scaleway/get_registry_image_tag.py +44 -44
- pulumiverse_scaleway/get_registry_namespace.py +36 -36
- pulumiverse_scaleway/get_secret.py +58 -58
- pulumiverse_scaleway/get_secret_version.py +46 -46
- pulumiverse_scaleway/get_tem_domain.py +103 -70
- pulumiverse_scaleway/get_vpc.py +58 -47
- pulumiverse_scaleway/get_vpc_gateway_network.py +55 -44
- pulumiverse_scaleway/get_vpc_private_network.py +58 -47
- pulumiverse_scaleway/get_vpc_public_gateway.py +56 -56
- pulumiverse_scaleway/get_vpc_public_gateway_dhcp.py +44 -44
- pulumiverse_scaleway/get_vpc_public_gateway_dhcp_reservation.py +44 -44
- pulumiverse_scaleway/get_vpc_public_gateway_ip.py +24 -24
- pulumiverse_scaleway/get_vpc_public_pat_rule.py +32 -32
- pulumiverse_scaleway/get_vpc_routes.py +47 -47
- pulumiverse_scaleway/get_vpcs.py +31 -31
- pulumiverse_scaleway/get_web_host_offer.py +34 -34
- pulumiverse_scaleway/get_webhosting.py +58 -58
- pulumiverse_scaleway/hosting/__init__.py +2 -1
- pulumiverse_scaleway/hosting/_inputs.py +117 -75
- pulumiverse_scaleway/hosting/get_hosting.py +58 -58
- pulumiverse_scaleway/hosting/get_offer.py +34 -34
- pulumiverse_scaleway/hosting/hosting.py +213 -212
- pulumiverse_scaleway/hosting/outputs.py +276 -192
- pulumiverse_scaleway/iam/__init__.py +3 -1
- pulumiverse_scaleway/iam/_inputs.py +26 -26
- pulumiverse_scaleway/iam/api_key.py +142 -141
- pulumiverse_scaleway/iam/application.py +98 -97
- pulumiverse_scaleway/iam/get_api_key.py +28 -28
- pulumiverse_scaleway/iam/get_application.py +32 -32
- pulumiverse_scaleway/iam/get_group.py +36 -36
- pulumiverse_scaleway/iam/get_policy.py +293 -0
- pulumiverse_scaleway/iam/get_ssh_key.py +34 -34
- pulumiverse_scaleway/iam/get_user.py +30 -30
- pulumiverse_scaleway/iam/group.py +140 -139
- pulumiverse_scaleway/iam/group_membership.py +54 -53
- pulumiverse_scaleway/iam/outputs.py +70 -18
- pulumiverse_scaleway/iam/policy.py +169 -168
- pulumiverse_scaleway/iam/ssh_key.py +107 -106
- pulumiverse_scaleway/iam/user.py +598 -134
- pulumiverse_scaleway/iam_api_key.py +142 -141
- pulumiverse_scaleway/iam_application.py +98 -97
- pulumiverse_scaleway/iam_group.py +140 -139
- pulumiverse_scaleway/iam_group_membership.py +54 -53
- pulumiverse_scaleway/iam_policy.py +169 -168
- pulumiverse_scaleway/iam_ssh_key.py +107 -106
- pulumiverse_scaleway/iam_user.py +598 -134
- pulumiverse_scaleway/inference/__init__.py +4 -1
- pulumiverse_scaleway/inference/_inputs.py +232 -50
- pulumiverse_scaleway/inference/deployment.py +319 -220
- pulumiverse_scaleway/inference/get_model.py +299 -0
- pulumiverse_scaleway/inference/model.py +684 -0
- pulumiverse_scaleway/inference/outputs.py +250 -34
- pulumiverse_scaleway/inference_deployment.py +319 -220
- pulumiverse_scaleway/instance/__init__.py +3 -1
- pulumiverse_scaleway/instance/_inputs.py +650 -437
- pulumiverse_scaleway/instance/get_image.py +58 -58
- pulumiverse_scaleway/instance/get_ip.py +42 -30
- pulumiverse_scaleway/instance/get_placement_group.py +38 -38
- pulumiverse_scaleway/instance/get_private_nic.py +53 -41
- pulumiverse_scaleway/instance/get_security_group.py +46 -46
- pulumiverse_scaleway/instance/get_server.py +108 -145
- pulumiverse_scaleway/instance/get_server_type.py +271 -0
- pulumiverse_scaleway/instance/get_servers.py +31 -31
- pulumiverse_scaleway/instance/get_snapshot.py +41 -41
- pulumiverse_scaleway/instance/get_volume.py +42 -42
- pulumiverse_scaleway/instance/image.py +260 -205
- pulumiverse_scaleway/instance/ip.py +116 -115
- pulumiverse_scaleway/instance/ip_reverse_dns.py +54 -53
- pulumiverse_scaleway/instance/outputs.py +981 -548
- pulumiverse_scaleway/instance/placement_group.py +123 -122
- pulumiverse_scaleway/instance/private_nic.py +163 -113
- pulumiverse_scaleway/instance/security_group.py +188 -187
- pulumiverse_scaleway/instance/security_group_rules.py +42 -41
- pulumiverse_scaleway/instance/server.py +590 -562
- pulumiverse_scaleway/instance/snapshot.py +150 -157
- pulumiverse_scaleway/instance/user_data.py +75 -74
- pulumiverse_scaleway/instance/volume.py +174 -159
- pulumiverse_scaleway/instance_image.py +260 -205
- pulumiverse_scaleway/instance_ip.py +116 -115
- pulumiverse_scaleway/instance_ip_reverse_dns.py +54 -53
- pulumiverse_scaleway/instance_placement_group.py +123 -122
- pulumiverse_scaleway/instance_private_nic.py +163 -113
- pulumiverse_scaleway/instance_security_group.py +188 -187
- pulumiverse_scaleway/instance_security_group_rules.py +42 -41
- pulumiverse_scaleway/instance_server.py +590 -562
- pulumiverse_scaleway/instance_snapshot.py +150 -157
- pulumiverse_scaleway/instance_user_data.py +75 -74
- pulumiverse_scaleway/instance_volume.py +174 -159
- pulumiverse_scaleway/iot/__init__.py +2 -1
- pulumiverse_scaleway/iot/_inputs.py +119 -119
- pulumiverse_scaleway/iot/device.py +156 -193
- pulumiverse_scaleway/iot/get_device.py +46 -46
- pulumiverse_scaleway/iot/get_hub.py +60 -60
- pulumiverse_scaleway/iot/hub.py +259 -258
- pulumiverse_scaleway/iot/network.py +115 -114
- pulumiverse_scaleway/iot/outputs.py +107 -107
- pulumiverse_scaleway/iot/route.py +109 -108
- pulumiverse_scaleway/iot_device.py +156 -193
- pulumiverse_scaleway/iot_hub.py +259 -258
- pulumiverse_scaleway/iot_network.py +115 -114
- pulumiverse_scaleway/iot_route.py +109 -108
- pulumiverse_scaleway/ipam/__init__.py +2 -1
- pulumiverse_scaleway/ipam/_inputs.py +108 -108
- pulumiverse_scaleway/ipam/get_ip.py +65 -65
- pulumiverse_scaleway/ipam/get_ips.py +56 -56
- pulumiverse_scaleway/ipam/ip.py +160 -131
- pulumiverse_scaleway/ipam/ip_reverse_dns.py +141 -70
- pulumiverse_scaleway/ipam/outputs.py +122 -122
- pulumiverse_scaleway/ipam_ip.py +160 -131
- pulumiverse_scaleway/ipam_ip_reverse_dns.py +141 -70
- pulumiverse_scaleway/job/__init__.py +2 -1
- pulumiverse_scaleway/job/_inputs.py +127 -14
- pulumiverse_scaleway/job/definition.py +283 -175
- pulumiverse_scaleway/job/outputs.py +98 -10
- pulumiverse_scaleway/job_definition.py +283 -175
- pulumiverse_scaleway/key_manager_key.py +939 -0
- pulumiverse_scaleway/kubernetes/__init__.py +3 -1
- pulumiverse_scaleway/kubernetes/_inputs.py +388 -185
- pulumiverse_scaleway/kubernetes/acl.py +447 -0
- pulumiverse_scaleway/kubernetes/cluster.py +594 -307
- pulumiverse_scaleway/kubernetes/get_cluster.py +98 -65
- pulumiverse_scaleway/kubernetes/get_pool.py +86 -75
- pulumiverse_scaleway/kubernetes/get_version.py +34 -23
- pulumiverse_scaleway/kubernetes/outputs.py +444 -250
- pulumiverse_scaleway/kubernetes/pool.py +427 -365
- pulumiverse_scaleway/kubernetes_cluster.py +594 -307
- pulumiverse_scaleway/kubernetes_node_pool.py +427 -365
- pulumiverse_scaleway/loadbalancer.py +329 -246
- pulumiverse_scaleway/loadbalancer_acl.py +102 -101
- pulumiverse_scaleway/loadbalancer_backend.py +534 -487
- pulumiverse_scaleway/loadbalancer_certificate.py +97 -96
- pulumiverse_scaleway/loadbalancer_frontend.py +271 -167
- pulumiverse_scaleway/loadbalancer_ip.py +124 -123
- pulumiverse_scaleway/loadbalancer_route.py +244 -119
- pulumiverse_scaleway/loadbalancers/__init__.py +3 -1
- pulumiverse_scaleway/loadbalancers/_inputs.py +367 -273
- pulumiverse_scaleway/loadbalancers/acl.py +102 -101
- pulumiverse_scaleway/loadbalancers/backend.py +534 -487
- pulumiverse_scaleway/loadbalancers/certificate.py +97 -96
- pulumiverse_scaleway/loadbalancers/frontend.py +271 -167
- pulumiverse_scaleway/loadbalancers/get_acls.py +29 -29
- pulumiverse_scaleway/loadbalancers/get_backend.py +75 -75
- pulumiverse_scaleway/loadbalancers/get_backends.py +29 -29
- pulumiverse_scaleway/loadbalancers/get_certificate.py +96 -36
- pulumiverse_scaleway/loadbalancers/get_frontend.py +76 -43
- pulumiverse_scaleway/loadbalancers/get_frontends.py +29 -29
- pulumiverse_scaleway/loadbalancers/get_ip.py +40 -40
- pulumiverse_scaleway/loadbalancers/get_ips.py +37 -37
- pulumiverse_scaleway/loadbalancers/get_load_balancer.py +78 -56
- pulumiverse_scaleway/loadbalancers/get_load_balancers.py +31 -31
- pulumiverse_scaleway/loadbalancers/get_route.py +36 -25
- pulumiverse_scaleway/loadbalancers/get_routes.py +23 -23
- pulumiverse_scaleway/loadbalancers/ip.py +124 -123
- pulumiverse_scaleway/loadbalancers/load_balancer.py +329 -246
- pulumiverse_scaleway/loadbalancers/outputs.py +870 -716
- pulumiverse_scaleway/loadbalancers/private_network.py +505 -0
- pulumiverse_scaleway/loadbalancers/route.py +244 -119
- pulumiverse_scaleway/mnq/__init__.py +2 -1
- pulumiverse_scaleway/mnq/_inputs.py +90 -38
- pulumiverse_scaleway/mnq/get_sns.py +18 -18
- pulumiverse_scaleway/mnq/get_sqs.py +18 -18
- pulumiverse_scaleway/mnq/nats_account.py +63 -62
- pulumiverse_scaleway/mnq/nats_credentials.py +63 -62
- pulumiverse_scaleway/mnq/outputs.py +73 -26
- pulumiverse_scaleway/mnq/sns.py +46 -45
- pulumiverse_scaleway/mnq/sns_credentials.py +82 -81
- pulumiverse_scaleway/mnq/sns_topic.py +174 -173
- pulumiverse_scaleway/mnq/sns_topic_subscription.py +182 -181
- pulumiverse_scaleway/mnq/sqs.py +46 -45
- pulumiverse_scaleway/mnq/sqs_credentials.py +82 -81
- pulumiverse_scaleway/mnq/sqs_queue.py +392 -232
- pulumiverse_scaleway/mnq_nats_account.py +63 -62
- pulumiverse_scaleway/mnq_nats_credentials.py +63 -62
- pulumiverse_scaleway/mnq_sns.py +46 -45
- pulumiverse_scaleway/mnq_sns_credentials.py +82 -81
- pulumiverse_scaleway/mnq_sns_topic.py +174 -173
- pulumiverse_scaleway/mnq_sns_topic_subscription.py +182 -181
- pulumiverse_scaleway/mnq_sqs.py +46 -45
- pulumiverse_scaleway/mnq_sqs_credentials.py +82 -81
- pulumiverse_scaleway/mnq_sqs_queue.py +392 -232
- pulumiverse_scaleway/mongo_db_instance.py +569 -268
- pulumiverse_scaleway/mongo_db_snapshot.py +127 -126
- pulumiverse_scaleway/mongodb/__init__.py +3 -1
- pulumiverse_scaleway/mongodb/_inputs.py +178 -51
- pulumiverse_scaleway/mongodb/get_instance.py +110 -55
- pulumiverse_scaleway/mongodb/instance.py +569 -268
- pulumiverse_scaleway/mongodb/outputs.py +192 -68
- pulumiverse_scaleway/mongodb/snapshot.py +127 -126
- pulumiverse_scaleway/mongodb/user.py +507 -0
- pulumiverse_scaleway/network/__init__.py +2 -1
- pulumiverse_scaleway/network/_inputs.py +209 -155
- pulumiverse_scaleway/network/acl.py +100 -101
- pulumiverse_scaleway/network/gateway_network.py +228 -180
- pulumiverse_scaleway/network/get_gateway_network.py +55 -44
- pulumiverse_scaleway/network/get_private_network.py +58 -47
- pulumiverse_scaleway/network/get_public_gateway.py +56 -56
- pulumiverse_scaleway/network/get_public_gateway_dhcp.py +44 -44
- pulumiverse_scaleway/network/get_public_gateway_dhcp_reservation.py +44 -44
- pulumiverse_scaleway/network/get_public_gateway_ip.py +24 -24
- pulumiverse_scaleway/network/get_public_gateway_pat_rule.py +32 -32
- pulumiverse_scaleway/network/get_routes.py +47 -47
- pulumiverse_scaleway/network/get_vpc.py +58 -47
- pulumiverse_scaleway/network/get_vpcs.py +31 -31
- pulumiverse_scaleway/network/outputs.py +338 -270
- pulumiverse_scaleway/network/private_network.py +203 -155
- pulumiverse_scaleway/network/public_gateway.py +261 -260
- pulumiverse_scaleway/network/public_gateway_dhcp.py +285 -284
- pulumiverse_scaleway/network/public_gateway_dhcp_reservation.py +107 -106
- pulumiverse_scaleway/network/public_gateway_ip.py +107 -106
- pulumiverse_scaleway/network/public_gateway_ip_reverse_dns.py +54 -53
- pulumiverse_scaleway/network/public_gateway_pat_rule.py +132 -131
- pulumiverse_scaleway/network/route.py +230 -141
- pulumiverse_scaleway/network/vpc.py +172 -124
- pulumiverse_scaleway/object/__init__.py +3 -1
- pulumiverse_scaleway/object/_inputs.py +202 -166
- pulumiverse_scaleway/object/bucket.py +162 -148
- pulumiverse_scaleway/object/bucket_acl.py +97 -94
- pulumiverse_scaleway/object/bucket_lock_configuration.py +57 -56
- pulumiverse_scaleway/object/bucket_policy.py +71 -134
- pulumiverse_scaleway/object/bucket_website_configuration.py +104 -85
- pulumiverse_scaleway/object/get.py +290 -0
- pulumiverse_scaleway/object/get_bucket.py +35 -35
- pulumiverse_scaleway/object/get_bucket_policy.py +22 -22
- pulumiverse_scaleway/object/item.py +301 -223
- pulumiverse_scaleway/object/outputs.py +210 -166
- pulumiverse_scaleway/object_bucket.py +162 -148
- pulumiverse_scaleway/object_bucket_acl.py +97 -94
- pulumiverse_scaleway/object_bucket_lock_configuration.py +57 -56
- pulumiverse_scaleway/object_bucket_policy.py +71 -134
- pulumiverse_scaleway/object_bucket_website_configuration.py +104 -85
- pulumiverse_scaleway/object_item.py +301 -223
- pulumiverse_scaleway/observability/__init__.py +4 -2
- pulumiverse_scaleway/observability/_inputs.py +106 -106
- pulumiverse_scaleway/observability/alert_manager.py +66 -65
- pulumiverse_scaleway/observability/cockpit.py +71 -49
- pulumiverse_scaleway/observability/get_grafana.py +103 -0
- pulumiverse_scaleway/observability/get_instance.py +16 -16
- pulumiverse_scaleway/observability/get_source.py +88 -45
- pulumiverse_scaleway/observability/get_sources.py +375 -0
- pulumiverse_scaleway/observability/grafana_user.py +72 -113
- pulumiverse_scaleway/observability/outputs.py +241 -101
- pulumiverse_scaleway/observability/source.py +142 -141
- pulumiverse_scaleway/observability/token.py +84 -83
- pulumiverse_scaleway/outputs.py +5235 -3665
- pulumiverse_scaleway/provider.py +107 -82
- pulumiverse_scaleway/pulumi-plugin.json +1 -1
- pulumiverse_scaleway/rdb_snapshot.py +127 -126
- pulumiverse_scaleway/redis/__init__.py +2 -1
- pulumiverse_scaleway/redis/_inputs.py +163 -69
- pulumiverse_scaleway/redis/cluster.py +273 -225
- pulumiverse_scaleway/redis/get_cluster.py +65 -54
- pulumiverse_scaleway/redis/outputs.py +199 -87
- pulumiverse_scaleway/redis_cluster.py +273 -225
- pulumiverse_scaleway/registry/__init__.py +2 -1
- pulumiverse_scaleway/registry/get_image.py +48 -48
- pulumiverse_scaleway/registry/get_image_tag.py +44 -44
- pulumiverse_scaleway/registry/get_namespace.py +36 -36
- pulumiverse_scaleway/registry/namespace.py +106 -105
- pulumiverse_scaleway/registry_namespace.py +106 -105
- pulumiverse_scaleway/sdb_database.py +97 -96
- pulumiverse_scaleway/secret.py +188 -179
- pulumiverse_scaleway/secret_version.py +107 -106
- pulumiverse_scaleway/secrets/__init__.py +2 -1
- pulumiverse_scaleway/secrets/_inputs.py +63 -63
- pulumiverse_scaleway/secrets/get_secret.py +58 -58
- pulumiverse_scaleway/secrets/get_version.py +46 -46
- pulumiverse_scaleway/secrets/outputs.py +84 -84
- pulumiverse_scaleway/secrets/secret.py +188 -179
- pulumiverse_scaleway/secrets/version.py +107 -106
- pulumiverse_scaleway/tem/__init__.py +2 -1
- pulumiverse_scaleway/tem/_inputs.py +33 -33
- pulumiverse_scaleway/tem/blocked_list.py +105 -104
- pulumiverse_scaleway/tem/domain.py +384 -257
- pulumiverse_scaleway/tem/domain_validation.py +63 -62
- pulumiverse_scaleway/tem/get_domain.py +103 -70
- pulumiverse_scaleway/tem/get_offer_subscription.py +34 -34
- pulumiverse_scaleway/tem/outputs.py +44 -44
- pulumiverse_scaleway/tem/webhook.py +132 -131
- pulumiverse_scaleway/tem_domain.py +384 -257
- pulumiverse_scaleway/tem_domain_validation.py +63 -62
- pulumiverse_scaleway/tem_webhook.py +132 -131
- pulumiverse_scaleway/vpc.py +172 -124
- pulumiverse_scaleway/vpc_gateway_network.py +228 -180
- pulumiverse_scaleway/vpc_private_network.py +203 -155
- pulumiverse_scaleway/vpc_public_gateway.py +261 -260
- pulumiverse_scaleway/vpc_public_gateway_dhcp.py +285 -284
- pulumiverse_scaleway/vpc_public_gateway_dhcp_reservation.py +107 -106
- pulumiverse_scaleway/vpc_public_gateway_ip.py +107 -106
- pulumiverse_scaleway/vpc_public_gateway_ip_reverse_dns.py +54 -53
- pulumiverse_scaleway/vpc_public_gateway_pat_rule.py +132 -131
- pulumiverse_scaleway/vpc_route.py +230 -141
- pulumiverse_scaleway/webhosting.py +213 -212
- {pulumiverse_scaleway-1.27.0a1744869119.dist-info → pulumiverse_scaleway-1.39.0a1764841480.dist-info}/METADATA +2 -2
- pulumiverse_scaleway-1.39.0a1764841480.dist-info/RECORD +513 -0
- {pulumiverse_scaleway-1.27.0a1744869119.dist-info → pulumiverse_scaleway-1.39.0a1764841480.dist-info}/WHEEL +1 -1
- pulumiverse_scaleway/get_cockpit_plan.py +0 -129
- pulumiverse_scaleway/observability/get_plan.py +0 -125
- pulumiverse_scaleway-1.27.0a1744869119.dist-info/RECORD +0 -483
- {pulumiverse_scaleway-1.27.0a1744869119.dist-info → pulumiverse_scaleway-1.39.0a1764841480.dist-info}/top_level.txt +0 -0
|
@@ -1,8 +1,8 @@
|
|
|
1
1
|
# coding=utf-8
|
|
2
|
-
# *** WARNING: this file was generated by
|
|
2
|
+
# *** WARNING: this file was generated by pulumi-language-python. ***
|
|
3
3
|
# *** Do not edit by hand unless you're certain you know what you are doing! ***
|
|
4
4
|
|
|
5
|
-
import
|
|
5
|
+
import builtins as _builtins
|
|
6
6
|
import warnings
|
|
7
7
|
import sys
|
|
8
8
|
import pulumi
|
|
@@ -13,22 +13,108 @@ if sys.version_info >= (3, 11):
|
|
|
13
13
|
else:
|
|
14
14
|
from typing_extensions import NotRequired, TypedDict, TypeAlias
|
|
15
15
|
from .. import _utilities
|
|
16
|
+
from . import outputs
|
|
16
17
|
|
|
17
18
|
__all__ = [
|
|
19
|
+
'AclAclRule',
|
|
18
20
|
'ClusterAutoUpgrade',
|
|
19
21
|
'ClusterAutoscalerConfig',
|
|
20
22
|
'ClusterKubeconfig',
|
|
21
23
|
'ClusterOpenIdConnectConfig',
|
|
22
24
|
'PoolNode',
|
|
25
|
+
'PoolNodePrivateIp',
|
|
23
26
|
'PoolUpgradePolicy',
|
|
24
27
|
'GetClusterAutoUpgradeResult',
|
|
25
28
|
'GetClusterAutoscalerConfigResult',
|
|
26
29
|
'GetClusterKubeconfigResult',
|
|
27
30
|
'GetClusterOpenIdConnectConfigResult',
|
|
28
31
|
'GetPoolNodeResult',
|
|
32
|
+
'GetPoolNodePrivateIpResult',
|
|
29
33
|
'GetPoolUpgradePolicyResult',
|
|
30
34
|
]
|
|
31
35
|
|
|
36
|
+
@pulumi.output_type
|
|
37
|
+
class AclAclRule(dict):
|
|
38
|
+
@staticmethod
|
|
39
|
+
def __key_warning(key: str):
|
|
40
|
+
suggest = None
|
|
41
|
+
if key == "scalewayRanges":
|
|
42
|
+
suggest = "scaleway_ranges"
|
|
43
|
+
|
|
44
|
+
if suggest:
|
|
45
|
+
pulumi.log.warn(f"Key '{key}' not found in AclAclRule. Access the value via the '{suggest}' property getter instead.")
|
|
46
|
+
|
|
47
|
+
def __getitem__(self, key: str) -> Any:
|
|
48
|
+
AclAclRule.__key_warning(key)
|
|
49
|
+
return super().__getitem__(key)
|
|
50
|
+
|
|
51
|
+
def get(self, key: str, default = None) -> Any:
|
|
52
|
+
AclAclRule.__key_warning(key)
|
|
53
|
+
return super().get(key, default)
|
|
54
|
+
|
|
55
|
+
def __init__(__self__, *,
|
|
56
|
+
description: Optional[_builtins.str] = None,
|
|
57
|
+
id: Optional[_builtins.str] = None,
|
|
58
|
+
ip: Optional[_builtins.str] = None,
|
|
59
|
+
scaleway_ranges: Optional[_builtins.bool] = None):
|
|
60
|
+
"""
|
|
61
|
+
:param _builtins.str description: A text describing this rule.
|
|
62
|
+
:param _builtins.str id: The ID of the ACL resource. It is the same as the ID of the cluster.
|
|
63
|
+
:param _builtins.str ip: The IP range to whitelist in [CIDR notation](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing#CIDR_notation)
|
|
64
|
+
|
|
65
|
+
> **Important:** If the `ip` field is set, `scaleway_ranges` cannot be set to true in the same rule.
|
|
66
|
+
:param _builtins.bool scaleway_ranges: Allow access to cluster from all Scaleway ranges as defined in [Scaleway Network Information - IP ranges used by Scaleway](https://www.scaleway.com/en/docs/console/account/reference-content/scaleway-network-information/#ip-ranges-used-by-scaleway).
|
|
67
|
+
Only one rule with this field set to true can be added.
|
|
68
|
+
|
|
69
|
+
> **Important:** If the `scaleway_ranges` field is set to true, the `ip` field cannot be set on the same rule.
|
|
70
|
+
"""
|
|
71
|
+
if description is not None:
|
|
72
|
+
pulumi.set(__self__, "description", description)
|
|
73
|
+
if id is not None:
|
|
74
|
+
pulumi.set(__self__, "id", id)
|
|
75
|
+
if ip is not None:
|
|
76
|
+
pulumi.set(__self__, "ip", ip)
|
|
77
|
+
if scaleway_ranges is not None:
|
|
78
|
+
pulumi.set(__self__, "scaleway_ranges", scaleway_ranges)
|
|
79
|
+
|
|
80
|
+
@_builtins.property
|
|
81
|
+
@pulumi.getter
|
|
82
|
+
def description(self) -> Optional[_builtins.str]:
|
|
83
|
+
"""
|
|
84
|
+
A text describing this rule.
|
|
85
|
+
"""
|
|
86
|
+
return pulumi.get(self, "description")
|
|
87
|
+
|
|
88
|
+
@_builtins.property
|
|
89
|
+
@pulumi.getter
|
|
90
|
+
def id(self) -> Optional[_builtins.str]:
|
|
91
|
+
"""
|
|
92
|
+
The ID of the ACL resource. It is the same as the ID of the cluster.
|
|
93
|
+
"""
|
|
94
|
+
return pulumi.get(self, "id")
|
|
95
|
+
|
|
96
|
+
@_builtins.property
|
|
97
|
+
@pulumi.getter
|
|
98
|
+
def ip(self) -> Optional[_builtins.str]:
|
|
99
|
+
"""
|
|
100
|
+
The IP range to whitelist in [CIDR notation](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing#CIDR_notation)
|
|
101
|
+
|
|
102
|
+
> **Important:** If the `ip` field is set, `scaleway_ranges` cannot be set to true in the same rule.
|
|
103
|
+
"""
|
|
104
|
+
return pulumi.get(self, "ip")
|
|
105
|
+
|
|
106
|
+
@_builtins.property
|
|
107
|
+
@pulumi.getter(name="scalewayRanges")
|
|
108
|
+
def scaleway_ranges(self) -> Optional[_builtins.bool]:
|
|
109
|
+
"""
|
|
110
|
+
Allow access to cluster from all Scaleway ranges as defined in [Scaleway Network Information - IP ranges used by Scaleway](https://www.scaleway.com/en/docs/console/account/reference-content/scaleway-network-information/#ip-ranges-used-by-scaleway).
|
|
111
|
+
Only one rule with this field set to true can be added.
|
|
112
|
+
|
|
113
|
+
> **Important:** If the `scaleway_ranges` field is set to true, the `ip` field cannot be set on the same rule.
|
|
114
|
+
"""
|
|
115
|
+
return pulumi.get(self, "scaleway_ranges")
|
|
116
|
+
|
|
117
|
+
|
|
32
118
|
@pulumi.output_type
|
|
33
119
|
class ClusterAutoUpgrade(dict):
|
|
34
120
|
@staticmethod
|
|
@@ -51,39 +137,39 @@ class ClusterAutoUpgrade(dict):
|
|
|
51
137
|
return super().get(key, default)
|
|
52
138
|
|
|
53
139
|
def __init__(__self__, *,
|
|
54
|
-
enable: bool,
|
|
55
|
-
maintenance_window_day: str,
|
|
56
|
-
maintenance_window_start_hour: int):
|
|
140
|
+
enable: _builtins.bool,
|
|
141
|
+
maintenance_window_day: _builtins.str,
|
|
142
|
+
maintenance_window_start_hour: _builtins.int):
|
|
57
143
|
"""
|
|
58
|
-
:param bool enable: Set to `true` to enable Kubernetes patch version auto upgrades.
|
|
144
|
+
:param _builtins.bool enable: Set to `true` to enable Kubernetes patch version auto upgrades.
|
|
59
145
|
> **Important:** When enabling auto upgrades, the `version` field take a minor version like x.y (ie 1.18).
|
|
60
|
-
:param str maintenance_window_day: The day of the auto upgrade maintenance window (`monday` to `sunday`, or `any`).
|
|
61
|
-
:param int maintenance_window_start_hour: The start hour (UTC) of the 2-hour auto upgrade maintenance window (0 to 23).
|
|
146
|
+
:param _builtins.str maintenance_window_day: The day of the auto upgrade maintenance window (`monday` to `sunday`, or `any`).
|
|
147
|
+
:param _builtins.int maintenance_window_start_hour: The start hour (UTC) of the 2-hour auto upgrade maintenance window (0 to 23).
|
|
62
148
|
"""
|
|
63
149
|
pulumi.set(__self__, "enable", enable)
|
|
64
150
|
pulumi.set(__self__, "maintenance_window_day", maintenance_window_day)
|
|
65
151
|
pulumi.set(__self__, "maintenance_window_start_hour", maintenance_window_start_hour)
|
|
66
152
|
|
|
67
|
-
@property
|
|
153
|
+
@_builtins.property
|
|
68
154
|
@pulumi.getter
|
|
69
|
-
def enable(self) -> bool:
|
|
155
|
+
def enable(self) -> _builtins.bool:
|
|
70
156
|
"""
|
|
71
157
|
Set to `true` to enable Kubernetes patch version auto upgrades.
|
|
72
158
|
> **Important:** When enabling auto upgrades, the `version` field take a minor version like x.y (ie 1.18).
|
|
73
159
|
"""
|
|
74
160
|
return pulumi.get(self, "enable")
|
|
75
161
|
|
|
76
|
-
@property
|
|
162
|
+
@_builtins.property
|
|
77
163
|
@pulumi.getter(name="maintenanceWindowDay")
|
|
78
|
-
def maintenance_window_day(self) -> str:
|
|
164
|
+
def maintenance_window_day(self) -> _builtins.str:
|
|
79
165
|
"""
|
|
80
166
|
The day of the auto upgrade maintenance window (`monday` to `sunday`, or `any`).
|
|
81
167
|
"""
|
|
82
168
|
return pulumi.get(self, "maintenance_window_day")
|
|
83
169
|
|
|
84
|
-
@property
|
|
170
|
+
@_builtins.property
|
|
85
171
|
@pulumi.getter(name="maintenanceWindowStartHour")
|
|
86
|
-
def maintenance_window_start_hour(self) -> int:
|
|
172
|
+
def maintenance_window_start_hour(self) -> _builtins.int:
|
|
87
173
|
"""
|
|
88
174
|
The start hour (UTC) of the 2-hour auto upgrade maintenance window (0 to 23).
|
|
89
175
|
"""
|
|
@@ -124,27 +210,27 @@ class ClusterAutoscalerConfig(dict):
|
|
|
124
210
|
return super().get(key, default)
|
|
125
211
|
|
|
126
212
|
def __init__(__self__, *,
|
|
127
|
-
balance_similar_node_groups: Optional[bool] = None,
|
|
128
|
-
disable_scale_down: Optional[bool] = None,
|
|
129
|
-
estimator: Optional[str] = None,
|
|
130
|
-
expander: Optional[str] = None,
|
|
131
|
-
expendable_pods_priority_cutoff: Optional[int] = None,
|
|
132
|
-
ignore_daemonsets_utilization: Optional[bool] = None,
|
|
133
|
-
max_graceful_termination_sec: Optional[int] = None,
|
|
134
|
-
scale_down_delay_after_add: Optional[str] = None,
|
|
135
|
-
scale_down_unneeded_time: Optional[str] = None,
|
|
136
|
-
scale_down_utilization_threshold: Optional[float] = None):
|
|
137
|
-
"""
|
|
138
|
-
:param bool balance_similar_node_groups: Detect similar node groups and balance the number of nodes between them.
|
|
139
|
-
:param bool disable_scale_down: Disables the scale down feature of the autoscaler.
|
|
140
|
-
:param str estimator: Type of resource estimator to be used in scale up.
|
|
141
|
-
:param str expander: Type of node group expander to be used in scale up.
|
|
142
|
-
:param int expendable_pods_priority_cutoff: Pods with priority below cutoff will be expendable. They can be killed without any consideration during scale down and they don't cause scale up. Pods with null priority (PodPriority disabled) are non expendable.
|
|
143
|
-
:param bool ignore_daemonsets_utilization: Ignore DaemonSet pods when calculating resource utilization for scaling down.
|
|
144
|
-
:param int max_graceful_termination_sec: Maximum number of seconds the cluster autoscaler waits for pod termination when trying to scale down a node
|
|
145
|
-
:param str scale_down_delay_after_add: How long after scale up that scale down evaluation resumes.
|
|
146
|
-
:param str scale_down_unneeded_time: How long a node should be unneeded before it is eligible for scale down.
|
|
147
|
-
:param float scale_down_utilization_threshold: Node utilization level, defined as sum of requested resources divided by capacity, below which a node can be considered for scale down
|
|
213
|
+
balance_similar_node_groups: Optional[_builtins.bool] = None,
|
|
214
|
+
disable_scale_down: Optional[_builtins.bool] = None,
|
|
215
|
+
estimator: Optional[_builtins.str] = None,
|
|
216
|
+
expander: Optional[_builtins.str] = None,
|
|
217
|
+
expendable_pods_priority_cutoff: Optional[_builtins.int] = None,
|
|
218
|
+
ignore_daemonsets_utilization: Optional[_builtins.bool] = None,
|
|
219
|
+
max_graceful_termination_sec: Optional[_builtins.int] = None,
|
|
220
|
+
scale_down_delay_after_add: Optional[_builtins.str] = None,
|
|
221
|
+
scale_down_unneeded_time: Optional[_builtins.str] = None,
|
|
222
|
+
scale_down_utilization_threshold: Optional[_builtins.float] = None):
|
|
223
|
+
"""
|
|
224
|
+
:param _builtins.bool balance_similar_node_groups: Detect similar node groups and balance the number of nodes between them.
|
|
225
|
+
:param _builtins.bool disable_scale_down: Disables the scale down feature of the autoscaler.
|
|
226
|
+
:param _builtins.str estimator: Type of resource estimator to be used in scale up.
|
|
227
|
+
:param _builtins.str expander: Type of node group expander to be used in scale up.
|
|
228
|
+
:param _builtins.int expendable_pods_priority_cutoff: Pods with priority below cutoff will be expendable. They can be killed without any consideration during scale down and they don't cause scale up. Pods with null priority (PodPriority disabled) are non expendable.
|
|
229
|
+
:param _builtins.bool ignore_daemonsets_utilization: Ignore DaemonSet pods when calculating resource utilization for scaling down.
|
|
230
|
+
:param _builtins.int max_graceful_termination_sec: Maximum number of seconds the cluster autoscaler waits for pod termination when trying to scale down a node
|
|
231
|
+
:param _builtins.str scale_down_delay_after_add: How long after scale up that scale down evaluation resumes.
|
|
232
|
+
:param _builtins.str scale_down_unneeded_time: How long a node should be unneeded before it is eligible for scale down.
|
|
233
|
+
:param _builtins.float scale_down_utilization_threshold: Node utilization level, defined as sum of requested resources divided by capacity, below which a node can be considered for scale down
|
|
148
234
|
"""
|
|
149
235
|
if balance_similar_node_groups is not None:
|
|
150
236
|
pulumi.set(__self__, "balance_similar_node_groups", balance_similar_node_groups)
|
|
@@ -167,81 +253,81 @@ class ClusterAutoscalerConfig(dict):
|
|
|
167
253
|
if scale_down_utilization_threshold is not None:
|
|
168
254
|
pulumi.set(__self__, "scale_down_utilization_threshold", scale_down_utilization_threshold)
|
|
169
255
|
|
|
170
|
-
@property
|
|
256
|
+
@_builtins.property
|
|
171
257
|
@pulumi.getter(name="balanceSimilarNodeGroups")
|
|
172
|
-
def balance_similar_node_groups(self) -> Optional[bool]:
|
|
258
|
+
def balance_similar_node_groups(self) -> Optional[_builtins.bool]:
|
|
173
259
|
"""
|
|
174
260
|
Detect similar node groups and balance the number of nodes between them.
|
|
175
261
|
"""
|
|
176
262
|
return pulumi.get(self, "balance_similar_node_groups")
|
|
177
263
|
|
|
178
|
-
@property
|
|
264
|
+
@_builtins.property
|
|
179
265
|
@pulumi.getter(name="disableScaleDown")
|
|
180
|
-
def disable_scale_down(self) -> Optional[bool]:
|
|
266
|
+
def disable_scale_down(self) -> Optional[_builtins.bool]:
|
|
181
267
|
"""
|
|
182
268
|
Disables the scale down feature of the autoscaler.
|
|
183
269
|
"""
|
|
184
270
|
return pulumi.get(self, "disable_scale_down")
|
|
185
271
|
|
|
186
|
-
@property
|
|
272
|
+
@_builtins.property
|
|
187
273
|
@pulumi.getter
|
|
188
|
-
def estimator(self) -> Optional[str]:
|
|
274
|
+
def estimator(self) -> Optional[_builtins.str]:
|
|
189
275
|
"""
|
|
190
276
|
Type of resource estimator to be used in scale up.
|
|
191
277
|
"""
|
|
192
278
|
return pulumi.get(self, "estimator")
|
|
193
279
|
|
|
194
|
-
@property
|
|
280
|
+
@_builtins.property
|
|
195
281
|
@pulumi.getter
|
|
196
|
-
def expander(self) -> Optional[str]:
|
|
282
|
+
def expander(self) -> Optional[_builtins.str]:
|
|
197
283
|
"""
|
|
198
284
|
Type of node group expander to be used in scale up.
|
|
199
285
|
"""
|
|
200
286
|
return pulumi.get(self, "expander")
|
|
201
287
|
|
|
202
|
-
@property
|
|
288
|
+
@_builtins.property
|
|
203
289
|
@pulumi.getter(name="expendablePodsPriorityCutoff")
|
|
204
|
-
def expendable_pods_priority_cutoff(self) -> Optional[int]:
|
|
290
|
+
def expendable_pods_priority_cutoff(self) -> Optional[_builtins.int]:
|
|
205
291
|
"""
|
|
206
292
|
Pods with priority below cutoff will be expendable. They can be killed without any consideration during scale down and they don't cause scale up. Pods with null priority (PodPriority disabled) are non expendable.
|
|
207
293
|
"""
|
|
208
294
|
return pulumi.get(self, "expendable_pods_priority_cutoff")
|
|
209
295
|
|
|
210
|
-
@property
|
|
296
|
+
@_builtins.property
|
|
211
297
|
@pulumi.getter(name="ignoreDaemonsetsUtilization")
|
|
212
|
-
def ignore_daemonsets_utilization(self) -> Optional[bool]:
|
|
298
|
+
def ignore_daemonsets_utilization(self) -> Optional[_builtins.bool]:
|
|
213
299
|
"""
|
|
214
300
|
Ignore DaemonSet pods when calculating resource utilization for scaling down.
|
|
215
301
|
"""
|
|
216
302
|
return pulumi.get(self, "ignore_daemonsets_utilization")
|
|
217
303
|
|
|
218
|
-
@property
|
|
304
|
+
@_builtins.property
|
|
219
305
|
@pulumi.getter(name="maxGracefulTerminationSec")
|
|
220
|
-
def max_graceful_termination_sec(self) -> Optional[int]:
|
|
306
|
+
def max_graceful_termination_sec(self) -> Optional[_builtins.int]:
|
|
221
307
|
"""
|
|
222
308
|
Maximum number of seconds the cluster autoscaler waits for pod termination when trying to scale down a node
|
|
223
309
|
"""
|
|
224
310
|
return pulumi.get(self, "max_graceful_termination_sec")
|
|
225
311
|
|
|
226
|
-
@property
|
|
312
|
+
@_builtins.property
|
|
227
313
|
@pulumi.getter(name="scaleDownDelayAfterAdd")
|
|
228
|
-
def scale_down_delay_after_add(self) -> Optional[str]:
|
|
314
|
+
def scale_down_delay_after_add(self) -> Optional[_builtins.str]:
|
|
229
315
|
"""
|
|
230
316
|
How long after scale up that scale down evaluation resumes.
|
|
231
317
|
"""
|
|
232
318
|
return pulumi.get(self, "scale_down_delay_after_add")
|
|
233
319
|
|
|
234
|
-
@property
|
|
320
|
+
@_builtins.property
|
|
235
321
|
@pulumi.getter(name="scaleDownUnneededTime")
|
|
236
|
-
def scale_down_unneeded_time(self) -> Optional[str]:
|
|
322
|
+
def scale_down_unneeded_time(self) -> Optional[_builtins.str]:
|
|
237
323
|
"""
|
|
238
324
|
How long a node should be unneeded before it is eligible for scale down.
|
|
239
325
|
"""
|
|
240
326
|
return pulumi.get(self, "scale_down_unneeded_time")
|
|
241
327
|
|
|
242
|
-
@property
|
|
328
|
+
@_builtins.property
|
|
243
329
|
@pulumi.getter(name="scaleDownUtilizationThreshold")
|
|
244
|
-
def scale_down_utilization_threshold(self) -> Optional[float]:
|
|
330
|
+
def scale_down_utilization_threshold(self) -> Optional[_builtins.float]:
|
|
245
331
|
"""
|
|
246
332
|
Node utilization level, defined as sum of requested resources divided by capacity, below which a node can be considered for scale down
|
|
247
333
|
"""
|
|
@@ -270,15 +356,15 @@ class ClusterKubeconfig(dict):
|
|
|
270
356
|
return super().get(key, default)
|
|
271
357
|
|
|
272
358
|
def __init__(__self__, *,
|
|
273
|
-
cluster_ca_certificate: Optional[str] = None,
|
|
274
|
-
config_file: Optional[str] = None,
|
|
275
|
-
host: Optional[str] = None,
|
|
276
|
-
token: Optional[str] = None):
|
|
359
|
+
cluster_ca_certificate: Optional[_builtins.str] = None,
|
|
360
|
+
config_file: Optional[_builtins.str] = None,
|
|
361
|
+
host: Optional[_builtins.str] = None,
|
|
362
|
+
token: Optional[_builtins.str] = None):
|
|
277
363
|
"""
|
|
278
|
-
:param str cluster_ca_certificate: The CA certificate of the Kubernetes API server.
|
|
279
|
-
:param str config_file: The raw kubeconfig file.
|
|
280
|
-
:param str host: The URL of the Kubernetes API server.
|
|
281
|
-
:param str token: The token to connect to the Kubernetes API server.
|
|
364
|
+
:param _builtins.str cluster_ca_certificate: The CA certificate of the Kubernetes API server.
|
|
365
|
+
:param _builtins.str config_file: The raw kubeconfig file.
|
|
366
|
+
:param _builtins.str host: The URL of the Kubernetes API server.
|
|
367
|
+
:param _builtins.str token: The token to connect to the Kubernetes API server.
|
|
282
368
|
"""
|
|
283
369
|
if cluster_ca_certificate is not None:
|
|
284
370
|
pulumi.set(__self__, "cluster_ca_certificate", cluster_ca_certificate)
|
|
@@ -289,33 +375,33 @@ class ClusterKubeconfig(dict):
|
|
|
289
375
|
if token is not None:
|
|
290
376
|
pulumi.set(__self__, "token", token)
|
|
291
377
|
|
|
292
|
-
@property
|
|
378
|
+
@_builtins.property
|
|
293
379
|
@pulumi.getter(name="clusterCaCertificate")
|
|
294
|
-
def cluster_ca_certificate(self) -> Optional[str]:
|
|
380
|
+
def cluster_ca_certificate(self) -> Optional[_builtins.str]:
|
|
295
381
|
"""
|
|
296
382
|
The CA certificate of the Kubernetes API server.
|
|
297
383
|
"""
|
|
298
384
|
return pulumi.get(self, "cluster_ca_certificate")
|
|
299
385
|
|
|
300
|
-
@property
|
|
386
|
+
@_builtins.property
|
|
301
387
|
@pulumi.getter(name="configFile")
|
|
302
|
-
def config_file(self) -> Optional[str]:
|
|
388
|
+
def config_file(self) -> Optional[_builtins.str]:
|
|
303
389
|
"""
|
|
304
390
|
The raw kubeconfig file.
|
|
305
391
|
"""
|
|
306
392
|
return pulumi.get(self, "config_file")
|
|
307
393
|
|
|
308
|
-
@property
|
|
394
|
+
@_builtins.property
|
|
309
395
|
@pulumi.getter
|
|
310
|
-
def host(self) -> Optional[str]:
|
|
396
|
+
def host(self) -> Optional[_builtins.str]:
|
|
311
397
|
"""
|
|
312
398
|
The URL of the Kubernetes API server.
|
|
313
399
|
"""
|
|
314
400
|
return pulumi.get(self, "host")
|
|
315
401
|
|
|
316
|
-
@property
|
|
402
|
+
@_builtins.property
|
|
317
403
|
@pulumi.getter
|
|
318
|
-
def token(self) -> Optional[str]:
|
|
404
|
+
def token(self) -> Optional[_builtins.str]:
|
|
319
405
|
"""
|
|
320
406
|
The token to connect to the Kubernetes API server.
|
|
321
407
|
"""
|
|
@@ -354,21 +440,21 @@ class ClusterOpenIdConnectConfig(dict):
|
|
|
354
440
|
return super().get(key, default)
|
|
355
441
|
|
|
356
442
|
def __init__(__self__, *,
|
|
357
|
-
client_id: str,
|
|
358
|
-
issuer_url: str,
|
|
359
|
-
groups_claims: Optional[Sequence[str]] = None,
|
|
360
|
-
groups_prefix: Optional[str] = None,
|
|
361
|
-
required_claims: Optional[Sequence[str]] = None,
|
|
362
|
-
username_claim: Optional[str] = None,
|
|
363
|
-
username_prefix: Optional[str] = None):
|
|
364
|
-
"""
|
|
365
|
-
:param str client_id: A client id that all tokens must be issued for
|
|
366
|
-
:param str issuer_url: URL of the provider which allows the API server to discover public signing keys
|
|
367
|
-
:param Sequence[str] groups_claims: JWT claim to use as the user's group
|
|
368
|
-
:param str groups_prefix: Prefix prepended to group claims
|
|
369
|
-
:param Sequence[str] required_claims: Multiple key=value pairs that describes a required claim in the ID Token
|
|
370
|
-
:param str username_claim: JWT claim to use as the user name
|
|
371
|
-
:param str username_prefix: Prefix prepended to username
|
|
443
|
+
client_id: _builtins.str,
|
|
444
|
+
issuer_url: _builtins.str,
|
|
445
|
+
groups_claims: Optional[Sequence[_builtins.str]] = None,
|
|
446
|
+
groups_prefix: Optional[_builtins.str] = None,
|
|
447
|
+
required_claims: Optional[Sequence[_builtins.str]] = None,
|
|
448
|
+
username_claim: Optional[_builtins.str] = None,
|
|
449
|
+
username_prefix: Optional[_builtins.str] = None):
|
|
450
|
+
"""
|
|
451
|
+
:param _builtins.str client_id: A client id that all tokens must be issued for
|
|
452
|
+
:param _builtins.str issuer_url: URL of the provider which allows the API server to discover public signing keys
|
|
453
|
+
:param Sequence[_builtins.str] groups_claims: JWT claim to use as the user's group
|
|
454
|
+
:param _builtins.str groups_prefix: Prefix prepended to group claims
|
|
455
|
+
:param Sequence[_builtins.str] required_claims: Multiple key=value pairs that describes a required claim in the ID Token
|
|
456
|
+
:param _builtins.str username_claim: JWT claim to use as the user name
|
|
457
|
+
:param _builtins.str username_prefix: Prefix prepended to username
|
|
372
458
|
"""
|
|
373
459
|
pulumi.set(__self__, "client_id", client_id)
|
|
374
460
|
pulumi.set(__self__, "issuer_url", issuer_url)
|
|
@@ -383,57 +469,57 @@ class ClusterOpenIdConnectConfig(dict):
|
|
|
383
469
|
if username_prefix is not None:
|
|
384
470
|
pulumi.set(__self__, "username_prefix", username_prefix)
|
|
385
471
|
|
|
386
|
-
@property
|
|
472
|
+
@_builtins.property
|
|
387
473
|
@pulumi.getter(name="clientId")
|
|
388
|
-
def client_id(self) -> str:
|
|
474
|
+
def client_id(self) -> _builtins.str:
|
|
389
475
|
"""
|
|
390
476
|
A client id that all tokens must be issued for
|
|
391
477
|
"""
|
|
392
478
|
return pulumi.get(self, "client_id")
|
|
393
479
|
|
|
394
|
-
@property
|
|
480
|
+
@_builtins.property
|
|
395
481
|
@pulumi.getter(name="issuerUrl")
|
|
396
|
-
def issuer_url(self) -> str:
|
|
482
|
+
def issuer_url(self) -> _builtins.str:
|
|
397
483
|
"""
|
|
398
484
|
URL of the provider which allows the API server to discover public signing keys
|
|
399
485
|
"""
|
|
400
486
|
return pulumi.get(self, "issuer_url")
|
|
401
487
|
|
|
402
|
-
@property
|
|
488
|
+
@_builtins.property
|
|
403
489
|
@pulumi.getter(name="groupsClaims")
|
|
404
|
-
def groups_claims(self) -> Optional[Sequence[str]]:
|
|
490
|
+
def groups_claims(self) -> Optional[Sequence[_builtins.str]]:
|
|
405
491
|
"""
|
|
406
492
|
JWT claim to use as the user's group
|
|
407
493
|
"""
|
|
408
494
|
return pulumi.get(self, "groups_claims")
|
|
409
495
|
|
|
410
|
-
@property
|
|
496
|
+
@_builtins.property
|
|
411
497
|
@pulumi.getter(name="groupsPrefix")
|
|
412
|
-
def groups_prefix(self) -> Optional[str]:
|
|
498
|
+
def groups_prefix(self) -> Optional[_builtins.str]:
|
|
413
499
|
"""
|
|
414
500
|
Prefix prepended to group claims
|
|
415
501
|
"""
|
|
416
502
|
return pulumi.get(self, "groups_prefix")
|
|
417
503
|
|
|
418
|
-
@property
|
|
504
|
+
@_builtins.property
|
|
419
505
|
@pulumi.getter(name="requiredClaims")
|
|
420
|
-
def required_claims(self) -> Optional[Sequence[str]]:
|
|
506
|
+
def required_claims(self) -> Optional[Sequence[_builtins.str]]:
|
|
421
507
|
"""
|
|
422
508
|
Multiple key=value pairs that describes a required claim in the ID Token
|
|
423
509
|
"""
|
|
424
510
|
return pulumi.get(self, "required_claims")
|
|
425
511
|
|
|
426
|
-
@property
|
|
512
|
+
@_builtins.property
|
|
427
513
|
@pulumi.getter(name="usernameClaim")
|
|
428
|
-
def username_claim(self) -> Optional[str]:
|
|
514
|
+
def username_claim(self) -> Optional[_builtins.str]:
|
|
429
515
|
"""
|
|
430
516
|
JWT claim to use as the user name
|
|
431
517
|
"""
|
|
432
518
|
return pulumi.get(self, "username_claim")
|
|
433
519
|
|
|
434
|
-
@property
|
|
520
|
+
@_builtins.property
|
|
435
521
|
@pulumi.getter(name="usernamePrefix")
|
|
436
|
-
def username_prefix(self) -> Optional[str]:
|
|
522
|
+
def username_prefix(self) -> Optional[_builtins.str]:
|
|
437
523
|
"""
|
|
438
524
|
Prefix prepended to username
|
|
439
525
|
"""
|
|
@@ -445,7 +531,9 @@ class PoolNode(dict):
|
|
|
445
531
|
@staticmethod
|
|
446
532
|
def __key_warning(key: str):
|
|
447
533
|
suggest = None
|
|
448
|
-
if key == "
|
|
534
|
+
if key == "privateIps":
|
|
535
|
+
suggest = "private_ips"
|
|
536
|
+
elif key == "publicIp":
|
|
449
537
|
suggest = "public_ip"
|
|
450
538
|
elif key == "publicIpV6":
|
|
451
539
|
suggest = "public_ip_v6"
|
|
@@ -462,20 +550,28 @@ class PoolNode(dict):
|
|
|
462
550
|
return super().get(key, default)
|
|
463
551
|
|
|
464
552
|
def __init__(__self__, *,
|
|
465
|
-
|
|
466
|
-
|
|
467
|
-
|
|
468
|
-
|
|
469
|
-
|
|
470
|
-
|
|
553
|
+
id: Optional[_builtins.str] = None,
|
|
554
|
+
name: Optional[_builtins.str] = None,
|
|
555
|
+
private_ips: Optional[Sequence['outputs.PoolNodePrivateIp']] = None,
|
|
556
|
+
public_ip: Optional[_builtins.str] = None,
|
|
557
|
+
public_ip_v6: Optional[_builtins.str] = None,
|
|
558
|
+
status: Optional[_builtins.str] = None):
|
|
559
|
+
"""
|
|
560
|
+
:param _builtins.str id: The ID of the IP address resource.
|
|
561
|
+
:param _builtins.str name: The name for the pool.
|
|
471
562
|
|
|
472
563
|
> **Important:** Updates to this field will recreate a new resource.
|
|
473
|
-
:param
|
|
474
|
-
:param str
|
|
475
|
-
:param str
|
|
564
|
+
:param Sequence['PoolNodePrivateIpArgs'] private_ips: The list of private IPv4 and IPv6 addresses associated with the node.
|
|
565
|
+
:param _builtins.str public_ip: The public IPv4. (Deprecated, Please use the official Kubernetes provider and the kubernetes_nodes data source)
|
|
566
|
+
:param _builtins.str public_ip_v6: The public IPv6. (Deprecated, Please use the official Kubernetes provider and the kubernetes_nodes data source)
|
|
567
|
+
:param _builtins.str status: The status of the node.
|
|
476
568
|
"""
|
|
569
|
+
if id is not None:
|
|
570
|
+
pulumi.set(__self__, "id", id)
|
|
477
571
|
if name is not None:
|
|
478
572
|
pulumi.set(__self__, "name", name)
|
|
573
|
+
if private_ips is not None:
|
|
574
|
+
pulumi.set(__self__, "private_ips", private_ips)
|
|
479
575
|
if public_ip is not None:
|
|
480
576
|
pulumi.set(__self__, "public_ip", public_ip)
|
|
481
577
|
if public_ip_v6 is not None:
|
|
@@ -483,9 +579,17 @@ class PoolNode(dict):
|
|
|
483
579
|
if status is not None:
|
|
484
580
|
pulumi.set(__self__, "status", status)
|
|
485
581
|
|
|
486
|
-
@property
|
|
582
|
+
@_builtins.property
|
|
487
583
|
@pulumi.getter
|
|
488
|
-
def
|
|
584
|
+
def id(self) -> Optional[_builtins.str]:
|
|
585
|
+
"""
|
|
586
|
+
The ID of the IP address resource.
|
|
587
|
+
"""
|
|
588
|
+
return pulumi.get(self, "id")
|
|
589
|
+
|
|
590
|
+
@_builtins.property
|
|
591
|
+
@pulumi.getter
|
|
592
|
+
def name(self) -> Optional[_builtins.str]:
|
|
489
593
|
"""
|
|
490
594
|
The name for the pool.
|
|
491
595
|
|
|
@@ -493,33 +597,72 @@ class PoolNode(dict):
|
|
|
493
597
|
"""
|
|
494
598
|
return pulumi.get(self, "name")
|
|
495
599
|
|
|
496
|
-
@property
|
|
600
|
+
@_builtins.property
|
|
601
|
+
@pulumi.getter(name="privateIps")
|
|
602
|
+
def private_ips(self) -> Optional[Sequence['outputs.PoolNodePrivateIp']]:
|
|
603
|
+
"""
|
|
604
|
+
The list of private IPv4 and IPv6 addresses associated with the node.
|
|
605
|
+
"""
|
|
606
|
+
return pulumi.get(self, "private_ips")
|
|
607
|
+
|
|
608
|
+
@_builtins.property
|
|
497
609
|
@pulumi.getter(name="publicIp")
|
|
498
610
|
@_utilities.deprecated("""Please use the official Kubernetes provider and the kubernetes_nodes data source""")
|
|
499
|
-
def public_ip(self) -> Optional[str]:
|
|
611
|
+
def public_ip(self) -> Optional[_builtins.str]:
|
|
500
612
|
"""
|
|
501
613
|
The public IPv4. (Deprecated, Please use the official Kubernetes provider and the kubernetes_nodes data source)
|
|
502
614
|
"""
|
|
503
615
|
return pulumi.get(self, "public_ip")
|
|
504
616
|
|
|
505
|
-
@property
|
|
617
|
+
@_builtins.property
|
|
506
618
|
@pulumi.getter(name="publicIpV6")
|
|
507
619
|
@_utilities.deprecated("""Please use the official Kubernetes provider and the kubernetes_nodes data source""")
|
|
508
|
-
def public_ip_v6(self) -> Optional[str]:
|
|
620
|
+
def public_ip_v6(self) -> Optional[_builtins.str]:
|
|
509
621
|
"""
|
|
510
622
|
The public IPv6. (Deprecated, Please use the official Kubernetes provider and the kubernetes_nodes data source)
|
|
511
623
|
"""
|
|
512
624
|
return pulumi.get(self, "public_ip_v6")
|
|
513
625
|
|
|
514
|
-
@property
|
|
626
|
+
@_builtins.property
|
|
515
627
|
@pulumi.getter
|
|
516
|
-
def status(self) -> Optional[str]:
|
|
628
|
+
def status(self) -> Optional[_builtins.str]:
|
|
517
629
|
"""
|
|
518
630
|
The status of the node.
|
|
519
631
|
"""
|
|
520
632
|
return pulumi.get(self, "status")
|
|
521
633
|
|
|
522
634
|
|
|
635
|
+
@pulumi.output_type
|
|
636
|
+
class PoolNodePrivateIp(dict):
|
|
637
|
+
def __init__(__self__, *,
|
|
638
|
+
address: Optional[_builtins.str] = None,
|
|
639
|
+
id: Optional[_builtins.str] = None):
|
|
640
|
+
"""
|
|
641
|
+
:param _builtins.str address: The private IP address.
|
|
642
|
+
:param _builtins.str id: The ID of the IP address resource.
|
|
643
|
+
"""
|
|
644
|
+
if address is not None:
|
|
645
|
+
pulumi.set(__self__, "address", address)
|
|
646
|
+
if id is not None:
|
|
647
|
+
pulumi.set(__self__, "id", id)
|
|
648
|
+
|
|
649
|
+
@_builtins.property
|
|
650
|
+
@pulumi.getter
|
|
651
|
+
def address(self) -> Optional[_builtins.str]:
|
|
652
|
+
"""
|
|
653
|
+
The private IP address.
|
|
654
|
+
"""
|
|
655
|
+
return pulumi.get(self, "address")
|
|
656
|
+
|
|
657
|
+
@_builtins.property
|
|
658
|
+
@pulumi.getter
|
|
659
|
+
def id(self) -> Optional[_builtins.str]:
|
|
660
|
+
"""
|
|
661
|
+
The ID of the IP address resource.
|
|
662
|
+
"""
|
|
663
|
+
return pulumi.get(self, "id")
|
|
664
|
+
|
|
665
|
+
|
|
523
666
|
@pulumi.output_type
|
|
524
667
|
class PoolUpgradePolicy(dict):
|
|
525
668
|
@staticmethod
|
|
@@ -542,28 +685,28 @@ class PoolUpgradePolicy(dict):
|
|
|
542
685
|
return super().get(key, default)
|
|
543
686
|
|
|
544
687
|
def __init__(__self__, *,
|
|
545
|
-
max_surge: Optional[int] = None,
|
|
546
|
-
max_unavailable: Optional[int] = None):
|
|
688
|
+
max_surge: Optional[_builtins.int] = None,
|
|
689
|
+
max_unavailable: Optional[_builtins.int] = None):
|
|
547
690
|
"""
|
|
548
|
-
:param int max_surge: The maximum number of nodes to be created during the upgrade
|
|
549
|
-
:param int max_unavailable: The maximum number of nodes that can be not ready at the same time
|
|
691
|
+
:param _builtins.int max_surge: The maximum number of nodes to be created during the upgrade
|
|
692
|
+
:param _builtins.int max_unavailable: The maximum number of nodes that can be not ready at the same time
|
|
550
693
|
"""
|
|
551
694
|
if max_surge is not None:
|
|
552
695
|
pulumi.set(__self__, "max_surge", max_surge)
|
|
553
696
|
if max_unavailable is not None:
|
|
554
697
|
pulumi.set(__self__, "max_unavailable", max_unavailable)
|
|
555
698
|
|
|
556
|
-
@property
|
|
699
|
+
@_builtins.property
|
|
557
700
|
@pulumi.getter(name="maxSurge")
|
|
558
|
-
def max_surge(self) -> Optional[int]:
|
|
701
|
+
def max_surge(self) -> Optional[_builtins.int]:
|
|
559
702
|
"""
|
|
560
703
|
The maximum number of nodes to be created during the upgrade
|
|
561
704
|
"""
|
|
562
705
|
return pulumi.get(self, "max_surge")
|
|
563
706
|
|
|
564
|
-
@property
|
|
707
|
+
@_builtins.property
|
|
565
708
|
@pulumi.getter(name="maxUnavailable")
|
|
566
|
-
def max_unavailable(self) -> Optional[int]:
|
|
709
|
+
def max_unavailable(self) -> Optional[_builtins.int]:
|
|
567
710
|
"""
|
|
568
711
|
The maximum number of nodes that can be not ready at the same time
|
|
569
712
|
"""
|
|
@@ -573,37 +716,37 @@ class PoolUpgradePolicy(dict):
|
|
|
573
716
|
@pulumi.output_type
|
|
574
717
|
class GetClusterAutoUpgradeResult(dict):
|
|
575
718
|
def __init__(__self__, *,
|
|
576
|
-
enable: bool,
|
|
577
|
-
maintenance_window_day: str,
|
|
578
|
-
maintenance_window_start_hour: int):
|
|
719
|
+
enable: _builtins.bool,
|
|
720
|
+
maintenance_window_day: _builtins.str,
|
|
721
|
+
maintenance_window_start_hour: _builtins.int):
|
|
579
722
|
"""
|
|
580
|
-
:param bool enable: True if Kubernetes patch version auto upgrades is enabled.
|
|
581
|
-
:param str maintenance_window_day: The day of the auto upgrade maintenance window (`monday` to `sunday`, or `any`).
|
|
582
|
-
:param int maintenance_window_start_hour: The start hour (UTC) of the 2-hour auto upgrade maintenance window (0 to 23).
|
|
723
|
+
:param _builtins.bool enable: True if Kubernetes patch version auto upgrades is enabled.
|
|
724
|
+
:param _builtins.str maintenance_window_day: The day of the auto upgrade maintenance window (`monday` to `sunday`, or `any`).
|
|
725
|
+
:param _builtins.int maintenance_window_start_hour: The start hour (UTC) of the 2-hour auto upgrade maintenance window (0 to 23).
|
|
583
726
|
"""
|
|
584
727
|
pulumi.set(__self__, "enable", enable)
|
|
585
728
|
pulumi.set(__self__, "maintenance_window_day", maintenance_window_day)
|
|
586
729
|
pulumi.set(__self__, "maintenance_window_start_hour", maintenance_window_start_hour)
|
|
587
730
|
|
|
588
|
-
@property
|
|
731
|
+
@_builtins.property
|
|
589
732
|
@pulumi.getter
|
|
590
|
-
def enable(self) -> bool:
|
|
733
|
+
def enable(self) -> _builtins.bool:
|
|
591
734
|
"""
|
|
592
735
|
True if Kubernetes patch version auto upgrades is enabled.
|
|
593
736
|
"""
|
|
594
737
|
return pulumi.get(self, "enable")
|
|
595
738
|
|
|
596
|
-
@property
|
|
739
|
+
@_builtins.property
|
|
597
740
|
@pulumi.getter(name="maintenanceWindowDay")
|
|
598
|
-
def maintenance_window_day(self) -> str:
|
|
741
|
+
def maintenance_window_day(self) -> _builtins.str:
|
|
599
742
|
"""
|
|
600
743
|
The day of the auto upgrade maintenance window (`monday` to `sunday`, or `any`).
|
|
601
744
|
"""
|
|
602
745
|
return pulumi.get(self, "maintenance_window_day")
|
|
603
746
|
|
|
604
|
-
@property
|
|
747
|
+
@_builtins.property
|
|
605
748
|
@pulumi.getter(name="maintenanceWindowStartHour")
|
|
606
|
-
def maintenance_window_start_hour(self) -> int:
|
|
749
|
+
def maintenance_window_start_hour(self) -> _builtins.int:
|
|
607
750
|
"""
|
|
608
751
|
The start hour (UTC) of the 2-hour auto upgrade maintenance window (0 to 23).
|
|
609
752
|
"""
|
|
@@ -613,27 +756,27 @@ class GetClusterAutoUpgradeResult(dict):
|
|
|
613
756
|
@pulumi.output_type
|
|
614
757
|
class GetClusterAutoscalerConfigResult(dict):
|
|
615
758
|
def __init__(__self__, *,
|
|
616
|
-
balance_similar_node_groups: bool,
|
|
617
|
-
disable_scale_down: bool,
|
|
618
|
-
estimator: str,
|
|
619
|
-
expander: str,
|
|
620
|
-
expendable_pods_priority_cutoff: int,
|
|
621
|
-
ignore_daemonsets_utilization: bool,
|
|
622
|
-
max_graceful_termination_sec: int,
|
|
623
|
-
scale_down_delay_after_add: str,
|
|
624
|
-
scale_down_unneeded_time: str,
|
|
625
|
-
scale_down_utilization_threshold: float):
|
|
626
|
-
"""
|
|
627
|
-
:param bool balance_similar_node_groups: True if detecting similar node groups and balance the number of nodes between them is enabled.
|
|
628
|
-
:param bool disable_scale_down: True if the scale down feature of the autoscaler is disabled.
|
|
629
|
-
:param str estimator: The type of resource estimator used in scale up.
|
|
630
|
-
:param str expander: The type of node group expander be used in scale up.
|
|
631
|
-
:param int expendable_pods_priority_cutoff: Pods with priority below cutoff will be expendable. They can be killed without any consideration during scale down and they don't cause scale up. Pods with null priority (PodPriority disabled) are non expendable.
|
|
632
|
-
:param bool ignore_daemonsets_utilization: True if ignoring DaemonSet pods when calculating resource utilization for scaling down is enabled.
|
|
633
|
-
:param int max_graceful_termination_sec: Maximum number of seconds the cluster autoscaler waits for pod termination when trying to scale down a node
|
|
634
|
-
:param str scale_down_delay_after_add: The duration after scale up that scale down evaluation resumes.
|
|
635
|
-
:param str scale_down_unneeded_time: The duration a node should be unneeded before it is eligible for scale down.
|
|
636
|
-
:param float scale_down_utilization_threshold: Node utilization level, defined as sum of requested resources divided by capacity, below which a node can be considered for scale down
|
|
759
|
+
balance_similar_node_groups: _builtins.bool,
|
|
760
|
+
disable_scale_down: _builtins.bool,
|
|
761
|
+
estimator: _builtins.str,
|
|
762
|
+
expander: _builtins.str,
|
|
763
|
+
expendable_pods_priority_cutoff: _builtins.int,
|
|
764
|
+
ignore_daemonsets_utilization: _builtins.bool,
|
|
765
|
+
max_graceful_termination_sec: _builtins.int,
|
|
766
|
+
scale_down_delay_after_add: _builtins.str,
|
|
767
|
+
scale_down_unneeded_time: _builtins.str,
|
|
768
|
+
scale_down_utilization_threshold: _builtins.float):
|
|
769
|
+
"""
|
|
770
|
+
:param _builtins.bool balance_similar_node_groups: True if detecting similar node groups and balance the number of nodes between them is enabled.
|
|
771
|
+
:param _builtins.bool disable_scale_down: True if the scale down feature of the autoscaler is disabled.
|
|
772
|
+
:param _builtins.str estimator: The type of resource estimator used in scale up.
|
|
773
|
+
:param _builtins.str expander: The type of node group expander be used in scale up.
|
|
774
|
+
:param _builtins.int expendable_pods_priority_cutoff: Pods with priority below cutoff will be expendable. They can be killed without any consideration during scale down and they don't cause scale up. Pods with null priority (PodPriority disabled) are non expendable.
|
|
775
|
+
:param _builtins.bool ignore_daemonsets_utilization: True if ignoring DaemonSet pods when calculating resource utilization for scaling down is enabled.
|
|
776
|
+
:param _builtins.int max_graceful_termination_sec: Maximum number of seconds the cluster autoscaler waits for pod termination when trying to scale down a node
|
|
777
|
+
:param _builtins.str scale_down_delay_after_add: The duration after scale up that scale down evaluation resumes.
|
|
778
|
+
:param _builtins.str scale_down_unneeded_time: The duration a node should be unneeded before it is eligible for scale down.
|
|
779
|
+
:param _builtins.float scale_down_utilization_threshold: Node utilization level, defined as sum of requested resources divided by capacity, below which a node can be considered for scale down
|
|
637
780
|
"""
|
|
638
781
|
pulumi.set(__self__, "balance_similar_node_groups", balance_similar_node_groups)
|
|
639
782
|
pulumi.set(__self__, "disable_scale_down", disable_scale_down)
|
|
@@ -646,81 +789,81 @@ class GetClusterAutoscalerConfigResult(dict):
|
|
|
646
789
|
pulumi.set(__self__, "scale_down_unneeded_time", scale_down_unneeded_time)
|
|
647
790
|
pulumi.set(__self__, "scale_down_utilization_threshold", scale_down_utilization_threshold)
|
|
648
791
|
|
|
649
|
-
@property
|
|
792
|
+
@_builtins.property
|
|
650
793
|
@pulumi.getter(name="balanceSimilarNodeGroups")
|
|
651
|
-
def balance_similar_node_groups(self) -> bool:
|
|
794
|
+
def balance_similar_node_groups(self) -> _builtins.bool:
|
|
652
795
|
"""
|
|
653
796
|
True if detecting similar node groups and balance the number of nodes between them is enabled.
|
|
654
797
|
"""
|
|
655
798
|
return pulumi.get(self, "balance_similar_node_groups")
|
|
656
799
|
|
|
657
|
-
@property
|
|
800
|
+
@_builtins.property
|
|
658
801
|
@pulumi.getter(name="disableScaleDown")
|
|
659
|
-
def disable_scale_down(self) -> bool:
|
|
802
|
+
def disable_scale_down(self) -> _builtins.bool:
|
|
660
803
|
"""
|
|
661
804
|
True if the scale down feature of the autoscaler is disabled.
|
|
662
805
|
"""
|
|
663
806
|
return pulumi.get(self, "disable_scale_down")
|
|
664
807
|
|
|
665
|
-
@property
|
|
808
|
+
@_builtins.property
|
|
666
809
|
@pulumi.getter
|
|
667
|
-
def estimator(self) -> str:
|
|
810
|
+
def estimator(self) -> _builtins.str:
|
|
668
811
|
"""
|
|
669
812
|
The type of resource estimator used in scale up.
|
|
670
813
|
"""
|
|
671
814
|
return pulumi.get(self, "estimator")
|
|
672
815
|
|
|
673
|
-
@property
|
|
816
|
+
@_builtins.property
|
|
674
817
|
@pulumi.getter
|
|
675
|
-
def expander(self) -> str:
|
|
818
|
+
def expander(self) -> _builtins.str:
|
|
676
819
|
"""
|
|
677
820
|
The type of node group expander be used in scale up.
|
|
678
821
|
"""
|
|
679
822
|
return pulumi.get(self, "expander")
|
|
680
823
|
|
|
681
|
-
@property
|
|
824
|
+
@_builtins.property
|
|
682
825
|
@pulumi.getter(name="expendablePodsPriorityCutoff")
|
|
683
|
-
def expendable_pods_priority_cutoff(self) -> int:
|
|
826
|
+
def expendable_pods_priority_cutoff(self) -> _builtins.int:
|
|
684
827
|
"""
|
|
685
828
|
Pods with priority below cutoff will be expendable. They can be killed without any consideration during scale down and they don't cause scale up. Pods with null priority (PodPriority disabled) are non expendable.
|
|
686
829
|
"""
|
|
687
830
|
return pulumi.get(self, "expendable_pods_priority_cutoff")
|
|
688
831
|
|
|
689
|
-
@property
|
|
832
|
+
@_builtins.property
|
|
690
833
|
@pulumi.getter(name="ignoreDaemonsetsUtilization")
|
|
691
|
-
def ignore_daemonsets_utilization(self) -> bool:
|
|
834
|
+
def ignore_daemonsets_utilization(self) -> _builtins.bool:
|
|
692
835
|
"""
|
|
693
836
|
True if ignoring DaemonSet pods when calculating resource utilization for scaling down is enabled.
|
|
694
837
|
"""
|
|
695
838
|
return pulumi.get(self, "ignore_daemonsets_utilization")
|
|
696
839
|
|
|
697
|
-
@property
|
|
840
|
+
@_builtins.property
|
|
698
841
|
@pulumi.getter(name="maxGracefulTerminationSec")
|
|
699
|
-
def max_graceful_termination_sec(self) -> int:
|
|
842
|
+
def max_graceful_termination_sec(self) -> _builtins.int:
|
|
700
843
|
"""
|
|
701
844
|
Maximum number of seconds the cluster autoscaler waits for pod termination when trying to scale down a node
|
|
702
845
|
"""
|
|
703
846
|
return pulumi.get(self, "max_graceful_termination_sec")
|
|
704
847
|
|
|
705
|
-
@property
|
|
848
|
+
@_builtins.property
|
|
706
849
|
@pulumi.getter(name="scaleDownDelayAfterAdd")
|
|
707
|
-
def scale_down_delay_after_add(self) -> str:
|
|
850
|
+
def scale_down_delay_after_add(self) -> _builtins.str:
|
|
708
851
|
"""
|
|
709
852
|
The duration after scale up that scale down evaluation resumes.
|
|
710
853
|
"""
|
|
711
854
|
return pulumi.get(self, "scale_down_delay_after_add")
|
|
712
855
|
|
|
713
|
-
@property
|
|
856
|
+
@_builtins.property
|
|
714
857
|
@pulumi.getter(name="scaleDownUnneededTime")
|
|
715
|
-
def scale_down_unneeded_time(self) -> str:
|
|
858
|
+
def scale_down_unneeded_time(self) -> _builtins.str:
|
|
716
859
|
"""
|
|
717
860
|
The duration a node should be unneeded before it is eligible for scale down.
|
|
718
861
|
"""
|
|
719
862
|
return pulumi.get(self, "scale_down_unneeded_time")
|
|
720
863
|
|
|
721
|
-
@property
|
|
864
|
+
@_builtins.property
|
|
722
865
|
@pulumi.getter(name="scaleDownUtilizationThreshold")
|
|
723
|
-
def scale_down_utilization_threshold(self) -> float:
|
|
866
|
+
def scale_down_utilization_threshold(self) -> _builtins.float:
|
|
724
867
|
"""
|
|
725
868
|
Node utilization level, defined as sum of requested resources divided by capacity, below which a node can be considered for scale down
|
|
726
869
|
"""
|
|
@@ -730,48 +873,48 @@ class GetClusterAutoscalerConfigResult(dict):
|
|
|
730
873
|
@pulumi.output_type
|
|
731
874
|
class GetClusterKubeconfigResult(dict):
|
|
732
875
|
def __init__(__self__, *,
|
|
733
|
-
cluster_ca_certificate: str,
|
|
734
|
-
config_file: str,
|
|
735
|
-
host: str,
|
|
736
|
-
token: str):
|
|
876
|
+
cluster_ca_certificate: _builtins.str,
|
|
877
|
+
config_file: _builtins.str,
|
|
878
|
+
host: _builtins.str,
|
|
879
|
+
token: _builtins.str):
|
|
737
880
|
"""
|
|
738
|
-
:param str cluster_ca_certificate: The CA certificate of the Kubernetes API server.
|
|
739
|
-
:param str config_file: The raw kubeconfig file.
|
|
740
|
-
:param str host: The URL of the Kubernetes API server.
|
|
741
|
-
:param str token: The token to connect to the Kubernetes API server.
|
|
881
|
+
:param _builtins.str cluster_ca_certificate: The CA certificate of the Kubernetes API server.
|
|
882
|
+
:param _builtins.str config_file: The raw kubeconfig file.
|
|
883
|
+
:param _builtins.str host: The URL of the Kubernetes API server.
|
|
884
|
+
:param _builtins.str token: The token to connect to the Kubernetes API server.
|
|
742
885
|
"""
|
|
743
886
|
pulumi.set(__self__, "cluster_ca_certificate", cluster_ca_certificate)
|
|
744
887
|
pulumi.set(__self__, "config_file", config_file)
|
|
745
888
|
pulumi.set(__self__, "host", host)
|
|
746
889
|
pulumi.set(__self__, "token", token)
|
|
747
890
|
|
|
748
|
-
@property
|
|
891
|
+
@_builtins.property
|
|
749
892
|
@pulumi.getter(name="clusterCaCertificate")
|
|
750
|
-
def cluster_ca_certificate(self) -> str:
|
|
893
|
+
def cluster_ca_certificate(self) -> _builtins.str:
|
|
751
894
|
"""
|
|
752
895
|
The CA certificate of the Kubernetes API server.
|
|
753
896
|
"""
|
|
754
897
|
return pulumi.get(self, "cluster_ca_certificate")
|
|
755
898
|
|
|
756
|
-
@property
|
|
899
|
+
@_builtins.property
|
|
757
900
|
@pulumi.getter(name="configFile")
|
|
758
|
-
def config_file(self) -> str:
|
|
901
|
+
def config_file(self) -> _builtins.str:
|
|
759
902
|
"""
|
|
760
903
|
The raw kubeconfig file.
|
|
761
904
|
"""
|
|
762
905
|
return pulumi.get(self, "config_file")
|
|
763
906
|
|
|
764
|
-
@property
|
|
907
|
+
@_builtins.property
|
|
765
908
|
@pulumi.getter
|
|
766
|
-
def host(self) -> str:
|
|
909
|
+
def host(self) -> _builtins.str:
|
|
767
910
|
"""
|
|
768
911
|
The URL of the Kubernetes API server.
|
|
769
912
|
"""
|
|
770
913
|
return pulumi.get(self, "host")
|
|
771
914
|
|
|
772
|
-
@property
|
|
915
|
+
@_builtins.property
|
|
773
916
|
@pulumi.getter
|
|
774
|
-
def token(self) -> str:
|
|
917
|
+
def token(self) -> _builtins.str:
|
|
775
918
|
"""
|
|
776
919
|
The token to connect to the Kubernetes API server.
|
|
777
920
|
"""
|
|
@@ -781,21 +924,21 @@ class GetClusterKubeconfigResult(dict):
|
|
|
781
924
|
@pulumi.output_type
|
|
782
925
|
class GetClusterOpenIdConnectConfigResult(dict):
|
|
783
926
|
def __init__(__self__, *,
|
|
784
|
-
client_id: str,
|
|
785
|
-
groups_claims: Sequence[str],
|
|
786
|
-
groups_prefix: str,
|
|
787
|
-
issuer_url: str,
|
|
788
|
-
required_claims: Sequence[str],
|
|
789
|
-
username_claim: str,
|
|
790
|
-
username_prefix: str):
|
|
791
|
-
"""
|
|
792
|
-
:param str client_id: A client id that all tokens must be issued for
|
|
793
|
-
:param Sequence[str] groups_claims: JWT claim to use as the user's group
|
|
794
|
-
:param str groups_prefix: Prefix prepended to group claims
|
|
795
|
-
:param str issuer_url: URL of the provider which allows the API server to discover public signing keys
|
|
796
|
-
:param Sequence[str] required_claims: Multiple key=value pairs that describes a required claim in the ID Token
|
|
797
|
-
:param str username_claim: JWT claim to use as the user name
|
|
798
|
-
:param str username_prefix: Prefix prepended to username
|
|
927
|
+
client_id: _builtins.str,
|
|
928
|
+
groups_claims: Sequence[_builtins.str],
|
|
929
|
+
groups_prefix: _builtins.str,
|
|
930
|
+
issuer_url: _builtins.str,
|
|
931
|
+
required_claims: Sequence[_builtins.str],
|
|
932
|
+
username_claim: _builtins.str,
|
|
933
|
+
username_prefix: _builtins.str):
|
|
934
|
+
"""
|
|
935
|
+
:param _builtins.str client_id: A client id that all tokens must be issued for
|
|
936
|
+
:param Sequence[_builtins.str] groups_claims: JWT claim to use as the user's group
|
|
937
|
+
:param _builtins.str groups_prefix: Prefix prepended to group claims
|
|
938
|
+
:param _builtins.str issuer_url: URL of the provider which allows the API server to discover public signing keys
|
|
939
|
+
:param Sequence[_builtins.str] required_claims: Multiple key=value pairs that describes a required claim in the ID Token
|
|
940
|
+
:param _builtins.str username_claim: JWT claim to use as the user name
|
|
941
|
+
:param _builtins.str username_prefix: Prefix prepended to username
|
|
799
942
|
"""
|
|
800
943
|
pulumi.set(__self__, "client_id", client_id)
|
|
801
944
|
pulumi.set(__self__, "groups_claims", groups_claims)
|
|
@@ -805,57 +948,57 @@ class GetClusterOpenIdConnectConfigResult(dict):
|
|
|
805
948
|
pulumi.set(__self__, "username_claim", username_claim)
|
|
806
949
|
pulumi.set(__self__, "username_prefix", username_prefix)
|
|
807
950
|
|
|
808
|
-
@property
|
|
951
|
+
@_builtins.property
|
|
809
952
|
@pulumi.getter(name="clientId")
|
|
810
|
-
def client_id(self) -> str:
|
|
953
|
+
def client_id(self) -> _builtins.str:
|
|
811
954
|
"""
|
|
812
955
|
A client id that all tokens must be issued for
|
|
813
956
|
"""
|
|
814
957
|
return pulumi.get(self, "client_id")
|
|
815
958
|
|
|
816
|
-
@property
|
|
959
|
+
@_builtins.property
|
|
817
960
|
@pulumi.getter(name="groupsClaims")
|
|
818
|
-
def groups_claims(self) -> Sequence[str]:
|
|
961
|
+
def groups_claims(self) -> Sequence[_builtins.str]:
|
|
819
962
|
"""
|
|
820
963
|
JWT claim to use as the user's group
|
|
821
964
|
"""
|
|
822
965
|
return pulumi.get(self, "groups_claims")
|
|
823
966
|
|
|
824
|
-
@property
|
|
967
|
+
@_builtins.property
|
|
825
968
|
@pulumi.getter(name="groupsPrefix")
|
|
826
|
-
def groups_prefix(self) -> str:
|
|
969
|
+
def groups_prefix(self) -> _builtins.str:
|
|
827
970
|
"""
|
|
828
971
|
Prefix prepended to group claims
|
|
829
972
|
"""
|
|
830
973
|
return pulumi.get(self, "groups_prefix")
|
|
831
974
|
|
|
832
|
-
@property
|
|
975
|
+
@_builtins.property
|
|
833
976
|
@pulumi.getter(name="issuerUrl")
|
|
834
|
-
def issuer_url(self) -> str:
|
|
977
|
+
def issuer_url(self) -> _builtins.str:
|
|
835
978
|
"""
|
|
836
979
|
URL of the provider which allows the API server to discover public signing keys
|
|
837
980
|
"""
|
|
838
981
|
return pulumi.get(self, "issuer_url")
|
|
839
982
|
|
|
840
|
-
@property
|
|
983
|
+
@_builtins.property
|
|
841
984
|
@pulumi.getter(name="requiredClaims")
|
|
842
|
-
def required_claims(self) -> Sequence[str]:
|
|
985
|
+
def required_claims(self) -> Sequence[_builtins.str]:
|
|
843
986
|
"""
|
|
844
987
|
Multiple key=value pairs that describes a required claim in the ID Token
|
|
845
988
|
"""
|
|
846
989
|
return pulumi.get(self, "required_claims")
|
|
847
990
|
|
|
848
|
-
@property
|
|
991
|
+
@_builtins.property
|
|
849
992
|
@pulumi.getter(name="usernameClaim")
|
|
850
|
-
def username_claim(self) -> str:
|
|
993
|
+
def username_claim(self) -> _builtins.str:
|
|
851
994
|
"""
|
|
852
995
|
JWT claim to use as the user name
|
|
853
996
|
"""
|
|
854
997
|
return pulumi.get(self, "username_claim")
|
|
855
998
|
|
|
856
|
-
@property
|
|
999
|
+
@_builtins.property
|
|
857
1000
|
@pulumi.getter(name="usernamePrefix")
|
|
858
|
-
def username_prefix(self) -> str:
|
|
1001
|
+
def username_prefix(self) -> _builtins.str:
|
|
859
1002
|
"""
|
|
860
1003
|
Prefix prepended to username
|
|
861
1004
|
"""
|
|
@@ -865,77 +1008,128 @@ class GetClusterOpenIdConnectConfigResult(dict):
|
|
|
865
1008
|
@pulumi.output_type
|
|
866
1009
|
class GetPoolNodeResult(dict):
|
|
867
1010
|
def __init__(__self__, *,
|
|
868
|
-
|
|
869
|
-
|
|
870
|
-
|
|
871
|
-
|
|
872
|
-
|
|
873
|
-
|
|
874
|
-
|
|
875
|
-
:param str
|
|
876
|
-
:param str
|
|
877
|
-
|
|
1011
|
+
id: _builtins.str,
|
|
1012
|
+
name: _builtins.str,
|
|
1013
|
+
private_ips: Sequence['outputs.GetPoolNodePrivateIpResult'],
|
|
1014
|
+
public_ip: _builtins.str,
|
|
1015
|
+
public_ip_v6: _builtins.str,
|
|
1016
|
+
status: _builtins.str):
|
|
1017
|
+
"""
|
|
1018
|
+
:param _builtins.str id: The ID of the pool.
|
|
1019
|
+
:param _builtins.str name: The pool name. Only one of `name` and `pool_id` should be specified. `cluster_id` should be specified with `name`.
|
|
1020
|
+
:param Sequence['GetPoolNodePrivateIpArgs'] private_ips: List of private IPv4 and IPv6 addresses associated with the node
|
|
1021
|
+
:param _builtins.str public_ip: The public IPv4.
|
|
1022
|
+
:param _builtins.str public_ip_v6: The public IPv6.
|
|
1023
|
+
:param _builtins.str status: The status of the node.
|
|
1024
|
+
"""
|
|
1025
|
+
pulumi.set(__self__, "id", id)
|
|
878
1026
|
pulumi.set(__self__, "name", name)
|
|
1027
|
+
pulumi.set(__self__, "private_ips", private_ips)
|
|
879
1028
|
pulumi.set(__self__, "public_ip", public_ip)
|
|
880
1029
|
pulumi.set(__self__, "public_ip_v6", public_ip_v6)
|
|
881
1030
|
pulumi.set(__self__, "status", status)
|
|
882
1031
|
|
|
883
|
-
@property
|
|
1032
|
+
@_builtins.property
|
|
884
1033
|
@pulumi.getter
|
|
885
|
-
def
|
|
1034
|
+
def id(self) -> _builtins.str:
|
|
1035
|
+
"""
|
|
1036
|
+
The ID of the pool.
|
|
1037
|
+
"""
|
|
1038
|
+
return pulumi.get(self, "id")
|
|
1039
|
+
|
|
1040
|
+
@_builtins.property
|
|
1041
|
+
@pulumi.getter
|
|
1042
|
+
def name(self) -> _builtins.str:
|
|
886
1043
|
"""
|
|
887
1044
|
The pool name. Only one of `name` and `pool_id` should be specified. `cluster_id` should be specified with `name`.
|
|
888
1045
|
"""
|
|
889
1046
|
return pulumi.get(self, "name")
|
|
890
1047
|
|
|
891
|
-
@property
|
|
1048
|
+
@_builtins.property
|
|
1049
|
+
@pulumi.getter(name="privateIps")
|
|
1050
|
+
def private_ips(self) -> Sequence['outputs.GetPoolNodePrivateIpResult']:
|
|
1051
|
+
"""
|
|
1052
|
+
List of private IPv4 and IPv6 addresses associated with the node
|
|
1053
|
+
"""
|
|
1054
|
+
return pulumi.get(self, "private_ips")
|
|
1055
|
+
|
|
1056
|
+
@_builtins.property
|
|
892
1057
|
@pulumi.getter(name="publicIp")
|
|
893
|
-
def public_ip(self) -> str:
|
|
1058
|
+
def public_ip(self) -> _builtins.str:
|
|
894
1059
|
"""
|
|
895
1060
|
The public IPv4.
|
|
896
1061
|
"""
|
|
897
1062
|
return pulumi.get(self, "public_ip")
|
|
898
1063
|
|
|
899
|
-
@property
|
|
1064
|
+
@_builtins.property
|
|
900
1065
|
@pulumi.getter(name="publicIpV6")
|
|
901
|
-
def public_ip_v6(self) -> str:
|
|
1066
|
+
def public_ip_v6(self) -> _builtins.str:
|
|
902
1067
|
"""
|
|
903
1068
|
The public IPv6.
|
|
904
1069
|
"""
|
|
905
1070
|
return pulumi.get(self, "public_ip_v6")
|
|
906
1071
|
|
|
907
|
-
@property
|
|
1072
|
+
@_builtins.property
|
|
908
1073
|
@pulumi.getter
|
|
909
|
-
def status(self) -> str:
|
|
1074
|
+
def status(self) -> _builtins.str:
|
|
910
1075
|
"""
|
|
911
1076
|
The status of the node.
|
|
912
1077
|
"""
|
|
913
1078
|
return pulumi.get(self, "status")
|
|
914
1079
|
|
|
915
1080
|
|
|
1081
|
+
@pulumi.output_type
|
|
1082
|
+
class GetPoolNodePrivateIpResult(dict):
|
|
1083
|
+
def __init__(__self__, *,
|
|
1084
|
+
address: _builtins.str,
|
|
1085
|
+
id: _builtins.str):
|
|
1086
|
+
"""
|
|
1087
|
+
:param _builtins.str address: The private IP address
|
|
1088
|
+
:param _builtins.str id: The ID of the pool.
|
|
1089
|
+
"""
|
|
1090
|
+
pulumi.set(__self__, "address", address)
|
|
1091
|
+
pulumi.set(__self__, "id", id)
|
|
1092
|
+
|
|
1093
|
+
@_builtins.property
|
|
1094
|
+
@pulumi.getter
|
|
1095
|
+
def address(self) -> _builtins.str:
|
|
1096
|
+
"""
|
|
1097
|
+
The private IP address
|
|
1098
|
+
"""
|
|
1099
|
+
return pulumi.get(self, "address")
|
|
1100
|
+
|
|
1101
|
+
@_builtins.property
|
|
1102
|
+
@pulumi.getter
|
|
1103
|
+
def id(self) -> _builtins.str:
|
|
1104
|
+
"""
|
|
1105
|
+
The ID of the pool.
|
|
1106
|
+
"""
|
|
1107
|
+
return pulumi.get(self, "id")
|
|
1108
|
+
|
|
1109
|
+
|
|
916
1110
|
@pulumi.output_type
|
|
917
1111
|
class GetPoolUpgradePolicyResult(dict):
|
|
918
1112
|
def __init__(__self__, *,
|
|
919
|
-
max_surge: int,
|
|
920
|
-
max_unavailable: int):
|
|
1113
|
+
max_surge: _builtins.int,
|
|
1114
|
+
max_unavailable: _builtins.int):
|
|
921
1115
|
"""
|
|
922
|
-
:param int max_surge: The maximum number of nodes to be created during the upgrade
|
|
923
|
-
:param int max_unavailable: The maximum number of nodes that can be not ready at the same time
|
|
1116
|
+
:param _builtins.int max_surge: The maximum number of nodes to be created during the upgrade
|
|
1117
|
+
:param _builtins.int max_unavailable: The maximum number of nodes that can be not ready at the same time
|
|
924
1118
|
"""
|
|
925
1119
|
pulumi.set(__self__, "max_surge", max_surge)
|
|
926
1120
|
pulumi.set(__self__, "max_unavailable", max_unavailable)
|
|
927
1121
|
|
|
928
|
-
@property
|
|
1122
|
+
@_builtins.property
|
|
929
1123
|
@pulumi.getter(name="maxSurge")
|
|
930
|
-
def max_surge(self) -> int:
|
|
1124
|
+
def max_surge(self) -> _builtins.int:
|
|
931
1125
|
"""
|
|
932
1126
|
The maximum number of nodes to be created during the upgrade
|
|
933
1127
|
"""
|
|
934
1128
|
return pulumi.get(self, "max_surge")
|
|
935
1129
|
|
|
936
|
-
@property
|
|
1130
|
+
@_builtins.property
|
|
937
1131
|
@pulumi.getter(name="maxUnavailable")
|
|
938
|
-
def max_unavailable(self) -> int:
|
|
1132
|
+
def max_unavailable(self) -> _builtins.int:
|
|
939
1133
|
"""
|
|
940
1134
|
The maximum number of nodes that can be not ready at the same time
|
|
941
1135
|
"""
|