dev-lxc 3.2.0 → 3.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/.gitattributes +1 -0
- data/.gitignore +17 -17
- data/CHANGELOG.md +352 -342
- data/Gemfile +4 -4
- data/LICENSE +201 -201
- data/README.md +235 -233
- data/Rakefile +1 -1
- data/bin/dl +9 -9
- data/dev-lxc.gemspec +26 -26
- data/docs/adhoc_clusters.md +17 -20
- data/docs/base_containers.md +44 -44
- data/docs/byobu_keybindings.md +22 -22
- data/docs/configuration.md +277 -235
- data/docs/dev-lxc_version_2.md +10 -10
- data/docs/manage_multiple_clusters.md +30 -30
- data/docs/mitmproxy.md +7 -7
- data/docs/usage.md +213 -213
- data/example-clusters/README.md +165 -35
- data/example-clusters/automate_dev-lxc.yml +98 -98
- data/example-clusters/chef-backend_dev-lxc.yml +81 -81
- data/example-clusters/conf-files/chef-server/elasticsearch-partial.rb +38 -0
- data/example-clusters/conf-files/chef-server/ldap-partial.rb +10 -0
- data/example-clusters/conf-files/chef-server/postgres-partial.rb +14 -0
- data/example-clusters/external_dev-lxc.yml +60 -0
- data/example-clusters/tier_dev-lxc.yml +88 -88
- data/lib/dev-lxc/cli.rb +542 -537
- data/lib/dev-lxc/cluster.rb +1310 -1215
- data/lib/dev-lxc/container.rb +128 -123
- data/lib/dev-lxc/server.rb +197 -197
- data/lib/dev-lxc/version.rb +3 -3
- data/lib/dev-lxc.rb +118 -118
- metadata +7 -2
data/docs/dev-lxc_version_2.md
CHANGED
@@ -1,10 +1,10 @@
|
|
1
|
-
# dev-lxc 2.0 is Available
|
2
|
-
|
3
|
-
Here are some of the new features which provide a significantly simplified and streamlined usage.
|
4
|
-
|
5
|
-
* mixlib-install library is used to automatically manage a cache of product packages
|
6
|
-
* Genuine container snapshot management (make as many snapshots as you want)
|
7
|
-
* New "nodes" server type which auto configures nodes for a Chef Server in the same cluster
|
8
|
-
* Removed all xc-... bash functions because the new "nodes" server type replaces this functionality
|
9
|
-
* Able to build Chef Server HA 2.0 cluster using chef-backend
|
10
|
-
* Updated and simplified READMEs
|
1
|
+
# dev-lxc 2.0 is Available
|
2
|
+
|
3
|
+
Here are some of the new features which provide a significantly simplified and streamlined usage.
|
4
|
+
|
5
|
+
* mixlib-install library is used to automatically manage a cache of product packages
|
6
|
+
* Genuine container snapshot management (make as many snapshots as you want)
|
7
|
+
* New "nodes" server type which auto configures nodes for a Chef Server in the same cluster
|
8
|
+
* Removed all xc-... bash functions because the new "nodes" server type replaces this functionality
|
9
|
+
* Able to build Chef Server HA 2.0 cluster using chef-backend
|
10
|
+
* Updated and simplified READMEs
|
@@ -1,30 +1,30 @@
|
|
1
|
-
### Maintain Uniqueness Across Multiple Clusters
|
2
|
-
|
3
|
-
The default cluster configs are already designed to be unique from each other but as you build
|
4
|
-
more clusters you have to maintain uniqueness across the YAML config files for the following items.
|
5
|
-
|
6
|
-
* Server names, `api_fqdn` and `analytics_fqdn`
|
7
|
-
|
8
|
-
Server names should really be unique across all clusters.
|
9
|
-
|
10
|
-
Even when cluster A is shutdown, if cluster B uses the same server names when it is created it
|
11
|
-
will use the already existing servers from cluster A.
|
12
|
-
|
13
|
-
`api_fqdn` and `analytics_fqdn` uniqueness only matters when clusters with the same `api_fqdn`
|
14
|
-
and `analytics_fqdn` are running.
|
15
|
-
|
16
|
-
If cluster B is started with the same `api_fqdn` or `analytics_fqdn` as an already running cluster A,
|
17
|
-
then cluster B will overwrite cluster A's DNS resolution of `api_fqdn` or `analytics_fqdn`.
|
18
|
-
|
19
|
-
* IP Addresses
|
20
|
-
|
21
|
-
IP addresses uniqueness only matters when clusters with the same IP's are running.
|
22
|
-
|
23
|
-
If cluster B is started with the same IP's as an already running cluster A, then cluster B
|
24
|
-
will overwrite cluster A's DHCP reservation of the IP's but dnsmasq will still refuse to
|
25
|
-
assign the IP's to cluster B because they already in use by cluster A. dnsmasq then assigns
|
26
|
-
random IP's from the DHCP pool to cluster B leaving it in an unexpected state.
|
27
|
-
|
28
|
-
The `dev-lxc-platform` creates the IP range 10.0.3.150 - 254 for DHCP reserved IP's.
|
29
|
-
|
30
|
-
Use unique IP's from that range when configuring clusters.
|
1
|
+
### Maintain Uniqueness Across Multiple Clusters
|
2
|
+
|
3
|
+
The default cluster configs are already designed to be unique from each other but as you build
|
4
|
+
more clusters you have to maintain uniqueness across the YAML config files for the following items.
|
5
|
+
|
6
|
+
* Server names, `api_fqdn` and `analytics_fqdn`
|
7
|
+
|
8
|
+
Server names should really be unique across all clusters.
|
9
|
+
|
10
|
+
Even when cluster A is shutdown, if cluster B uses the same server names when it is created it
|
11
|
+
will use the already existing servers from cluster A.
|
12
|
+
|
13
|
+
`api_fqdn` and `analytics_fqdn` uniqueness only matters when clusters with the same `api_fqdn`
|
14
|
+
and `analytics_fqdn` are running.
|
15
|
+
|
16
|
+
If cluster B is started with the same `api_fqdn` or `analytics_fqdn` as an already running cluster A,
|
17
|
+
then cluster B will overwrite cluster A's DNS resolution of `api_fqdn` or `analytics_fqdn`.
|
18
|
+
|
19
|
+
* IP Addresses
|
20
|
+
|
21
|
+
IP addresses uniqueness only matters when clusters with the same IP's are running.
|
22
|
+
|
23
|
+
If cluster B is started with the same IP's as an already running cluster A, then cluster B
|
24
|
+
will overwrite cluster A's DHCP reservation of the IP's but dnsmasq will still refuse to
|
25
|
+
assign the IP's to cluster B because they already in use by cluster A. dnsmasq then assigns
|
26
|
+
random IP's from the DHCP pool to cluster B leaving it in an unexpected state.
|
27
|
+
|
28
|
+
The `dev-lxc-platform` creates the IP range 10.0.3.150 - 254 for DHCP reserved IP's.
|
29
|
+
|
30
|
+
Use unique IP's from that range when configuring clusters.
|
data/docs/mitmproxy.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1
|
-
### Use mitmproxy to view HTTP traffic
|
2
|
-
|
3
|
-
Run `mitmproxy` in a terminal on the host instance.
|
4
|
-
|
5
|
-
Uncomment the `https_proxy` line in the chef-repo's `.chef/knife.rb` or in a node's `/etc/chef/client.rb` so traffic from knife commands or chef-client runs will be proxied through mitmproxy making the HTTP requests visible in the mitmproxy console.
|
6
|
-
|
7
|
-
If you enabled local port forwarding for port 8080 in your workstation's SSH config file and configured your web browser to use `127.0.0.1:8080` for HTTP and HTTPS proxies as described in the [dev-lxc-platform README.md](https://github.com/jeremiahsnapp/dev-lxc-platform) then you should be able to see the HTTP requests appear in the mitmproxy console.
|
1
|
+
### Use mitmproxy to view HTTP traffic
|
2
|
+
|
3
|
+
Run `mitmproxy` in a terminal on the host instance.
|
4
|
+
|
5
|
+
Uncomment the `https_proxy` line in the chef-repo's `.chef/knife.rb` or in a node's `/etc/chef/client.rb` so traffic from knife commands or chef-client runs will be proxied through mitmproxy making the HTTP requests visible in the mitmproxy console.
|
6
|
+
|
7
|
+
If you enabled local port forwarding for port 8080 in your workstation's SSH config file and configured your web browser to use `127.0.0.1:8080` for HTTP and HTTPS proxies as described in the [dev-lxc-platform README.md](https://github.com/jeremiahsnapp/dev-lxc-platform) then you should be able to see the HTTP requests appear in the mitmproxy console.
|
data/docs/usage.md
CHANGED
@@ -1,213 +1,213 @@
|
|
1
|
-
## Usage
|
2
|
-
|
3
|
-
### dl Command and Subcommands
|
4
|
-
|
5
|
-
`dl` is the dev-lxc command line tool.
|
6
|
-
|
7
|
-
`dev-lxc` subcommands and some options can be auto-completed by pressing the `Tab` key.
|
8
|
-
|
9
|
-
You only have to type enough of a `dev-lxc` subcommand to make it unique.
|
10
|
-
|
11
|
-
For example, the following commands are equivalent:
|
12
|
-
|
13
|
-
```
|
14
|
-
dl help
|
15
|
-
dl he
|
16
|
-
```
|
17
|
-
|
18
|
-
### Display dev-lxc help
|
19
|
-
|
20
|
-
```
|
21
|
-
dl help
|
22
|
-
|
23
|
-
dl help <subcommand>
|
24
|
-
```
|
25
|
-
|
26
|
-
### Configure a cluster
|
27
|
-
|
28
|
-
See the [configuration docs](docs/configuration.md) to learn how to use the `dl init` command to create and configure a `dev-lxc.yml` file.
|
29
|
-
|
30
|
-
### cluster-view, tks, tls commands
|
31
|
-
|
32
|
-
The dev-lxc-platform comes with some commands that create and manage helpful
|
33
|
-
tmux/byobu sessions to more easily see the state of a cluster.
|
34
|
-
|
35
|
-
Running the `cluster-view` command in the same directory as a `dev-lxc.yml` file
|
36
|
-
creates a tmux/byobu session with the same name as the cluster's directory.
|
37
|
-
|
38
|
-
`cluster-view` can also be run with the parent directory of a `dev-lxc.yml` file
|
39
|
-
as the first argument and `cluster-view` will change to that directory before
|
40
|
-
creating the tmux/byobu session.
|
41
|
-
|
42
|
-
The session's first window is named "cluster".
|
43
|
-
|
44
|
-
The left side is for running dev-lxc commands.
|
45
|
-
|
46
|
-
The right side updates every 0.5 seconds with the cluster's status provided by `dl status`.
|
47
|
-
|
48
|
-
The session's second window is named "shell". It opens in the same directory as the
|
49
|
-
cluster's `dev-lxc.yml` file.
|
50
|
-
|
51
|
-
The `tls` and `tks` commands are really aliases.
|
52
|
-
|
53
|
-
`tls` is an alias for `tmux list-sessions` and is used to see what tmux/byobu sessions
|
54
|
-
are running.
|
55
|
-
|
56
|
-
`tks` is an alias for `tmux kill-session -t` and is used to kill tmux/byobu sessions.
|
57
|
-
When specifying the session to be killed you only need as many characters of the session
|
58
|
-
name that are required to make the name unique among the list of running sessions.
|
59
|
-
|
60
|
-
I recommend switching to a different running tmux/byobu session before killing the current
|
61
|
-
tmux/byobu session. Otherwise you will need to reattach to the remaining tmux/byobu session.
|
62
|
-
Use the keyboard shortcuts Alt-Up/Down to easily switch between tmux/byobu sessions.
|
63
|
-
|
64
|
-
### Cluster status
|
65
|
-
|
66
|
-
Run the following command to see the status of the cluster.
|
67
|
-
|
68
|
-
```
|
69
|
-
dl status
|
70
|
-
```
|
71
|
-
|
72
|
-
This is an example of the output.
|
73
|
-
|
74
|
-
```
|
75
|
-
chef.lxc NOT_CREATED
|
76
|
-
|
77
|
-
analytics.lxc NOT_CREATED
|
78
|
-
|
79
|
-
supermarket.lxc NOT_CREATED
|
80
|
-
|
81
|
-
node-1.lxc NOT_CREATED
|
82
|
-
```
|
83
|
-
|
84
|
-
### Specifying a Subset of Servers
|
85
|
-
|
86
|
-
Many dev-lxc subcommands can act on a subset of the cluster's servers by specifying a regular expression that matches the desired server names.
|
87
|
-
|
88
|
-
For example, the following command will show the status of the Chef Server.
|
89
|
-
|
90
|
-
```
|
91
|
-
dl status chef
|
92
|
-
```
|
93
|
-
|
94
|
-
### Start cluster
|
95
|
-
|
96
|
-
Starting the cluster the first time takes awhile since it has a lot to download and build.
|
97
|
-
|
98
|
-
```
|
99
|
-
dl up
|
100
|
-
```
|
101
|
-
|
102
|
-
A test org, users, knife.rb and keys are automatically created in
|
103
|
-
the bootstrap backend server in `/root/chef-repo/.chef` for testing purposes.
|
104
|
-
|
105
|
-
The `knife-opc` plugin is installed in the embedded ruby environment of the
|
106
|
-
Private Chef and Enterprise Chef server to facilitate the creation of the test
|
107
|
-
org and user.
|
108
|
-
|
109
|
-
Note: You also have the option of running the `prepare-product-cache` subcommand which downloads required product packages to the cache.
|
110
|
-
This can be helpful when you don't want to start building the cluster yet but you want the package cache ready when you build the cluster later.
|
111
|
-
|
112
|
-
```
|
113
|
-
dl prepare-product-cache
|
114
|
-
```
|
115
|
-
|
116
|
-
### Print Chef Automate Credentials
|
117
|
-
|
118
|
-
If the cluster has a Chef Automate server you can use the `print-automate-credentials` subcommand to see what the login credentials.
|
119
|
-
|
120
|
-
```
|
121
|
-
dl print-automate-credentials
|
122
|
-
```
|
123
|
-
|
124
|
-
### Create chef-repo
|
125
|
-
|
126
|
-
Create a `.chef` directory in the current directory with appropriate knife.rb and pem files.
|
127
|
-
|
128
|
-
Use the `-p` option to also get pivotal.pem and pivotal.rb files.
|
129
|
-
|
130
|
-
Use the `-f` option to overwrite existing knife.rb and pivotal.rb files.
|
131
|
-
|
132
|
-
```
|
133
|
-
dl chef-repo
|
134
|
-
```
|
135
|
-
|
136
|
-
Now you can easily use knife to access the cluster.
|
137
|
-
|
138
|
-
```
|
139
|
-
knife client list
|
140
|
-
```
|
141
|
-
|
142
|
-
### Stop and start the cluster
|
143
|
-
|
144
|
-
```
|
145
|
-
dl halt
|
146
|
-
dl up
|
147
|
-
```
|
148
|
-
|
149
|
-
### Run arbitrary commands in each server
|
150
|
-
|
151
|
-
```
|
152
|
-
dl run-command chef 'uptime'
|
153
|
-
```
|
154
|
-
|
155
|
-
### Attach the terminal to a server
|
156
|
-
|
157
|
-
Attach the terminal to a server in the cluster that matches the REGEX pattern given.
|
158
|
-
|
159
|
-
```
|
160
|
-
dl attach chef
|
161
|
-
```
|
162
|
-
|
163
|
-
### Create a snapshot of the servers
|
164
|
-
|
165
|
-
Save the changes in the servers to snapshots with a comment.
|
166
|
-
|
167
|
-
```
|
168
|
-
dl halt
|
169
|
-
dl snapshot -c 'this is a snapshot comment'
|
170
|
-
```
|
171
|
-
|
172
|
-
### List snapshots
|
173
|
-
|
174
|
-
```
|
175
|
-
dl snapshot -l
|
176
|
-
```
|
177
|
-
|
178
|
-
### Restore snapshots
|
179
|
-
|
180
|
-
Restore snapshots by name.
|
181
|
-
|
182
|
-
Leave out the snapshot name or specify `LAST` to restore the most recent snapshot.
|
183
|
-
|
184
|
-
```
|
185
|
-
dl snapshot -r
|
186
|
-
dl up
|
187
|
-
```
|
188
|
-
|
189
|
-
### Destroy snapshots
|
190
|
-
|
191
|
-
Destroy snapshots by name or destroy all snapshots by specifying `ALL`.
|
192
|
-
|
193
|
-
Leave out the snapshot name or specify `LAST` to destroy the most recent snapshots.
|
194
|
-
|
195
|
-
```
|
196
|
-
dl snapshot -d
|
197
|
-
```
|
198
|
-
|
199
|
-
### Destroy cluster
|
200
|
-
|
201
|
-
Use the following command to destroy the cluster's servers.
|
202
|
-
|
203
|
-
```
|
204
|
-
dl destroy
|
205
|
-
```
|
206
|
-
|
207
|
-
### Show Calculated Configuration
|
208
|
-
|
209
|
-
Mostly for debugging purposes you have the ability to print the calculated cluster configuration.
|
210
|
-
|
211
|
-
```
|
212
|
-
dl show-config
|
213
|
-
```
|
1
|
+
## Usage
|
2
|
+
|
3
|
+
### dl Command and Subcommands
|
4
|
+
|
5
|
+
`dl` is the dev-lxc command line tool.
|
6
|
+
|
7
|
+
`dev-lxc` subcommands and some options can be auto-completed by pressing the `Tab` key.
|
8
|
+
|
9
|
+
You only have to type enough of a `dev-lxc` subcommand to make it unique.
|
10
|
+
|
11
|
+
For example, the following commands are equivalent:
|
12
|
+
|
13
|
+
```
|
14
|
+
dl help
|
15
|
+
dl he
|
16
|
+
```
|
17
|
+
|
18
|
+
### Display dev-lxc help
|
19
|
+
|
20
|
+
```
|
21
|
+
dl help
|
22
|
+
|
23
|
+
dl help <subcommand>
|
24
|
+
```
|
25
|
+
|
26
|
+
### Configure a cluster
|
27
|
+
|
28
|
+
See the [configuration docs](docs/configuration.md) to learn how to use the `dl init` command to create and configure a `dev-lxc.yml` file.
|
29
|
+
|
30
|
+
### cluster-view, tks, tls commands
|
31
|
+
|
32
|
+
The dev-lxc-platform comes with some commands that create and manage helpful
|
33
|
+
tmux/byobu sessions to more easily see the state of a cluster.
|
34
|
+
|
35
|
+
Running the `cluster-view` command in the same directory as a `dev-lxc.yml` file
|
36
|
+
creates a tmux/byobu session with the same name as the cluster's directory.
|
37
|
+
|
38
|
+
`cluster-view` can also be run with the parent directory of a `dev-lxc.yml` file
|
39
|
+
as the first argument and `cluster-view` will change to that directory before
|
40
|
+
creating the tmux/byobu session.
|
41
|
+
|
42
|
+
The session's first window is named "cluster".
|
43
|
+
|
44
|
+
The left side is for running dev-lxc commands.
|
45
|
+
|
46
|
+
The right side updates every 0.5 seconds with the cluster's status provided by `dl status`.
|
47
|
+
|
48
|
+
The session's second window is named "shell". It opens in the same directory as the
|
49
|
+
cluster's `dev-lxc.yml` file.
|
50
|
+
|
51
|
+
The `tls` and `tks` commands are really aliases.
|
52
|
+
|
53
|
+
`tls` is an alias for `tmux list-sessions` and is used to see what tmux/byobu sessions
|
54
|
+
are running.
|
55
|
+
|
56
|
+
`tks` is an alias for `tmux kill-session -t` and is used to kill tmux/byobu sessions.
|
57
|
+
When specifying the session to be killed you only need as many characters of the session
|
58
|
+
name that are required to make the name unique among the list of running sessions.
|
59
|
+
|
60
|
+
I recommend switching to a different running tmux/byobu session before killing the current
|
61
|
+
tmux/byobu session. Otherwise you will need to reattach to the remaining tmux/byobu session.
|
62
|
+
Use the keyboard shortcuts Alt-Up/Down to easily switch between tmux/byobu sessions.
|
63
|
+
|
64
|
+
### Cluster status
|
65
|
+
|
66
|
+
Run the following command to see the status of the cluster.
|
67
|
+
|
68
|
+
```
|
69
|
+
dl status
|
70
|
+
```
|
71
|
+
|
72
|
+
This is an example of the output.
|
73
|
+
|
74
|
+
```
|
75
|
+
chef.lxc NOT_CREATED
|
76
|
+
|
77
|
+
analytics.lxc NOT_CREATED
|
78
|
+
|
79
|
+
supermarket.lxc NOT_CREATED
|
80
|
+
|
81
|
+
node-1.lxc NOT_CREATED
|
82
|
+
```
|
83
|
+
|
84
|
+
### Specifying a Subset of Servers
|
85
|
+
|
86
|
+
Many dev-lxc subcommands can act on a subset of the cluster's servers by specifying a regular expression that matches the desired server names.
|
87
|
+
|
88
|
+
For example, the following command will show the status of the Chef Server.
|
89
|
+
|
90
|
+
```
|
91
|
+
dl status chef
|
92
|
+
```
|
93
|
+
|
94
|
+
### Start cluster
|
95
|
+
|
96
|
+
Starting the cluster the first time takes awhile since it has a lot to download and build.
|
97
|
+
|
98
|
+
```
|
99
|
+
dl up
|
100
|
+
```
|
101
|
+
|
102
|
+
A test org, users, knife.rb and keys are automatically created in
|
103
|
+
the bootstrap backend server in `/root/chef-repo/.chef` for testing purposes.
|
104
|
+
|
105
|
+
The `knife-opc` plugin is installed in the embedded ruby environment of the
|
106
|
+
Private Chef and Enterprise Chef server to facilitate the creation of the test
|
107
|
+
org and user.
|
108
|
+
|
109
|
+
Note: You also have the option of running the `prepare-product-cache` subcommand which downloads required product packages to the cache.
|
110
|
+
This can be helpful when you don't want to start building the cluster yet but you want the package cache ready when you build the cluster later.
|
111
|
+
|
112
|
+
```
|
113
|
+
dl prepare-product-cache
|
114
|
+
```
|
115
|
+
|
116
|
+
### Print Chef Automate Credentials
|
117
|
+
|
118
|
+
If the cluster has a Chef Automate server you can use the `print-automate-credentials` subcommand to see what the login credentials.
|
119
|
+
|
120
|
+
```
|
121
|
+
dl print-automate-credentials
|
122
|
+
```
|
123
|
+
|
124
|
+
### Create chef-repo
|
125
|
+
|
126
|
+
Create a `.chef` directory in the current directory with appropriate knife.rb and pem files.
|
127
|
+
|
128
|
+
Use the `-p` option to also get pivotal.pem and pivotal.rb files.
|
129
|
+
|
130
|
+
Use the `-f` option to overwrite existing knife.rb and pivotal.rb files.
|
131
|
+
|
132
|
+
```
|
133
|
+
dl chef-repo
|
134
|
+
```
|
135
|
+
|
136
|
+
Now you can easily use knife to access the cluster.
|
137
|
+
|
138
|
+
```
|
139
|
+
knife client list
|
140
|
+
```
|
141
|
+
|
142
|
+
### Stop and start the cluster
|
143
|
+
|
144
|
+
```
|
145
|
+
dl halt
|
146
|
+
dl up
|
147
|
+
```
|
148
|
+
|
149
|
+
### Run arbitrary commands in each server
|
150
|
+
|
151
|
+
```
|
152
|
+
dl run-command chef 'uptime'
|
153
|
+
```
|
154
|
+
|
155
|
+
### Attach the terminal to a server
|
156
|
+
|
157
|
+
Attach the terminal to a server in the cluster that matches the REGEX pattern given.
|
158
|
+
|
159
|
+
```
|
160
|
+
dl attach chef
|
161
|
+
```
|
162
|
+
|
163
|
+
### Create a snapshot of the servers
|
164
|
+
|
165
|
+
Save the changes in the servers to snapshots with a comment.
|
166
|
+
|
167
|
+
```
|
168
|
+
dl halt
|
169
|
+
dl snapshot -c 'this is a snapshot comment'
|
170
|
+
```
|
171
|
+
|
172
|
+
### List snapshots
|
173
|
+
|
174
|
+
```
|
175
|
+
dl snapshot -l
|
176
|
+
```
|
177
|
+
|
178
|
+
### Restore snapshots
|
179
|
+
|
180
|
+
Restore snapshots by name.
|
181
|
+
|
182
|
+
Leave out the snapshot name or specify `LAST` to restore the most recent snapshot.
|
183
|
+
|
184
|
+
```
|
185
|
+
dl snapshot -r
|
186
|
+
dl up
|
187
|
+
```
|
188
|
+
|
189
|
+
### Destroy snapshots
|
190
|
+
|
191
|
+
Destroy snapshots by name or destroy all snapshots by specifying `ALL`.
|
192
|
+
|
193
|
+
Leave out the snapshot name or specify `LAST` to destroy the most recent snapshots.
|
194
|
+
|
195
|
+
```
|
196
|
+
dl snapshot -d
|
197
|
+
```
|
198
|
+
|
199
|
+
### Destroy cluster
|
200
|
+
|
201
|
+
Use the following command to destroy the cluster's servers.
|
202
|
+
|
203
|
+
```
|
204
|
+
dl destroy
|
205
|
+
```
|
206
|
+
|
207
|
+
### Show Calculated Configuration
|
208
|
+
|
209
|
+
Mostly for debugging purposes you have the ability to print the calculated cluster configuration.
|
210
|
+
|
211
|
+
```
|
212
|
+
dl show-config
|
213
|
+
```
|